AN EFFECTIVE APPROACH FOR LOW-ALTITUDE AERIAL OBJECT DETECTION

  • Nguyen Do Khai Hoan Hung Yen University of Technology and Education
  • Tran Do Thu Ha Hung Yen University of Technology and Education
  • Luu Hoang Minh Hung Yen University of Technology and Education
  • Nguyen Xuan Mong Hung Yen University of Technology and Education
  • Nguyen Van Dat Hung Yen University of Technology and Education
  • Truong Quoc Huy Hung Yen University of Technology and Education
  • Nguyen Thanh Binh Hung Yen University of Technology and Education

Abstract

Over the past decade, advances in aerial surveillance systems have driven increasing interest in low-altitude flying object detection within the computer vision community. However, the effectiveness of detecting and classifying flying objects, such as unmanned aerial vehicles (UAVs) and kites, is strongly influenced by the availability of specialized datasets, which remain limited. To overcome this challenge, we introduce a new dataset named UKD (UAV and Kite Dataset). The UKD dataset includes 2,121 real-world images of UAVs and kites with bounding box annotations and is employed to train and assess four representative modern detection models, such as YOLOv7, YOLOv10, RT-DETRv3, and Salience-DERT. Experimental results demonstrate that leveraging the UKD dataset leads to notable performance improvements, with the detection results achieving up to 80.3%.

References

Hsiao, C. J., “Uav Low-Altitude Agricultural Information Remote Sensing Monitoring,” J. Comput. Sci. Electr. Eng, 6, pp. 1-8, 2024.

Zhou, Y., “Unmanned aerial vehicles based low-altitude economy with lifecycle techno-economic- environmental analysis for sustainable and smart cities,” Journal of Cleaner Production, 145050, 2025.

Uzuntaş, K. N., & Aslan, M., “Unmanned aerial vehicles and aviation safety: a qualitative study,” The Aeronautical Journal, 1, 19, 2025.

Girshick, R., Donahue, J., Darrell, T., Malik, J.: “Rich feature hierarchies for accurate object detection and semantic segmentation,” In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 580–587.

Ren, S., He, K., Girshick, R., & Sun, J., “Faster r-cnn: Towards real-time object detection with region proposal networks,” Advances in neural information processing systems, 28, 2015.

Wang, C. Y., Bochkovskiy, A., & Liao, H. Y. M., “YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors,” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 7464-7475.

Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., & Han, J., “Yolov10: Real-time end-to-end object detection,” Advances in Neural Information Processing Systems, 37, pp. 107984-108011, 2024.

Wang, S., Xia, C., Lv, F., & Shi, Y., “RT-DETRv3: Real-time end-to-end object detection with hierarchical dense positive supervision,” In 2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), February, 2025, pp. 1628-1636.

Hou, X., Liu, M., Zhang, S., Wei, P., & Chen, B., “Salience detr: Enhancing detection transformer with hierarchical salience filtering refinement,” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2024, pp. 17574-17583.

Published
2025-12-10