Ainvestigation of computer vision capabilities in autonomous unmanned aerial vehicles control systems
- Authors: Karelin Е.А.1, Lyubashenko T.D.1, Palilov M.R.1, Zhiglova N.S.1, Pachin A.V.1
-
Affiliations:
- Bonch-Bruevich Saint Petersburg State University of Telecommunications
- Issue: Vol 22, No 1 (2024)
- Pages: 102-110
- Section: Radio telecommunication, radiobroadcasting and television technologies
- URL: https://journals.eco-vector.com/2073-3909/article/view/689829
- DOI: https://doi.org/10.18469/ikt.2024.22.1.14
- ID: 689829
Cite item
Full Text
Abstract
The development of modern hardware and software has led to a rapid increase in the use of unmanned aerial vehicles, primarily aircraft. One of the promising areas for improving the efficiency of such platforms is the development of autonomous control systems, eliminating the need for human operators. The article presents the results of a study on the possibility of constructing elements of an automatic motion control system based on computer vision for unmanned aerial vehicles. The authors used comparative analysis to evaluate the advantages of two popular instruments: YOLO and SSD. They have also collected data for training the chosen model and tested it in various conditions, described the methodology for creating a training set and presented the model testing results on video images captured by a UAV's camera. The test results confirm that the YOLOv8n model is suitable for object detection on the UAV board using a Raspberry Pi 4 Model B as the single-board hardware platform. The object detection accuracy was from 80% to 90%, with the power consumption of 15-25 watts.
About the authors
Е. А. Karelin
Bonch-Bruevich Saint Petersburg State University of Telecommunications
Author for correspondence.
Email: evgeniikarelin01@mail.ru
Student of Software Engineering and Computer Engineering Department
Russian Federation, Saint PetersburgT. D. Lyubashenko
Bonch-Bruevich Saint Petersburg State University of Telecommunications
Email: tima50879@gmail.com
Student of Secure Communication Systems Department
Russian Federation, Saint PetersburgM. R. Palilov
Bonch-Bruevich Saint Petersburg State University of Telecommunications
Email: palilovfox@gmail.com
Student of Secure Communication Systems Department
Russian Federation, Saint PetersburgN. S. Zhiglova
Bonch-Bruevich Saint Petersburg State University of Telecommunications
Email: zhiglova.natalia@yandex.ru
Student of Secure Communication Systems Department
Russian Federation, Saint PetersburgA. V. Pachin
Bonch-Bruevich Saint Petersburg State University of Telecommunications
Email: pachin.andrej@bk.ru
Associate Professor of Software Engineering and Computer Engineering Department, PhD in Technical Science
Russian Federation, Saint PetersburgReferences
- Alkentar S.M. et al. Practical comparation of the accuracy and speed of YOLO, SSD and Faster RCNN for drone detection. Journal of Engineering, 2021, vol. 27, no. 8, pp. 19–31. doi: 10.31026/j.eng.2021.08.02
- Choudhari V. et al. Comparison between YOLO and SSD MobileNet for object detection in a surveillance drone. URL: https://www.researchgate.net/publication/355336797_Comparison_between_YOLO_and_SSD_MobileNet_for_Object_Detection_in_a_Surveillance_Drone (accessed: 01.06.2024).
- Li M. et al. Agricultural greenhouses detection in high-resolution satellite images based on convolutional neural networks: comparison of faster R-CNN, YOLO v3 and SSD. Sensors, 2020, vol. 20, no. 17. URL: https://www.sci-hub.ru/10.3390/s20174938 (accessed: 01.06.2024).
- Soylu E., Soylu T. A performance comparison of YOLOv8 models for traffic sign detection in the Robotaxi-full scale autonomous vehicle competition. Multimedia Tools and Applications, 2024, vol. 83, pp. 25005–25035. doi: 10.1007/s11042-023-16451-1
- Long X. et al. PP-YOLO: an effective and efficient implementation of object detector. URL: https://www.researchgate.net/publication/343178810_PP-YOLO_An_Effective_and_Efficient_Implementation_of_Object_Detector (accessed: 05.06.2024).
- Osco L.P. et al. A review on deep learning in UAV remote sensing. International Journal of Applied Earth Observation and Geoinformation. URL: https://www.sci-hub.ru/10.1016/j.jag.2021.102456 (accessed: 05.06.2024).
- Joseph R. et al. You only look once: unified, real-time object detection. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, 2015, pp. 779–788.
- Selcuk B., Serif T. A comparison of YOLOV5 and YOLOV8 in the context of mobile UI detection. Lecture Notes in Computer Science, 2023, vol. 13977, pp. 161–174. doi: 10.1007/978-3-031-39764-6_11
- Zagitov A. et al. Comparative analysis of neural network models performance on low-power devices for a real-time object detection task. Journal Computer Optics, 2024, vol. 48, no. 2, pp. 242–252. doi: 10.18287/2412-6179-CO-1343
- Zhang H. et al. A review of unmanned aerial vehicle low-altitude remote sensing (UAV-LARS) use in agricultural monitoring in China. Remote Sensing, 2021, vol. 13, no. 6, pp. 1221. URL: https://doi.org/10.3390/rs13061221 (accessed: 05.06.2024).
- COCO – Common Objects in context. URL: https://cocodataset.org/#home (accessed: 18.02.2024).
- Quadrocopters for schools «Pioneer Mini» – GC «Geoscan». URL: https://www.geoscan.ru/ru/products/pioneer/mini. (accessed: 18.02.2024). (In Russ.)
- Set: F450 (E305) ARF KIT+NAZA-M V2+GPS+Landing Skid+PMU v2+LED. URL: https://www.djimsk.ru/catalog/products/components/flame_wheel_arf_kit/komplekt_f450_e305_arf_kit_naza_m_v2_gps_landing_skid_pmu_v2_led.html (accessed: 18.02.2024). (In Russ.)
- GitHub – opencv/opencv: Open Source Computer Vision Library. URL: https://github.com/opencv/opencv?ysclid=lsqmv32jem25767816 (accessed: 18.02.2024).
- GitHub – pytorch/ pytorch: Tensors and Dinamic Neural Networks in Python. URL: https://github.com/pytorch/pytorch?ysclid=lsqmug3hgs789690537 (accessed: 20.04.2024).
- Buy a Raspberry Pi 4 Model B – Raspberry Pi. URL: https://www.raspberrypi.com/products/raspberry-pi-4-model-b/?parent-baobab-id=lsqmt1h7lk (accessed: 20.04.2024).
- Raspberry Pi OS – Raspberry Pi. URL: https://www.raspberrypi.com/software/ (accessed: 25.05.2024).
- Overview of datasets – Ultralytics YOLO Docs. URL: https://docs.ultralytics.com/ru/datasets/ (accessed: 25.05.2024). (In Russ.)
- GitHub – Ultralytics. URL: https://github.com/ultralytics/ultralytics (accessed: 28.05.2024).
Supplementary files
