基于YOLOv3的无人机识别与定位追踪

Drone identification and location tracking based on YOLOv3

  • 摘要: 近年来,无人机入侵的事件经常发生,无人机跌落碰撞的事件也屡见不鲜,在人群密集的地方容易引发安全事故,所以无人机监测是目前安防领域的研究热点。虽然目前有很多种无人机监测方案,但大多成本高昂,实施困难。在5G背景下,针对此问题提出了一种利用城市已有的监控网络去获取数据的方法,基于深度学习的算法进行无人机目标检测,进而识别无人机,并追踪定位无人机。该方法采用改进的YOLOv3模型检测视频帧中是否存在无人机,YOLOv3算法是YOLO(You only look once,一次到位)系列的第三代版本,属于one-stage目标检测算法这一类,在速度上相对于two-stage类型的算法有着明显的优势。YOLOv3输出视频帧中存在的无人机的位置信息。根据位置信息用PID(Proportion integration differentiation,比例积分微分)算法调节摄像头的中心朝向追踪无人机,再由多个摄像头的参数解算出无人机的实际坐标,从而实现定位。本文通过拍摄无人机飞行的照片、从互联网上搜索下载等方式构建了数据集,并且使用labelImg工具对图片中的无人机进行了标注,数据集按照无人机的旋翼数量进行了分类。实验中采用按旋翼数量分类后的数据集对检测模型进行训练,训练后的模型在测试集上能达到83.24%的准确率和88.15%的召回率,在配备NVIDIA GTX 1060的计算机上能达到每秒20帧的速度,可实现实时追踪。

     

    Abstract: In recent years, increasing incidents of drone intrusion have occurred, and the drone collisions have become common. As a result, accidents may occur in densely populated areas. Therefore, drone monitoring is an important research topic in the field of security. Although many types of drone monitoring programs exist, most of them are costly and difficult to implement. To solve this problem, in the 5G context, this study proposed a method of using a city’s existing monitoring network to acquire data based on a deep learning algorithm for drone target detection, constructing a recognizable drone, and tracking the unmanned aerial vehicle. The method used the improved YOLOv3 (You only look once) model to detect the presence of drones in video frames. The YOLOv3 algorithm is the third generation version of the YOLO series, belonging to the one-stage target detection algorithm. This algorithm has significant advantages over the two-stage type of algorithm in speed. YOLOv3 outputs the position information of the drone in the video frame. According to the position information, the PID (Proportion integration differentiation) algorithm was used to adjust the center of the camera to track the drone. Then, the parameters of the plurality of cameras were used to calculate the actual coordinates of the drone, thereby realizing the positioning. We built the dataset by taking photos of the drone's flight, searching and downloading drone pictures from the Internet, and labeling the drones in the image by using the labelImg tool. The dataset was classified according to the number of rotors of the drone. In the experiment, the detection model was trained by the dataset classified by the number of rotors. The trained model can achieve 83.24% accuracy and 88.15% recall rate on the test set, and speed of 20 frames per second on the computer equipped with NVIDIA GTX 1060 for real-time tracking.

     

/

返回文章
返回