高级检索+

基于改进YOLO检测与时序跟踪的草地空斑高精度定位方法

A high-precision positioning method for grassland patches based on joint optimization of recognition and tracking

  • 摘要: 为实现草地空斑补播作业中的高精度识别与定位,解决复杂草地环境下空斑目标易受光照变化、草种差异及季节干扰影响而导致跨场景检测与跟踪精度不足的问题,该研究基于深度学习检测与多目标跟踪技术,提出一种融合检测与跟踪的草地空斑高精度定位方法SCTD-YOLO。通过骨干网络引入空间金字塔池化融合卷积模块,实现多尺度特征提取并提升小空斑纹理感知能力。替换部分特征提取模块为融合滑动窗口自注意力的跨层结构,有效建模不规则空斑与背景间的长距离依赖关系。构建纹理特征提取模块,通过轻量级卷积和池化提取高频纹理特征,抑制颜色干扰,结合改进的自适应特征金字塔检测头与可变形卷积实现特征对齐,从而提升空斑的检测能力。其次,利用DeepOCSORT目标跟踪算法,实现跨帧稳定跟踪,增强在动态航拍过程中的目标一致性。试验结果表明,SCTD-YOLO模型平均精度值为95.9%,相比基础检测模型提升了2.2个百分点。SCTD-YOLO+DeepOCSORT算法平均多目标跟踪准确率为88.2%,多目标跟踪精度为86.71%,平均每个测试视频的ID切换次数为35,能够满足实时跟踪需求。空斑定位的平均误差为0.316m,满足无人机作业对草地空斑精准定位的需求,该研究不仅能为无人机智能补播提供技术支撑,还为农业机器人在复杂多样农艺模式和环境条件下的高适应性与通用性提供了有效技术支撑。

     

    Abstract: The current degradation of grasslands is severe, posing a serious threat to ecological stability, forage yield, and the sustainable development of animal husbandry. The use of agricultural robots (such as unmanned aerial vehicle (UAV) platforms) for monitoring and reseeding degraded grassland patches has become an important means of grassland ecological restoration. Given the insufficient accuracy of patch identification in existing UAV operations and the low accuracy of patch positioning due to easy target loss or unstable tracking, this study proposes a computer vision method that integrates advanced object detection and tracking algorithms to achieve efficient detection and positioning of grassland patches, providing technical support for precise reseeding operations. This study uses the YOLO series model for patch detection and introduces an adaptive feature pyramid network to enhance multi-scale feature fusion capabilities, thereby improving detection accuracy in complex grassland backgrounds. Lightweight convolution is used to achieve channel compression and nonlinear activation, while optimizing the expression ability of deep features and computational efficiency. The Swin Transformer sliding window self-attention mechanism and residual convolution structure are adopted to establish long-range dependencies between patches and the grassland background within the local receptive field. The input RGB patch images are converted to the YUV color space, and the single-channel luminance map is extracted as the input to enhance the luminance contrast of the grassland and patch texture structures. In terms of temporal consistency and target tracking, the DeepOCSORT algorithm is combined to perform multi-target tracking of patches in consecutive image sequences. The improved SCTD-YOLO detector significantly outperforms the baseline YOLOv8 in detection performance. In ablation experiments, the addition of different module combinations improved the model performance to varying degrees. When using the improved SCTD-YOLO model, the accuracy rate increased to 94.7%, the recall rate was 86.2%, and the effect was the best; the average precision increased by 2.4 percentage points, and the mAP increased by 2.2 percentage points. Under different lighting conditions, complex grassland texture backgrounds, partial patch occlusion by plants, and significant changes in patch size, the model still maintains high robustness, with significantly reduced missed detection and false detection rates. In multi-target tracking, after combining with the DeepOCSORT tracker, the algorithm effectively maintains the consistency of patches in multi-frame sequences, with an IDF1 of 85.15%, an MOTA of 88.2%, and an MOTP of 86.71%, significantly reducing false detections and ID switches. Experiments show that in scenarios of rapid camera panning, grass leaf shaking caused by wind, and dynamic background changes, DeepOCSORT can still stably maintain target trajectories, significantly improving the reliability of long-term monitoring. Finally, the predicted latitude and longitude coordinates are obtained through image modeling and external parameter calculation, and compared with high-precision RTK-GPS measurement coordinates. The spatial positioning error is calculated using the Vincenty formula. The results show that the combination of the improved SCTD-YOLO and stable tracking algorithm significantly reduces the overall positioning error, with an average error of 0.316m and an error range of 0.278m to 0.423m. The improved algorithm shows higher positioning accuracy and error stability in all test samples, effectively improving the overall consistency of target detection, tracking, and positioning. The overall results meet the requirement of less than 0.5m for positioning accuracy in intelligent reseeding operations of degraded grassland patches. The integrated framework of SCTD-YOLO and DeepOCSORT proposed in this study provides a high-precision and stable solution for patch detection, tracking, and positioning in complex multi-scenario and multi-interference environments for agricultural robots. This research not only improves the efficiency and reliability of grassland patch reseeding but also provides technical support for the generalization application of agricultural robot vision algorithms in different agronomic models and environmental conditions, which is of great significance for sustainable grassland management, resource conservation, and ecological restoration. In the future, further exploration will be conducted on seasonal monitoring and adaptive reseeding decision-making methods based on multi-source remote sensing data.

     

/

返回文章
返回