高级检索+

基于CBP-YOLO的玉米田间草地贪夜蛾侵染痕迹检测与试验

Detection and experiment of spodoptera frugiperda infestation traces in maize fields based on CBP-YOLO

  • 摘要: 针对无人机检测玉米田间草地贪夜蛾侵染痕迹时,因目标尺寸小、尺度变化大、特征表征不明显导致识别精度低的问题,该研究提出了一种面向无人机多尺度图像的侵染痕迹检测模型CBP-YOLO(Coordinated-BiFPN-P2-YOLO,CBP-YOLO)。通过引入了真实增强超分辨率生成对抗网络(real-enhanced super-resolution generative adversarial network,Real-ESRGAN)重建原始低分辨率图像,恢复叶片啃食区域的纹理,在YOLOv8主干网络嵌入协调注意力机制(coordinated attention, CA),提升小目标局部细节提取与全局位置信息感知能力,在颈部网络引入双向特征金字塔网络(Bi-directional feature pyramid network, BiFPN),降低跨层级特征信息损耗,增强多尺度目标特征表达一致性,新增了尺寸为160×160×64的小目标检测头,拓展模型对小尺度目标的特征感知能力。结果表明,地面采样距离(Ground Sampling Distance,GSD)为0.38cm/像素时检测性能最佳,Real-ESRGAN有效弥补了分辨率不足引起的特征信息缺失。CBP-YOLO在无人机多尺度叶片数据集上mAP@0.5达到76.5%,较原模型提升3.4个百分点,较YOLOv9m、YOLOv10m、YOLOv11m、Faster R-CNN和RetinaNet分别提升10.1、7.2、5.1、9.3和17.9个百分点,验证了其对玉米田间草地贪夜蛾侵染痕迹的检测能力。

     

    Abstract: Early and accurate detection of fall armyworm (Spodoptera frugiperda) infestation in maize fields using unmanned aerial vehicle (UAV) imagery is of paramount importance for implementing timely and effective pest management strategies. Despite its significance, achieving reliable detection remains highly challenging due to several factors. Firstly, the feeding marks caused by the larvae are often very small and subtle, making them difficult to identify at high altitudes. Secondly, significant variations in object scale occur across different flight heights, which complicates consistent recognition. Finally, under real-world field conditions, the low contrast between damaged leaf tissue and surrounding healthy foliage further hampers accurate detection, especially when lighting and environmental conditions vary. Collectively, these challenges underscore the need for advanced computer vision techniques capable of robustly identifying early signs of infestation across diverse scales and complex backgrounds. This study aimed to develop a robust deep learning model capable of reliably identifying these subtle infestation traces in multi-scale UAV images, thereby supporting precision agriculture applications. A novel detection architecture, termed Coordinated-BiFPN-P2-YOLO (CBP-YOLO), was developed based on YOLOv8. To address image degradation from low ground sampling distance, the Real-Enhanced Super-Resolution Generative Adversarial Network (Real-ESRGAN) was applied as a preprocessing step to reconstruct high-fidelity textures of leaf damage from original low-resolution inputs. The backbone of YOLOv8 was enhanced with the Coordinated Attention (CA) mechanism, which jointly captures spatial and channel-wise feature dependencies to improve localization and discrimination of minute lesions. Furthermore, the neck component was upgraded with a Bi-directional Feature Pyramid Network (BiFPN)to enable efficient top-down and bottom-up cross-scale feature fusion, minimizing information loss during hierarchical propagation and ensuring consistent representation across scales. In addition, a dedicated detection head operating at a 160×160 spatial resolution with 64-channel output was added to specifically strengthen sensitivity to small targets. The model was trained and evaluated on a custom UAV-collected dataset acquired over maize fields naturally infested by fall armyworm, encompassing diverse lighting conditions and flight heights. Extensive experiments demonstrated that CBP-YOLO achieved peak performance on imagery with a ground sampling distance (GSD)of 0.38 centimeters per pixel. The integration of Real-ESRGAN significantly alleviated texture blurring and edge ambiguity in low-resolution images, leading to clearer delineation of feeding scars. Ablation studies were conducted to evaluate the effectiveness of the proposed improvements, and the results demonstrate that the enhanced model achieves outstanding performance on the UAV multi-scale blade dataset. Specifically, the model attains an average precision (AP@0.5) of 76.5%, reflecting a notable increase of 3.4 percentage points compared to the baseline model. These findings indicate that the proposed modifications substantially enhance the model’s detection capability across blades of varying scales, highlighting the robustness and practical applicability of the improved approach in aerial inspection scenarios. In comparative evaluations, CBP-YOLO outperformed state-of-the-art detectors—including YOLOv9 medium, YOLOv10 medium, YOLOv11 medium, Faster Region-based Convolutional Neural Network, and RetinaNet—by margins of 10.1, 7.2, 5.1, 9.3, and 17.9 percentage points in AP@0.5, respectively. Notably, the model maintained high precision under varying illumination and partial occlusion, demonstrating strong generalization in real-world agricultural environments. The proposed CBP-YOLO framework effectively addresses the core challenges of detecting subtle, multi-scale fall armyworm infestation signs in UAV-based maize monitoring. By synergistically combining super-resolution enhancement, attention-aware feature extraction, fine-grained detection heads, and bidirectional multi-scale fusion, the model delivers superior accuracy and robustness. This approach provides a practical and scalable solution for early pest outbreak detection, enabling timely intervention and reducing crop losses in large-scale maize production systems.

     

/

返回文章
返回