高级检索+

基于YOLOv8n-dy的玫瑰花花期多尺度目标检测方法

Multi-scale object detection method for rose flowering stages based on YOLOv8n-dy

  • 摘要: 为应对玫瑰采摘劳动强度大、采花农户老龄化和人工成本上升等挑战,针对玫瑰花采摘过程中检测识别准确性不足的问题,提出一种基于YOLOv8n模型的玫瑰花花期检测模型(YOLOv8n-dy),实现对花苞期、盛花期和败花期的精确判别。在主干网络中优化C2f模块,构建高效的多分支特征提取卷积结构,提高模型对多尺度、多形态花期目标的识别能力;引入ELA高效注意力机制(efficient lightweight attention),增添小目标检测层并将损失函数替换为WIOUv3损失函数(weighted intersection over union version 3)强化深层网络对花苞期玫瑰花等小尺寸目标的定位检测能力;采用Adamax优化器(adaptive maximum optimization)解决模型陷入局部最优解问题。试验结果表明,YOLOv8n-dy模型的准确率、召回率和在50%阈值下(IoU=50%)平均精度均值mAP0.5分别达到79.2%、71.3%和77.3%,较原始YOLOv8n模型分别提升4.4个百分点、6.7个百分点和6.1个百分点。同检测精度最高的YOLOv9比较,YOLOv8n-dy模型计算复杂度仅为9.8 GFLOPs,同时保持了与YOLOv9相当的50%交并比阈值下平均精度均值77.3%,该模型的大小仅为YOLOv9模型的4.22%,显著提高模型部署的应用性能。该研究提出的YOLOv8n-dy模型在复杂田间环境下表现出优异性能,能高效准确地识别玫瑰花的花苞期(bud)、盛花期(flower)和败花期(withered),同时实现检测精度和计算效率的良好平衡,为玫瑰花采摘提供可靠的技术支持。

     

    Abstract: Rose is one of most favorite flowers in modern floriculture worldwide. However, significant challenges have been posed to the rose cultivation in China. Particularly, manual harvesting cannot fully meet the large-scale intensive production in recent years, due mainly to the increasing labor costs. It is still lacking in accurate and efficient machine vision to reliably identify the rose flowering stages. In this study, an accurate model was proposed to detect the rose flowering stage using the YOLOv8n framework, designated as YOLOv8n-dy. The backbone network was optimized to enhance the C2f module. A more efficient multi-branch convolutional block was constructed to improve the recognition for the multi-scale and multi-morphology flowering targets. The Efficient Local Attention (ELA) mechanism was introduced to detect the small targets. An additional layer was also added for the detection of the small target. The loss function was replaced with the Weighted Intersection over Union version 3 (WIOUv3) to accurately locate and detect the small-sized targets, such as the bud-stage roses. The Adamax Optimizer 5 was employed to avoid the local optima. Experimental results demonstrated that the YOLOv8n-dy model achieved significant improvements over the original YOLOv8n model, with gains of 4.4 percentage points in precision, 6.7 percentage points in recall, and 6.1 percentage points in mAP0.5, respectively. A comparison was made of the state-of-the-art detectors. The YOLOv8n-dy model also maintained a highly competitive mAP0.5 value of 77.3%. While the YOLOv9 achieved the highest accuracy in raw detection. The computational footprint was dramatically reduced to require only 9.8 Giga Floating Point Operations (GFLOPs). Additionally, the weight of the YOLOv8n-dy model was only 4.22% of that of the YOLOv9 model, significantly enhancing the deployment and application performance. This exceptional efficiency was also observed in the field. All three flowering stages were consistently and accurately identified under diverse and challenging field conditions, including the varying lighting and occlusions. In conclusion, the YOLOv8n-dy model was optimized to balance between high detection accuracy and low computational complexity. The strategic architecture greatly contributed to the robust performance, including the optimal blocks of the feature extraction, attention mechanism, and advanced loss function. The improved model with markedly low computation can be real-time deployed, suitable for the embedded and mobile platforms, which are very critical for practical applications. Consequently, this finding can provide a reliable, efficient, and scalable technological solution for rose harvesting, in order to reduce the labor shortages and operational costs in sustainable and precision agriculture.

     

/

返回文章
返回