高级检索+

基于HCM-UNet的复杂环境下花生地表残膜检测方法

Detection of peanut surface residual film in complex environments based on HCM-UNet

  • 摘要: 针对现有残膜分割方法精度低、检测速度慢的问题,该研究提出一种基于HCM-UNet (haar-wavelet cbam multi-scale attention aggregation UNet)的花生地表残膜分割方法。以UNet模型为基础,将MobileNetV3引入UNet模型中作为轻量级骨干特征提取网络,同时为减小模型轻量化后精度下降的影响,采用基于Haar小波的下采样模块(haar wavelet downsampling,HWD)减少下采样过程中丢失的细节;其次,为改善模型对于小目标残膜的检测效果,引入MSAA(multi-scale attention aggregation)架构融合残膜多尺度特征;最后,在解码器中引入CBAM(convolutional block attention module),增强模型对残膜边缘的关注度,进一步提升模型的分割精度。试验结果表明:HCM-UNet模型在花生地表残膜图像测试集上的平均交并比为85.72%,平均像素准确率为84.26%,F1分数为83.68%,模型参数量为43.22 M,每幅图像的平均分割时间为127.17 ms,均优于Deeplabv3、PSPNet、UNet、Segformer、Mask2former和HRNet主流分割模型。该模型在轻量化的基础上提高了花生地表残膜的分割精度,在不同作业时期和光照条件下均表现出良好的稳定性和鲁棒性,可为评估花生地残膜污染情况提供技术支撑。

     

    Abstract: Aiming at the problems of low accuracy and slow detection speed of existing residual film segmentation methods, a peanut surface residual film segmentation method based on Haar-Wavelet CBAM Multi-Scale Attention Aggregation UNet (HCM-UNet) is proposed. Based on the UNet model, MobileNetV3 is introduced into the UNet model as a lightweight backbone feature extraction network, and in order to minimize the effect of accuracy degradation after the model is lightweighted, the Haar Wavelet Downsampling (HWD) module is used to reduce the details lost in the downsampling process; secondly, to improve the detection effect of the model for small target residual film, MSAA (Multi-subsampling) is introduced to improve the detection effect of the model for small target residual film. model for small target residual film detection, Multi-Scale Attention Aggregation (MSAA) architecture is introduced to fuse the multi-scale features of residual film; finally, Convolutional Block Attention Module (CBAM) is introduced in the decoder to enhance the model's attention to the edges of the residual film, which further improve the segmentation accuracy of the model. The experimental results show that the HCM-UNet model has an average intersection and merger ratio of 85.72%, an average pixel accuracy of 84.26%, an F1 score of 83.68%, a model parameter count of 43.22 M, and an average segmentation time of 127.17 ms per image on a test set of peanut surface residual film images, all of which outperform Deeplabv3, PSPNet, and UNet, Segformer, Mask2former and HRNet mainstream segmentation models. The model improves the segmentation accuracy of peanut ground residual film on the basis of lightweight, and shows good stability and robustness under different operating periods and light conditions, providing data support for evaluating the contamination of peanut ground residual film.

     

/

返回文章
返回