基于深度学习的装配零部件识别与定位方法

An assembly part identification and localization method based on deep learning

  • 摘要: 工业4.0和中国制造强调智能化生产与装配,而在机床加工装配过程中存在装配零部件识别与定位精度低、效率低的问题。因此,基于轻量化网络、注意力和信息融合机制,提出一种零部件识别与定位算法LAI YOLOv5。首先,在YOLOv5网络结构中通过对卷积层进行轻量操作,有效解决神经网络参数量多、浮点运算量大、显存占用高和实时检测速度慢的问题;然后在主干网络中引入注意力机制,增加特征提取的倾向性,提升被检测物体的显著度;最后在特征融合网络中增加跨通道信息融合机制,增强特征检测能力。实验结果表明:与原始算法相比,在模型结构方面,改进后的LAI YOLOv5算法参数量和网络层数分别减少了约45.98%和28.46%,浮点运算量减少了约55.82%,显存占用降低了15.51%,训练时间缩短约32.27%。同时,训练精确度达到96.80%,训练覆盖率达到95.01%,实时检测效率提升至100.739 fps,检测准确率高达98.62%。

     

    Abstract: Industry 4.0 and Made in China emphasize intelligent production and assembly, and there is the problem of low accuracy and slow efficiency of assembly parts identification and localization in machine tool machining and assembly process. Therefore, based on the lightweight network, attention and information fusion mechanism, a component identification and localization algorithm LAI YOLOv5 is proposed. Firstly, in the YOLOv5 network structure, the problems of large number of neural network parameters, large floating-point operations, high graphics memory usage and slow real-time detection speed are effectively solved by lightweight operations of the convolutional layer. Then the attention mechanism is introduced into the backbone network to increase the tendency of feature extraction and enhance the saliency of the detected object. Finally, the feature fusion network is added with the cross channel information fusion mechanism to enhance the feature detection capability. The experimental results show that compared with the original algorithm, in terms of model structure, the number of parameters and network layers of the improved LAI YOLOv5 algorithm are reduced by about 45.98% and 28.46%, respectively, and the amount of floating-point operations is reduced by about 55.82%, while the graphics memory usage is reduced by 15.51% and the length of the training time is reduced by about 32.27%. Meanwhile, the training accuracy reaches 96.80%, the training coverage reaches 95.01%, and the real-time detection efficiency is improved to 100.739 fps with a detection accuracy of 98.62%.

     

/

返回文章
返回