系统仿真学报 ›› 2022, Vol. 34 ›› Issue (6): 1219-1229.doi: 10.16182/j.issn1004731x.joss.21-0907

• 仿真建模理论与方法 • 上一篇    下一篇

多尺度特征融合车辆检测方法

王银(), 王飞翔(), 孙前来   

  1. 太原科技大学 电子信息工程学院,山西 太原 030024
  • 收稿日期:2021-09-03 修回日期:2021-10-16 出版日期:2022-06-30 发布日期:2022-06-16
  • 通讯作者: 王飞翔 E-mail:xpw417@163.com;2601741160@qq.com
  • 作者简介:王银(1982-),男,博士,副教授,研究方向为计算机视觉、智能控制。E-mail:xpw417@163.com
  • 基金资助:
    国家自然科学基金(61905172);山西省科技成果转化引导专项(201904D131023);山西省重点研发计划(201903D121130);山西省面上青年基金(201901D211304);山西省研究生教育创新(2020SY422)

Vehicle Detection Method Based on Multi Scale Feature Fusion

Yin Wang(), Feixiang Wang(), Qianlai Sun   

  1. College of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan 030024, China
  • Received:2021-09-03 Revised:2021-10-16 Online:2022-06-30 Published:2022-06-16
  • Contact: Feixiang Wang E-mail:xpw417@163.com;2601741160@qq.com

摘要:

针对传统的车辆目标检测算法检测精度低,小尺度目标识别效果差等问题,提出了一种基于YOLOv4(you only look once v4)算法的目标检测方法,以提升对交通场景小目标车辆的检测性能。通过对YOLOv4网络进行再设计,使用MobileNetv2深度可分离卷积模块代替传统卷积,将CBAM(convolutional block attention module)注意力模块融合到特征提取网络中,在保证模型检测精度的同时减少模型参数。采用PANet-D特征融合网络融合获取到4个尺度特征图深浅层的语义信息,增强对小目标物体的检测能力。通过使用Focal loss优化分类损失函数,加快网络模型的收敛速度。实验结果表明,改进后的网络识别准确率达到96.55%,网络模型大小较原YOLOv4网络降低了92.49 M,同时检测速度比原网络提升了17%,充分证明了本算法的可行性。

关键词: 车辆检测, YOLOv4, 多尺度融合, 深度可分离卷积, 注意力机制

Abstract:

Vehicle detection is the important research content and hotspot in the intelligent transportation. Aiming at the low detection accuracy and poor small-scale recognition effect of the traditional vehicle detection algorithm, an improved detection method based on YOLOv4(you only look once v4) is proposed to improve the detection performance of small target vehicles in traffic scenes. By redesigning the YOLOv4 network, the MobileNetv2 deep separable convolution module is used to replace the traditional convolution, and the convolutional block attention module (CBAM) attention module is integrated into the feature extraction network to ensure the detection accuracy of the model and reduce the model parameters. The deep and shallow semantic information of the four scale feature maps is fused by using PANet-D feature fusion to enhance the detection ability of small objects. By using Focal loss to optimize the classification loss function, the convergence speed of the network model is accelerated. The experimental results show that the recognition accuracy of the improved network reaches 96.55%, and the size of the network model is 92.49 M lower, and the detection speed is 17% higher than those of the original YOLOv4 network, which fully proves the feasibility of the algorithm.

Key words: vehicle detection, YOLOv4, multiscale fusion, depth separable convolution, attention mechanism

中图分类号: