论文:2022,Vol:40,Issue(2):433-441
引用本文:
张艳, 李星汕, 孙叶美, 刘树东. 基于通道注意力与特征融合的水下目标检测算法[J]. 西北工业大学学报
ZHANG Yan, LI Xingshan, SUN Yemei, LIU Shudong. Underwater object detection algorithm based on channel attention and feature fusion[J]. Northwestern polytechnical university

基于通道注意力与特征融合的水下目标检测算法
张艳, 李星汕, 孙叶美, 刘树东
天津城建大学 计算机与信息工程学院, 天津 300384
摘要:
水下光学图像存在色偏、低对比度、目标模糊的现象,导致水下目标检测时存在漏检、误检等问题。针对上述问题,提出了一种基于通道注意力与特征融合的水下目标检测算法。基于通道注意力设计了激励残差模块,将前向传播的特征信息进行自适应分配权重,以突出不同通道特征图的显著性,提高了网络对水下图像高频信息的提取能力;设计了多尺度特征融合模块,增加了大尺度特征图用于目标检测,利用其对应的小尺度感受野提高了网络对小尺寸目标的检测性能,进一步提高了网络对水下不同尺寸目标的检测精度;为提高网络对水下环境的泛化性能,设计了基于拼接和融合的数据增强方法,模拟水下目标的重叠、遮挡和模糊情况,增强了网络对水下环境的适应性。通过在公共数据集URPC上的实验,与YOLOv3、YOLOv4和YOLOv5相比,所提算法的平均精度均值分别提升5.42%,3.20%和0.9%,有效改善了水下复杂环境中不同尺寸目标漏检、误检的问题。
关键词:    水下图像    目标检测    通道注意力    多尺度特征融合    深度学习   
Underwater object detection algorithm based on channel attention and feature fusion
ZHANG Yan, LI Xingshan, SUN Yemei, LIU Shudong
School of Computer and Information Engineering, Tianjin Chengjian University, Tianjin 300384, China
Abstract:
Due to the color deviation, low contrast and fuzzy object in underwater optical images, there are some problems in underwater object detection, such as missed detection and false detection. In order to solve the above-mentioned problems, an underwater object detection algorithm is proposed based on the channel attention and feature fusion for underwater optical images. The excitation residual module is designed based on the channel attention, and the forward propagation feature information is adaptively allocated weights to highlight the salience of different channel feature maps, which improves the network ability to extract high-frequency information from the underwater images. The multi-scale feature fusion module is designed to add a large scale feature map for object detection, which improves the detection performance of the network for small size objects by using its corresponding small size receptive field, and further improves the detection accuracy of the network for different size objects in the underwater environment. To improve the generalization performance of the network to the underwater environment, the data augmentation method based on the stitching and fusion is designed to simulate the overlap, occlusion and blurring of underwater objects, which improves the adaptability of the network to the underwater environment. Through experiments on the public dataset URPC, the algorithm in this paper improves the mean average precision by 5.42%, 3.20% and 0.9% compared with YOLOv3, YOLOv4 and YOLOv5, respectively, effectively improving the missed and false detection of objects of different sizes in complex underwater environments.
Key words:    underwater image    object detection    channel attention    multi-scale feature fusion    deep learning   
收稿日期: 2021-07-09     修回日期:
DOI: 10.1051/jnwpu/20224020433
基金项目: 国家自然科学基金(61902273)与天津市科技计划项目(20YDTPJC01310)资助
通讯作者: 孙叶美(1990-),天津城建大学助理实验师,主要从事图像处理和人工智能研究。e-mail:sunyemei1216@163.com     Email:sunyemei1216@163.com
作者简介: 张艳(1982-),天津城建大学副教授,主要从事图像处理和机器视觉研究。
相关功能
PDF(2975KB) Free
打印本文
把本文推荐给朋友
作者相关文章
张艳  在本刊中的所有文章
李星汕  在本刊中的所有文章
孙叶美  在本刊中的所有文章
刘树东  在本刊中的所有文章

参考文献:
[1] GIRSHICK R, DONAHUE J, DARRELL T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014
[2] GIRSHICK R. Fast R-CNN[C]//Proceedings of the IEEE International Conference on Computer Vision, 2015
[3] REN S, HE K, GIRSHICK R, et al. Faster R-CNN:towards real-time object detection with region proposal networks[J]. Advances in Neural Information Processing Systems, 2015, 28:91-99
[4] HE K, GKIOXARI G, DOLLÁR P, et al. Mask R-CNN[C]//Proceedings of the IEEE International Conference on Computer Vision, 2017
[5] REDMON J, DIVVALA S, GIRSHICK R, et al. You only look once:unified, real-time object detection[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016
[6] REDMON J, FARHADI A. YOLO9000:better, faster, stronger[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017
[7] REDMON J, FARHADI A. Yolov3:an incremental improvement[EB/OL]. (2018-04-08)[2021-07-09]. https://arxiv.org/abs/1804.02767.
[8] LIU W, ANGUELOV D, ERHAN D, et al. SSD:single shot multibox detector[C]//European Conference on Computer Vision, 2016
[9] 赵德安, 刘晓洋, 孙月平, 等. 基于机器视觉的水下河蟹识别方法[J]. 农业机械学报, 2019, 50(3):151-158 ZHAO Dean, LIU Xiaoyang, SUN Yueping,et al. Underwater crab recognition method based on machine vision[J]. Transactions of the Chinese Society for Agricultural Machinery, 2019, 50(3):151-158 (in Chinese)
[10] 高英杰. 基于"SSD"模型的水下目标检测网络设计与实现[J]. 电子世界, 2019, 566(8):112-113 GAO Yingjie. Design and implementation of underwater target detection network based on "SSD" model[J]. Electronic World, 2019, 566(8):112-113 (in Chinese)
[11] 刘萍, 杨鸿波, 宋阳. 改进YOLOv3网络的海洋生物识别算法[J]. 计算机应用研究, 2020, 37(增刊1):394-397 LIU Ping, YANG Hongbo, SONG Yang. Marine biometric identification algorithm based on improved YOLOv3 network[J]. Computer Application Research, 2020, 37(suppl 1):394-397 (in Chinese)
[12] JADERBERG M, SIMONYAN K, ZISSERMAN A. Spatial transformer networks[J]. Advances in Neural Information Processing Systems, 2015, 28:2017-2025
[13] WOO S, PARK J, LEE J Y, et al. CBAM:convolutional block attention module[C]//Proceedings of the European Conference on Computer Vision, 2018
[14] HU J, SHEN L, SUN G. Squeeze-and-excitation networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018
[15] ZHANG H, CISSE M, DAUPHIN Y N, et al. Mixup:beyond empirical risk minimization[EB/OL]. (2018-04-27)[2021-07-09]. https://arxiv.org/abs/1710.09412
[16] YUN S, HAN D, OH S J, et al. Cutmix:regularization strategy to train strong classifiers with localizable features[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019
[17] LIN W H, ZHONG J X, LIU S, et al. Roimix:proposal-fusion among multiple images for underwater object detection[C]//IEEE International Conference on Acoustics, Speech and Signal Processing, 2020
[18] BOCHKOVSKIY A, WANG C Y, LIAO H Y M. Yolov4:optimal speed and accuracy of object detection[EB/OL]. (2020-04-23)[2021-07-09]. https://arxiv.org/abs/2004.10934
[19] GLENN Jocher. YOLOv5. Github Repository[EB/OL]. (2020-06-09)[2021-07-09]. https://github.com/ultralytics/yolov5
相关文献:
1.杨振舰, 尚佳美, 张众维, 张艳, 刘树东.基于残差注意力机制的图像去雾算法[J]. 西北工业大学学报, 2021,39(4): 901-908