兵工学报 ›› 2022, Vol. 43 ›› Issue (10): 2687-2704.doi: 10.12382/bgxb.2021.0610
• 综述 • 上一篇
杨传栋, 钱立志, 薛松, 陈栋, 凌冲
上线日期:
2022-05-19
通讯作者:
钱立志(1963—),男,教授,博士生导师
E-mail:qianlizhi1@hotmail.com
作者简介:
杨传栋(1994—), 男, 博士研究生。 E-mail: yangcd19941030@163.com
基金资助:
YANG Chuandong, QIAN Lizhi, XUE Song, CHEN Dong, LING Chong
Online:
2022-05-19
摘要: 弹载图像目标检测方法是实现图像自寻的弹药“发射后不管”、对目标进行自主打击的关键技术。弹药图像自寻的面临着成像环境恶劣,目标特性变化快,对算法体积、速度要求苛刻等问题。围绕弹载目标检测难点问题进行综述,将基于深度学习的目标检测方法区分为基于候选框、无候选框和基于transformer的方法,回顾了各类方法主要研究进展;对特征提取网络轻量化、预测特征图增强、非极大值抑制后处理算法、训练中样本均衡、模型压缩等弹载图像目标检测模型部署中的关键技术进行了梳理;对比了典型目标检测方法在ImageNet、COCO及弹载图像目标数据集上的性能,并对未来发展进行展望。
中图分类号:
杨传栋, 钱立志, 薛松, 陈栋, 凌冲. 图像自寻的弹药目标检测方法综述[J]. 兵工学报, 2022, 43(10): 2687-2704.
YANG Chuandong, QIAN Lizhi, XUE Song, CHEN Dong, LING Chong. Review on Target Detection of Image Homing Ammunition[J]. Acta Armamentarii, 2022, 43(10): 2687-2704.
[1] 钱立志.电视末制导炮弹武器系统关键技术研究[D].合肥:中国科学技术大学,2006:2-6. QIAN L Z.Study on the key technique of the TV-guided projectile weapon system[D].Hefei: University of Science and Technology of China, 2006:2-6.(in Chinese) [2] 文苏丽, 宋怡然. DARPA导引头成本转换项目分析[J]. 战术导弹技术, 2016(2):5-9. WEN S L, SONG Y R.Research on seeker cost transfer program of DARPA[J].Tactical Missile Technology, 2016(2): 5-9.(in Chinese) [3] GIRSHICK R, DONABUE J, DARRELL T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Columbus, OH, US:IEEE, 2014: 580-587. [4] FELZENSZWALB P F, GIRSHICK R B, et al. Object detection with discriminatively trained part-based models[J].IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010,32(9): 1627-1645. [5] 徐英, 谷雨, 彭冬亮,等. 面向合成孔径雷达图像任意方向舰船检测的改进YOLOv3模型[J]. 兵工学报, 2021, 42(8): 1698-1707. XU Y, GU Y, PENG D L,et al. An improved YOLOv3 model for arbitrary-oriented ship detection in SAR image[J]. Acta Armamentarii, 2021, 42(8):1698-1707. (in Chinese) [6] 杨传栋, 刘桢, 石胜斌. 基于CNN的弹载图像目标检测方法研究[J]. 战术导弹技术, 2019(4): 85-92. YANG C D, LIU Z, SHI S B. Research on object detection for missile-borne image based on CNN[J]. Tactical Missile Technology, 2019(4): 85-92.(in Chinese) [7] 雷鸣, 王曙光, 凌冲, 等. 基于模型压缩YOLOv4的弹载图像舰船目标实时检测[J].兵器装备工程学报,2021,42(9):225- 230. LEI M, WANG S G, LING C, et al. Real-time detection of ship target in missile-borne platform images based on model compressed YOLOv4[J]. Journal of Ordnance Equipment Engineering, 2021, 42(9): 225-230. (in Chinese) [8] 陈栋,田宗浩.面向深度学习的弹载图像处理异构加速现状分析[J]. 航空兵器,2021,28(3):10-17. CHEN D, TIAN Z H.Research on heterogeneous acceleration of deep learning method for missile-borne image processing [J]. Aero Weaponry, 2021, 28(3): 10-17. (in Chinese) [9] 侯凯强, 李俊山, 王雪博, 等. 弹载人工智能目标识别算法的嵌入式实现方法研究[J]. 制导与引信, 2019, 40(3): 40-45. HOU K Q, LI J S, WANG X B, et al. Research on embedded implementation method of missle-borne artifical intelligence target recognition algorithms[J]. Guidance & Fuze, 2019, 40(3): 40- 45.(in Chinese) [10] 钱立志, 王曙光, 张江辉,等. 一种弹载视频图像实时消旋方法[J]. 弹箭与制导学报, 2009,29(3):26-28. QIAN L Z, WANG S G, ZHANG J H, et al. A real-time despinning method for onboard video image[J]. Journal of Projectiles, Rockets, Missiles and Guidance, 2009,29(3):26-28.(in Chinese) [11] REN S, HE K, GIRSHICK R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[J]. IEEE Transactions on Pattern Analysis & Machine Intelligence, 2015, 39(6):1137-1149. [12] LIU W, ANGUELOV D, ERHAN D, et al. SSD: single shot multibox detector[C]∥Proceedings of the 14th European Conference on Computer Vision. Amsterdam, Netherlands: Springer, 2016:21-37. [13] REDMON J, FARHADI A. YOLO9000: better, faster, stronger[C]]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI,US: IEEE, 2017: 6517-6525. [14] REDMON J, FARHADI A.Yolov3: an incremental improvement [EB/OL].[2018-04-08].https:∥arxiv.org/pdf/1804.02767. pdf. [15] BOCHKOVSKIY A, WANG C Y, LIAO H. YOLOv4: optimal speed and accuracy of object detection[EB/OL]. [2020-04-23].https:∥arxiv.org/pdf/2004.10934.pdf. [16] LIU S, QI L, QIN H, et al. Path aggregation network for instance segmentation[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City,VT,US: IEEE, 2018:8759-8768. [17] LIN T Y, GOYAL P, GIRSHICK R, et al. Focal loss for dense object detection[J].IEEE Transactions on Pattern Analysis & Machine Intelligence, 2017, 99: 2999-3007. [18] REDMON J, DIVVALA S, GIRSHICK R, et al. You only look once:unified, real-time object detection[C]∥/Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, CA,US: IEEE, 2016: 779-788. [19] HUANG L, YANG Y, DDENG Y, et al.DenseBox: unifying landmark localization with end to end object detection[EB/OL].[2015-09-16]. https:∥arxiv.org/pdf/1509. 04874v1.pdf. [20] TIAN Z, SHEN C, CHEN H, et al. FCOS: fully convolutional one-stage object detection[EB/OL].[2018-01-18]. https:∥ arxiv.org/abs/1904. 01355.pdf. [21] ZHOU X, WANG D, KRHENBHL P. Objects as points[EB/OL]. [2019-04-25].https:∥arxiv.org/pdf/1904.07850.pdf. [22] LAW H, DENG J. CornerNet: detecting objects as paired keypoints[C]∥Proceedings of the 15th European Conference on Computer Vision. Munich, Germany: Springer, 2018: 765-781. [23] DUAN K, BAI S, XIE L, et al. CenterNet: keypoint triplets for object detection[C]∥Proceeding of IEEE International Conference on Computer Vision. Seoul, Korea: IEEE, 2019:6568-6577. [24] DONG Z, LI G, LIAO Y, et al. CentripetalNet: pursuing high-quality keypoint pairs for object detection[C]∥ Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seattle, WA,US: IEEE, 2020: 10516- 10525. [25] HEI L, YUN T, OLGA R, et al. CornerNet-Lite: efficient keypoint based object detection[EB/OL].[2019-04-18]. https:∥arxiv.org/abs/1904. 08900.pdf. [26] ZHOU X, ZHUO J,KRHENBHL P. Bottom-Up object detection by grouping extreme and center points[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.Long Beach, CA,US: IEEE, 2019: 850-859. [27] ARION N, MASSA F, SYNNAEVE G, et al. End-to-End object detection with transformers[C]∥Proceedings of the 16th European Conference on Computer Vision.Cham, Switzerland: Springer, 2020: 213-229. [28] FANG Y, LIAO B, WANG X, et al. You only look at one sequence: rethinking transformer in vision through object detection[EB/OL].[2021-06-01].https:∥arxiv.org/pdf/2106. 00666.pdf. [29] DAI X, CHEN Y, XIAO B, et al. Dynamic head: unifying object detection heads with attentions[EB/OL]. [2021-06-15].https:∥arxiv.org/pdf/2106.08322.pdf. [30] LIU Z, LIN Y, CAO Y, et al. Swin transformer: hierarchical vision transformer using shifted windows[EB/OL]. [2021-05-25].https:∥arxiv.org/pdf/2103. 14030v1.pdf. [31] KRIZHEVSKY A, SUTSKEVER I, HINTON G. ImageNet classification with deep convolutional neural networks[C]∥Proceedings of Advances in Neural Information Processing Systems. Lake Tahoe,CA, US: NIPS, 2012:1106-1114. [32] SIMONYAN K, ZISSERMAN A. Very deep convolutional networks for large-scale image recognition[J]. Computer Science, 2014, 4(1): 1-14. [33] SZEGEDY C, VANHOUCKE V, IOFFE S, et al. Rethinking the inception architecture for computer vision[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV,US: IEEE, 2016: 2818-2826. [34] HOWARD A, ZHU M, CHEN B, et al. MobileNets: efficient convolutional neural networks for mobile vision applications[EB/OL].[2017-04-17].https:∥arxiv.org/pdf/ 1704.04861.pdf. [35] HE K, ZHANG X, REN S, et al. Deep residual learning for image recognition[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, CA,US: IEEE, 2016: 770-778. [36] XIE S, GIRSHICK R, DDLLáR P, et al.Aggregated residual transformations for deep neural networks[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI,US: IEEE, 2017: 5987-5995. [37] HUANG G, LIU Z, LAURENS V, et al. Densely connected convolutional networks[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI,US: IEEE, 2017: 2261-2269. [38] HUANG G, LIU S, VAN D, et al. CondenseNet: An efficient DenseNet using learned group convolutions[C]∥ Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City, CA,US: IEEE, 2018:2752-2761. [39] LEE Y, PARK J. CenterMask: real-time anchor-free instance segmentation[EB/OL].[2020-05-14].https:∥arxiv.org/abs/ 1911.06667.pdf. [40] YANG L, JIANG H, CAI R, et al. CondenseNet V2: sparse feature reactivation for deep networks[EB/OL].[2021-04-09] https:∥arxiv.org/pdf/2104.04382.pdf. [41] HU J, SHEN L, SUN G, et al. Squeeze-and-excitation networks[J].IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020, 42(8): 2011-2023. [42] WOO S, PARK J, LEE J Y, et al. CBAM: convolutional block attention module[C]∥Proceedings of the 15th European Conference on Computer Vision. Munich, Germany: Springer, 2018:3-19. [43] SELVARJU R, COGSWELL M, DDS A, et al. Grad-CAM: visual explanations from deep networks via gradient-based localization[J]. International Journal of Computer Vision, 2020, 128(2):336-359. [44] LANDOLA F N, HAN S, MOSKEWICZ M W, et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and<0.5 MB model size[EB/OL].(2016-11-04)[2021-05-31]. https:∥arxiv. org/abs/1602. 07360.pdf. [45] CHOLLET F. Xception:deep learning with depthwise separable convolutions[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI,US: IEEE, 2017: 1800-1807. [46] SANDLER M, HOWARD A, ZHU M, et al. Mobilenetv2: Inverted residuals and linear bottlenecks[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City, VT,US: IEEE, 2018: 4510-4520. [47] ZHANG X Y,ZHOU X Y,LIN M. ShuffleNet: an extremely efficient convolutional neural network for mobile devices[C]∥Proceedings of Conference on Computer Vision and Pattern Recognition. Salt Lake City, CA,US: IEEE, 2018: 6848-6856. [48] MA N, ZHANG X, ZHENG H, et al. ShuffleNet V2: practical guidelines for efficient CNN architecture design[C]∥Proceedings of the 15th European Conference on Computer Vision. Munich, Germany: Springer, 2018:122-138. [49] HAN K, WANG Y, TIAN Q, et al. GhostNet: more features from cheap operations[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seattle,WA,US: IEEE, 2020:1577-1586. [50] HOWARD A, SANDLER M, CHU G, et al. Searching for mobilenetv3[C]∥Proceedings of IEEE International Conference on Computer Vision. Seoul, Korea: IEEE, 2019: 1314-1324. [51] LI Y, CHEN Y, DAI X, et al. MicroNet: improving image recognition with extremely low FLOPs[EB/OL]. [2020-11-24].https:∥arxiv.org/pdf/2108.05894.pdf. [52] ZOPH B, VASUDEVAN V, SHLENS J, et al. Learning transferable architectures for scalable image recognition[C]∥Proceeding of Conference on Computer Vision and Pattern Recognition. Salt Lake City, CA,US: IEEE, 2018: 8697-8710. [53] WU B, DAI X, ZHANG P, et al. FBNet: hardware-aware efficient convnet design via differentiable neural architecture search[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seoul, Korea: IEEE, 2019: 10726-10734. [54] WAN A, DAI X, ZHANG P, et al. FBNetV2: differentiable neural architecture search for spatial and channel dimensions[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seattle, WA,US: IEEE, 2020:12962-12971. [55] DAI X, WAN A, ZHANG P, et al. FBNetV3: joint architecture-recipe search using neural acquisition function [EB/OL].[2021-03-30].https:∥arxiv.org/pdf/2006.02049.pdf. [56] LIN T Y, DOLLáR P, GIRSHICK R, et al. Feature pyramid networks for object detection[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI,US: IEEE, 2017:936-944. [57] TAN M, PANG R, LE Q. EfficientDet: scalable and efficient object detection[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seattle,WA, US: IEEE, 2020: 10778-10787. [58] GHIASI G, LIN T Y, PANG R, et al.NAS-FPN: learning scalable feature pyramid architecture for object detection[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seoul, Korea: IEEE, 2019: 7036-7045. [59] LIU S, HUANG D, WANG Y. Learning spatial fusion for single-shot object detection[EB/OL].[2019-11-21].https:∥ arxiv.org/pdf/1911.09516.pdf. [60] LI Q, LIN Y, HE W. SSD7-FFAM: a real-time object detection network friendly to embedded devices from scratch[J]. Applied Sciences, 2021, 11(3):1096. [61] LIU S, HUANG D, WANG Y. receptive field block net for accurate and fast object detection[C]∥Proceedings of the 15th European Conference on Computer Vision. Munich, Germany: Springer, 2018:404-419. [62] LI Y, CHEN Y, WANG N, et al. Scale-aware trident networks for object detection[C]∥Proceedings of IEEE International Conference on Computer Vision. Seoul, Korea: IEEE, 2019: 6053-6062. [63] CAO J, CHEN Q, GUO J, et al. Attention-guided context feature pyramid network for object detection[EB/OL]. [2020-04-06].https:∥arxiv.org/pdf/2005.11475.pdf. [64] CHEN Q, WANG Y, YANG T, et al. You only look one-level feature[EB/OL].[2021-05-17].https:∥arxiv.org/pdf/2103. 09460.pdf. [65] SHRIVASTAVA A, GUPTA A, GIRSHICK R. Training region-based object detectors with online hard example mining[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, CA,US:IEEE, 2016:761-769. [66] CAO Y, CHEN K, LOY C C, et al.Prime sample attention in object detection[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Long Beach,CA, US:IEEE, 2019: 11580-11588. [67] ZHANG S, CHI C, YAO Y, et al. Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Seattle, WA,US:IEEE, 2020:9756-9765. [68] LI B, LIU Y, WANG X.Gradient harmonized single-stage detector[C]∥Proceedings of AAAI Conference on Artificial Intelligence. Honolulu,HI,US: IEEE, 2019: 8577-8584. [69] LI X, WANG W, WU L, et al. Generalized focal loss: learning qualified and distributed bounding boxes for dense object detection[C]∥Proceedings of Advances in Neural Information Processing Systems. US: NIPS, 2020: 21002-21012. [70] CHEN K, LI J, LIN W, et al. Towards accurate one-stage object detection with AP-loss[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Long Beach, CA,US:IEEE, 2019: 5119-5127. [71] CHEN J, LIU D, LUO B, et al. Residual objectness for imbalance reduction[EB/OL].[2019-08-24].https:∥arxiv. org/pdf/1908.09075.pdf. [72] BODLA N, SINGH B, CHELLAPPA R, et al. Soft-NMS-improving object detection with one line of code[C]∥Proceedings of IEEE International Conference on Computer Vision. Venice, Italy: IEEE, 2017:5562-5570. [73] JIANG B, LUO R, MAO J, et al. Acquisition of localization confidence for accurate object detection[C]∥Proceedings of the 15th European Conference on Computer Vision. Munich, Germany: Springer, 2018:816-832. [74] WU S, LI X, WANG X. IoU-aware single-stage object detector for accurate localization[J]. Image and Vision Computing, 2020, 97: 103911. [75] LIU S, HUANG D, WANG Y. Adaptive NMS: refining pedestrian detection in a crowd[C]∥Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Long Beach, CA,US: IEEE, 2019, 6452-6461. [76] ZHENG Z, WANG P, LIU W, et al. Distance-IoU loss: faster and better learning for bounding box regression[C]∥ Proceedings of AAAI Conference on Artificial Intelligence. New York, NY,US: AAAI, 2020. [77] BOLYA D, ZHOU C, XIAO F, et al. YOLACT: real-time instance segmentation[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019,99:1-1. [78] ZHENG Z, WANG P, REN D, et al. Enhancing geometric factors in model learning and inference for object detection and instance segmentation[EB/OL].[2020-05-07].https:∥arxiv. org/pdf/2005.03572v3.pdf. [79] HE Y, LIN J, LIU Z, et al. AMC: autoML for model compression and acceleration on mobile devices[EB/OL]. [2020-04-06].https:∥arxiv.org/pdf/1802.03494.pdf. [80] HOROWITZ M.Computing's energy problem (and what we can do about it)[C]∥Proceedings of IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC). San Francisco, CA,US: IEEE, 2014:10-14. |
[1] | 李博, 王博, 韩京冶, 杨宗睿, 栗霖雲, 傅畅之. 基于车载计算机的红外图像移动目标检测[J]. 兵工学报, 2022, 43(S1): 66-73. |
[2] | 李良福, 陈卫东, 高强, 许开銮, 刘轩, 何曦, 钱钧. 基于深度学习的光电系统智能目标识别[J]. 兵工学报, 2022, 43(S1): 162-168. |
[3] | 严继伟, 苏娟, 李义红. 基于[KG-*9]Ghost卷积与注意力机制的SAR图像建筑物检测算法[J]. 兵工学报, 2022, 43(7): 1667-1675. |
[4] | 王少博, 张成, 苏迪, 冀瑞静. 基于改进YOLOv3和核相关滤波算法的旋转弹目标探测算法[J]. 兵工学报, 2022, 43(5): 1032-1045. |
[5] | 李锦青, 刘泽飞, 满振龙. 基于生成对抗网络的密钥生成方法及其在微光图像加密中的应用[J]. 兵工学报, 2022, 43(2): 337-344. |
[6] | 于博文, 吕明. 改进的YOLOv3算法及其在军事目标检测中的应用[J]. 兵工学报, 2022, 43(2): 345-354. |
[7] | 吴超, 吴绍斌, 李子睿, 孙冬. 基于人机交互的免锚检测和跟踪系统设计[J]. 兵工学报, 2022, 43(10): 2565-2575. |
[8] | 龚诗雄, 王旭, 孔国杰, 龚建伟. 多车协同目标跟踪方法[J]. 兵工学报, 2022, 43(10): 2429-2442. |
[9] | 普运伟, 刘涛涛, 郭江, 吴海潇. 基于卷积神经网络和模糊函数主脊坐标变换的雷达辐射源信号识别[J]. 兵工学报, 2021, 42(8): 1680-1689. |
[10] | 徐英, 谷雨, 彭冬亮, 刘俊, 陈华杰. 面向合成孔径雷达图像任意方向舰船检测的改进YOLOv3模型[J]. 兵工学报, 2021, 42(8): 1698-1707. |
[11] | 张永梅, 赖裕平, 马健喆, 冯超, 束颉. 基于视频的装甲车和飞机检测跟踪及轨迹预测算法[J]. 兵工学报, 2021, 42(3): 545-554. |
[12] | 朱敏, 许爱强, 许晴, 李睿峰. 基于改进多层核超限学习机的模拟电路故障诊断[J]. 兵工学报, 2021, 42(2): 356-369. |
[13] | 马月红, 孔梦瑶. 基于改进快速区域卷积网络的目标检测轻量化算法[J]. 兵工学报, 2021, 42(12): 2664-2674. |
[14] | 张凤, 张超, 杨华民, 王发斌. 作战辅助系统中基于深度学习的投影图像颜色补偿方法[J]. 兵工学报, 2021, 42(11): 2418-2423. |
[15] | 薛灵芝, 曾向阳, 杨爽. 基于生成对抗网络的水声目标识别算法[J]. 兵工学报, 2021, 42(11): 2444-2452. |
阅读次数 | ||||||
全文 |
|
|||||
摘要 |
|
|||||