A Review of Automatic Driving Target Detection Based on Camera and Millimeter Wave Radar Fusion Technology
DOI:
https://doi.org/10.15282/ijame.22.1.2025.3.0920Keywords:
Sensor fusion, Target detection, Automatic driving, Camera, MMW radarAbstract
Autonomous driving relies heavily on precise target detection to ensure safety and efficiency in navigating complex environments. It typically utilizes multiple sensors to achieve comprehensive environmental perception. This review explores advancements in integrating these complementary sensors, focusing on state-of-the-art fusion methods, challenges, and applications. The combination of these sensors addresses the limitations of individual modalities: cameras excel in capturing detailed textures and colors, while millimeter wave radar provides reliable distance, velocity, and motion information under adverse weather conditions. Key findings reveal that the sparse radar data, lack of comprehensive multimodal datasets, and difficulties in correlating radar with image data pose significant hurdles. Future research should focus on developing comprehensive multimodal datasets, 4D millimeter-wave radar, and refining fusion algorithms for robustness in diverse environments. This review provides a comprehensive understanding of the current state and challenges in target detection, serving as a foundation for future innovation in autonomous driving technology.
References
[1] Y. Huang, J. Du, Z. Yang, Z. Zhou, L. Zhang, and H. Chen, “A survey on trajectory-prediction methods for autonomous driving,” IEEE Transactions on Intelligent Vehicles, vol. 7, no. 3, pp. 652–674, 2022.
[2] S. Teng, X. Hu, P. Deng, B. Li, Y. Li, Y. Ai et al., “Motion planning for autonomous driving: The state of the art and future perspectives,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 6, pp. 3692–3711, 2023.
[3] D. J. Yeong, G. Velasco-Hernandez, J. Barry, and J. Walsh, “Sensor and sensor fusion technology in autonomous vehicles: A review,” Sensors, vol. 21, no. 6, p. 2140, 2021.
[4] D. Feng, C. Haase-Schütz, L. Rosenbaum, et al., “Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets, methods, and challenges,” IEEE Transactions on Intelligent Transportation Systems, vol. 22, no. 3, pp. 1341–1360, 2021.
[5] M. Liang, B. Yang, Y. Chen, R. Hu, and R. Urtasun, “Multi-task multi-sensor fusion for 3D object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA: IEEE, pp. 7345-7353. 2019.
[6] H. Caesar, V. Bankiti, A. H. Lang, S. Vora, V. E. Liong, Q. Xu et al., “nuScenes: A multimodal dataset for autonomous driving,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA: IEEE, pp. 11621-11631, 2020.
[7] H. Liu, C. Wu, and H. Wang, “Real time object detection using Lidar and camera fusion for autonomous driving,” Scientific Reports, vol. 13, no. 1, p. 8056, 2023.
[8] X. Chen, H. Ma, J. Wan, B. Li, and T. Xia, “Multi-view 3D object detection network for autonomous driving,” in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI: IEEE pp. 1907-1915, 2017.
[9] J. Ku, M. Mozifian, J. Lee, A. Harakeh, and S. L. Waslander, “Joint 3D proposal generation and object detection from view aggregation,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, Madrid: IEEE, pp. 1-8, 2018.
[10] B. Li, T. Zhang, and T. Xia, “Vehicle detection from 3D lidar using fully convolutional network.” arXiv, Aug. 29, 2016. Accessed: Jul. 23, 2023. [Online]. Available: http://arxiv.org/abs/1608.07916
[11] R. Nabati and H. Qi, “RRPN: Radar Region Proposal Network for object detection in autonomous vehicles,” in 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan: IEEE, pp. 3093–3097, 2019.
[12] P. Li, P. Wang, K. Berntorp, and H. Liu, “Exploiting temporal relations on radar perception for autonomous driving,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA: IEEE, pp. 17071-17080, 2022.
[13] B. Major, D. Fontijne, A. Ansari, R.T. Sukhavasi, R. Gowaikar, M. Hamilton et al., “Vehicle detection with automotive radar using deep learning on range-azimuth-doppler tensors,” in Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Korea (South): IEEE, pp. 924-932, 2019.
[14] J. Vargas, S. Alsweiss, O. Toker, R. Razdan, and J. Santos, “An overview of autonomous vehicles sensors and their vulnerability to weather conditions,” Sensors, vol. 21, no. 16, p. 5397, 2021.
[15] Z. Wei, F. Zhang, S. Chang, Y. Liu, H. Wu, and Z. Feng, “MmWave radar and vision fusion for object detection in autonomous driving: A review,” Sensors, vol. 22, no. 7, p. 2542, 2022.
[16] E. Arnold, O. Y. Al-Jarrah, M. Dianati, S. Fallah, D. Oxtoby, and A. Mouzakitis, “A survey on 3D object detection methods for autonomous driving applications,” IEEE Transactions on Intelligent Transportation Systems, vol. 20, no. 10, pp. 3782–3795, 2019.
[17] B. Shahian Jahromi, T. Tulabandhula, and S. Cetin, “Real-time hybrid multi-sensor fusion framework for perception in autonomous vehicles,” Sensors, vol. 19, no. 20, p. 4357, 2019.
[18] “Image sensor,” Wikipedia. Mar. 24, 2023. Accessed: Aug. 14, 2023. [Online]. Available: https://en.wikipedia.org/w/index.php?title=Image_sensor&oldid=1146373856
[19] S. Saponara, M.S. Greco, and F. Gini, “Radar-on-Chip/in-package in autonomous driving vehicles and intelligent transport systems: Opportunities and challenges,” IEEE Signal Processing Magazine, vol. 36, no. 5, pp. 71–84, 2019.
[20] A. Soumya, C. Krishna Mohan, and L.R. Cenkeramaddi, “Recent advances in mmWave-radar-based sensing, its applications, and machine learning techniques: A review,” Sensors, vol. 23, no. 21, p. 8901, 2023.
[21] Z. Wang, Y. Wu, and Q. Niu, “Multi-sensor fusion in automated driving: A survey,” IEEE Access, vol. 8, pp. 2847–2868, 2020.
[22] “Detecting static objects in view using—electrical engineering stack exchange,” [Online]. Available: https://electronics. stackexchange.com/questions/236484/detecting-static-objects-in-view-using-radar
[23] M. Cho, “A study on the obstacle recognition for autonomous driving RC car using lidar and thermal infrared camera,” in 2019 Eleventh International Conference on Ubiquitous and Future Networks (ICUFN), Zagreb, Croatia: IEEE, pp. 544–546, 2019.
[24] S. Sun, A. P. Petropulu, and H. V. Poor, “MIMO radar for advanced driver-assistance systems and autonomous driving: Advantages and challenges,” IEEE Signal Processing Magazine, vol. 37, no. 4, pp. 98–117, 2020.
[25] O. Schumann, M. Hahn, J. Dickmann, and C. Wohler, “Semantic segmentation on radar point clouds,” in 2018 21st International Conference on Information Fusion (FUSION), Cambridge: IEEE, pp. 2179–2186, 2018.
[26] F. Fent, P. Bauerschmidt, and M. Lienkamp, “RadarGNN: Transformation invariant graph neural network for radar-based perception,” in 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada: IEEE, pp. 182–191, 2023.
[27] S. Rahman and D. A. Robertson, “Radar micro-Doppler signatures of drones and birds at K-band and W-band,” Scientific Reports, vol. 8, no. 1, p. 17396, 2018.
[28] A. Ouaknine, A. Newson, J. Rebut, F. Tupin, and P. Pérez, “CARRADA Dataset: Camera and automotive radar with range-angle-doppler annotations,” arXiv, 2021. Accessed: Jan. 29, 2024. [Online]. Available: http://arxiv.org/abs/2005.01456
[29] D. Gusland, J. M. Christiansen, B. Torvik, F. Fioranelli, S. Z. Gurbuz, and M. Ritchie, “Open radar initiative: Large scale dataset for benchmarking of micro-Doppler recognition algorithms,” in 2021 IEEE Radar Conference (RadarConf21), Atlanta, GA, USA: IEEE, pp. 1–6, 2021.
[30] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA: IEEE, pp. 580–587, 2014.
[31] R. Girshick, “Fast R-CNN,” in 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile: IEEE, pp. 1440–1448, 2015.
[32] S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137–1149, 2017.
[33] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Y. Fu et al., “SSD: Single shot multibox detector,” in Computer Vision – ECCV 2016, B. Leibe, J. Matas, N. Sebe, and M. Welling, Eds., in Lecture Notes in Computer Science, Cham: Springer International Publishing, vol. 9905, pp. 21–37, 2016.
[34] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA: IEEE, 2016.
[35] N. Abdul Razak, N. A. A. Sabri, J. Johari, F. Ahmat Ruslan, M. Md. Kamal, and M. A. Abdul Aziz, “Investigation of object detection and identification at different lighting conditions for autonomous vehicle application,” International Journal of Automotive and Mechanical Engineering, vol. 20, no. 3, pp. 10649–10658, 2023.
[36] W. Lv, Y. Zhao, S. Xu, J. Wei, G. Wang, Q. Dang et al., “DETRs Beat YOLOs on real-time object detection,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16965-16974, 2024.
[37] X. Zhu, W. Su, L. Lu, B. Li, X. Wang, and J. Dai, “Deformable DETR: Deformable transformers for end-to-end object detection,” arXiv, 2021. Accessed: Jan. 28, 2024. [Online]. Available: http://arxiv.org/abs/2010.04159
[38] F. Liu, H. Wei, W. Zhao, G. Li, J. Peng, and Z. Li, “WB-DETR: Transformer-based detector without backbone,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada: IEEE, pp. 2959–2967, 2021.
[39] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang et al., “Swin Transformer: Hierarchical vision transformer using shifted windows,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada: IEEE, pp. 9992–10002, 2021.
[40] Z. Peng, W. Huang, S. Gu, L. Xie, Y. Wang, J. Jiao, et al., “Conformer: Local features coupling global representations for visual recognition,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada: IEEE, pp. 357–366, 2021.
[41] S. Mehta and M. Rastegari, “MobileViT: Light-weight, general-purpose, and mobile-friendly vision transformer.” arXiv, Mar. 04, 2022. Accessed: Jan. 28, 2024. [Online]. Available: http://arxiv.org/abs/2110.02178
[42] Z. Chen, L. Xie, J. Niu, X. Liu, L. Wei, and Q. Tian, “Visformer: The vision-friendly transformer,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada: IEEE, pp. 569–578, 2021.
[43] R. Nabati and H. Qi, “CenterFusion: Center-based radar and camera fusion for 3D object detection,” in 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA: IEEE, pp. 1526–1535, 2021.
[44] Y. Wang, Z. Jiang, X. Gao, J.-N. Hwang, G. Xing, and H. Liu, “RODNet: Radar object detection using cross-modal supervision,” in 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA: IEEE, pp. 504–513, 2021.
[45] Y. Kim and T. Moon, “Human detection and activity classification based on micro-doppler signatures using deep convolutional neural networks,” IEEE Geoscience and Remote Sensing Letters, vol. 13, no. 1, pp. 8–12, 2016.
[46] S. Lee, Y. Yoon, J. Lee, and S. Kim, “Human–vehicle classification using feature‐based SVM in 77‐GHz automotive FMCW radar,” IET Radar, Sonar & Navigation, vol. 11, no. 10, pp. 1589–1596, 2017.
[47] J. Redmon and A. Farhadi, “YOLOv3: An incremental improvement,” arXiv, Apr. 08, 2018. Accessed: Jan. 28, 2024. [Online]. Available: http://arxiv.org/abs/1804.02767
[48] X. Dong, P. Wang, P. Zhang, and L. Liu, “Probabilistic oriented object detection in automotive radar,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA: IEEE, pp. 458–467, 2020.
[49] W. Ng, G. Wang, Siddhartha, Z. Lin, and B. J. Dutta, “Range-doppler detection in automotive radar with deep learning,” in 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, United Kingdom: IEEE, pp. 1–8, 2020.
[50] C. Decourt, R. VanRullen, D. Salle, and T. Oberlin, “DAROD: A deep automotive radar object detector on range-doppler maps,” in 2022 IEEE Intelligent Vehicles Symposium (IV), Aachen, Germany: IEEE, pp. 112–118, 2022.
[51] X. Gao, G. Xing, S. Roy, and H. Liu, “RAMP-CNN: A novel neural network for enhanced automotive radar object recognition,” IEEE Sensors Journal, vol. 21, no. 4, pp. 5119–5132, 2021.
[52] J. Rebut, A. Ouaknine, W. Malik, and P. Perez, “Raw high-definition radar for multi-task learning,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA: IEEE, pp. 17000–17009, 2022.
[53] A. Danzer, T. Griebel, M. Bach, and K. Dietmayer, “2D Car Detection in Radar Data with PointNets,” in 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand: IEEE, pp. 61–66, 2019.
[54] J. F. Tilly, S. Haag, O. Schumann, F. Weishaupt, B. Duraisamy, J. Dickmann et al., “Detection and tracking on automotive radar data with deep learning,” in 2020 IEEE 23rd International Conference on Information Fusion (FUSION), Rustenburg, South Africa: IEEE, pp. 1–7, 2020.
[55] A. Dubey, A. Santra, J. Fuchs, M. Lübke, R. Weigel, and F. Lurz, “HARadNet: Anchor-free target detection for radar point clouds using hierarchical attention and multi-task learning,” Machine Learning with Applications, vol. 8, p. 100275, 2022.
[56] R. Q. Charles, H. Su, M. Kaichun, and L. J. Guibas, “PointNet: Deep learning on point sets for 3D classification and segmentation,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI: IEEE, pp. 77–85, 2017.
[57] C. R. Qi, L. Yi, H. Su, and L. J. Guibas, “PointNet++: Deep hierarchical feature learning on point sets in a metric space,” Jun. 07, 2017, arXiv: arXiv:1706.02413.
[58] C. R. Qi, W. Liu, C. Wu, H. Su, and L. J. Guibas, “Frustum PointNets for 3D object detection from RGB-D data,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA: IEEE, pp. 918–927, 2018.
[59] M. Dreher, E. Ercelik, T. Banziger, and A. Knoll, “Radar-based 2D car detection using deep neural networks,” in 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), Rhodes, Greece: IEEE, pp. 1–8, 2020.
[60] D. Köhler, M. Quach, M. Ulrich, F. Meinl, B. Bischoff, and H. Blume, “Improved multi-scale grid rendering of point clouds for radar object detection networks,” in 2023 26th International Conference on Information Fusion (FUSION), Charleston, SC, USA: IEEE, pp. 1–8, 2023.
[61] J. Liu, Q. Zhao, W. Xiong, T. Huang, Q.-L. Han, and B. Zhu, “SMURF: Spatial multi-representation fusion for 3D object detection with 4D imaging radar,” IEEE Transactions on Intelligent Vehicles, pp. 1–14, 2023.
[62] Y. Zhou and O. Tuzel, “VoxelNet: End-to-end learning for point cloud based 3D object detection,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA: IEEE, pp. 4490–4499, 2018.
[63] P. Svenningsson, F. Fioranelli, and A. Yarovoy, “Radar-PointGNN: Graph based object recognition for unstructured radar point-cloud data,” in 2021 IEEE Radar Conference (RadarConf21), Atlanta, GA, USA: IEEE, pp. 1–6, 2021.
[64] S. Saponara and B. Neri, “Radar sensor signal acquisition and multidimensional FFT processing for surveillance applications in transport systems,” IEEE Transactions on Instrumentation and Measurement, vol. 66, no. 4, pp. 604–615, 2017.
[65] A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? The KITTI vision benchmark suite,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI: IEEE, pp. 3354–3361, 2012.
[66] M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson et al., “The cityscapes dataset for semantic urban scene understanding,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA: IEEE, pp. 3213–3223, 2016.
[67] X. Huang, X. Cheng, Q. Geng, B. Cao, D. Zhou, P. Wang et al., “The ApolloScape dataset for autonomous driving,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT: IEEE, pp. 1067–10676, 2018.
[68] P. Sun, H. Kretzschmar, X. Dotiwalla, A. Chouard, V. Patnaik, P. Tsui et al., “Scalability in perception for autonomous driving: Waymo open dataset,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA: IEEE, pp. 2443–2451, 2020.
[69] Y. Wang, G. Wang, H.-M. Hsu, H. Liu, and J.-N. Hwang, “Rethinking of Radar’s Role: A camera-radar dataset and systematic annotator via coordinate alignment,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Nashville, TN, USA: IEEE, pp. 2809–2818, 2021.
[70] A. Ouaknine, A. Newson, J. Rebut, F. Tupin, and P. Perez, “CARRADA Dataset: Camera and automotive radar with range- angle- doppler annotations,” in 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy: IEEE, pp. 5068–5075, 2021.
[71] Palffy, E. Pool, S. Baratam, J. F. P. Kooij, and D. M. Gavrila, “Multi-class road user detection with 3+1D radar in the view-of-delft dataset,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 4961–4968, 2022.
[72] S. Chadwick, W. Maddern, and P. Newman, “Distant vehicle detection using radar and vision,” in 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada: IEEE, pp. 8311–8317, 2019.
[73] T. Roddick, A. Kendall, and R. Cipolla, “Orthographic feature transform for monocular 3D object detection.” arXiv, Nov. 20, 2018. Accessed: Feb. 05, 2024. [Online]. Available: http://arxiv.org/abs/1811.08188
[74] A. Simonelli, S. R. Bulo, L. Porzi, M. L. Antequera, and P. Kontschieder, “Disentangling monocular 3D object detection: From single to multi-class recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 3, pp. 1219–1231, 2022.
[75] X. Zhou, D. Wang, and P. Krähenbühl, “Objects as points.” arXiv, Apr. 25, 2019. Accessed: Feb. 05, 2024. [Online]. Available: http://arxiv.org/abs/1904.07850
[76] R. Yadav, A. Vierling, and K. Berns, “Radar + RGB fusion for robust object detection in autonomous vehicle,” in 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates: IEEE, pp. 1986–1990, 2020.
[77] G. Alessandretti, A. Broggi, and P. Cerri, “Vehicle and guard rail detection using radar and vision data fusion,” IEEE Transactions on Intelligent Transportation Systems, vol. 8, no. 1, pp. 95–105, 2007.
[78] F. Nobis, M. Geisslinger, M. Weber, J. Betz, and M. Lienkamp, “A deep learning-based radar and camera sensor fusion architecture for object detection,” in 2019 Sensor Data Fusion: Trends, Solutions, Applications (SDF), Bonn, Germany: IEEE, pp. 1–7, 2019.
[79] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv, Apr. 10, 2015. Accessed: May 28, 2023. [Online]. Available: http://arxiv.org/abs/1409.1556
[80] K. Bansal, K. Rungta, and D. Bharadia, “RadSegNet: A reliable approach to radar camera fusion,” arXiv, Aug. 07, 2022.
[81] Y. Long, D. Morris, X. Liu, M. Castro, P. Chakravarty, and P. Narayanan, “Radar-camera pixel depth association for depth completion,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12502–12511, 2021.
[82] V. John and S. Mita, “RVNet: Deep sensor fusion of monocular camera and radar for image-based obstacle detection in challenging environments,” in Image and Video Technology, C. Lee, Z. Su, and A. Sugimoto, Eds., in Lecture Notes in Computer Science, Cham: Springer International Publishing, vol. 11854, pp. 351–364, 2019.
[83] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, 2016.
[84] X. Guo, J. Du, J. Y. Gao, and W. Wang, “Pedestrian detection based on fusion of millimeter wave radar and vision,” Proceedings of the 2018 International Conference on Artificial Intelligence and Pattern Recognition, pp. 38-42, 2018.
[85] S. Chang, Y. Zhang, F. Zhang, X. Zhao, S. Huang, Z. Feng et al., “Spatial attention fusion for obstacle detection using MmWave radar and vision sensor,” Sensors, vol. 20, no. 4, p. 956, 2020.
[86] DU Xiaoyu, Research on vehicle front target detection algorithm based on MMW radar and visual information fusion [D]. Chongqing University of Science and Technology, 2021.
[87] S. Li, J. Yan, H. Chen, and K. Zheng, “Radar-camera fusion network for depth estimation in structured driving scenes,” Sensors, vol. 23, no. 17, p. 7560, 2023.
[88] T. Kurniawan, “ClusterFusion: Leveraging radar spatial features for radar-camera 3D object detection in autonomous vehicles,” IEEE Access, vol. 11, pp. 121511–121528, 2023.
[89] S. Pang, D. Morris, and H. Radha, “TransCAR: Transformer-based camera-and-radar fusion for 3D object detection,” arXiv, Apr. 30, 2023. Accessed: Feb. 12, 2024. [Online]. Available: http://arxiv.org/abs/2305.00397
[90] Kim, Y. Kim, and D. Kum, “Low-level sensor fusion network for 3D vehicle detection using radar range-azimuth heatmap and monocular image,” In Proceedings of the Asian Conference on Computer Vision, 2020.
[91] H. Jha, V. Lodhi, and D. Chakravarty, “Object detection and identification using vision and radar data fusion system for ground-based navigation,” in 2019 6th International Conference on Signal Processing and Integrated Networks (SPIN), pp. 590–593, 2019.
[92] X. Dong, B. Zhuang, Y. Mao, and L. Liu, “Radar camera fusion via representation learning in autonomous driving,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1672–1681, 2021.
[93] Philion, A. Kar, and S. Fidler, “Learning to evaluate perception models using planner-centric metrics,” arXiv, Apr. 18, 2020. Accessed: Feb. 12, 2024. [Online]. Available: http://arxiv.org/abs/2004.08745
[94] J. Kim, M. Seong, G. Bang, D. Kum, and J. W. Choi, “RCM-Fusion: Radar-camera multi-level fusion for 3D object detection.” arXiv, Feb. 05, 2024. Accessed: Feb. 12, 2024. [Online]. Available: http://arxiv.org/abs/2307.10249
[95] Lei, Z. Chen, S. Jia, and X. Zhang, “HVDetFusion: A simple and robust camera-radar fusion framework,” arXiv, Jul. 20, 2023. Accessed: Feb. 08, 2024. [Online]. Available: http://arxiv.org/abs/2307.11323
[96] Y. Kim, J. Shin, S. Kim, I.-J. Lee, J. W. Choi, and D. Kum, “CRN: Camera radar net for accurate, robust, efficient 3D perception,” in 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France: IEEE, pp. 17569–17580, 2023.
[97] Y. Kim, S. Kim, J. W. Choi, and D. Kum, “CRAFT: Camera-radar 3D object detection with spatio-contextual fusion transformer,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 1, pp. 1160–1168, 2023.
[98] K. Duan, S. Bai, L. Xie, H. Qi, Q. Huang, and Q. Tian, “CenterNet: Keypoint triplets for object detection,” in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea (South): IEEE, pp. 6568–6577, 2019.
[99] T. Zhou, J. Chen, Y. Shi, K. Jiang, M. Yang, and D. Yang, “Bridging the View Disparity Between Radar and Camera Features for Multi-Modal Fusion 3D Object Detection,” IEEE Trans. Intell. Veh., vol. 8, no. 2, pp. 1523–1535, Feb. 2023.
[100] Cui, Y. Ma, J. Lu, and Z. Wang, “Radar enlighten the dark: Enhancing low-visibility perception for automated vehicles with camera-radar fusion.” arXiv, May 26, 2023. Accessed: Feb. 05, 2024. [Online]. Available: http://arxiv.org/abs/2305.17318
[101] A. Sengupta, A. Yoshizawa, and S. Cao, “Automatic radar-camera dataset generation for sensor-fusion applications,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 2875–2882, 2022.
[102] E. Haussmann, M. Fenzi, K. Chitta, J. Ivanecky, H. Xu, D. Roy et al., “Scalable active learning for object detection,” in 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA: IEEE, pp. 1430–1435, 2020.
[103] Z. Zhu, K. Lin, A. K. Jain, and J. Zhou, “Transfer learning in deep reinforcement learning: A survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 11, pp. 13344–13362, 2023.
[104] Y. Zhao, Y. Yang, and S. Chen, “An active semi-supervised learning for object detection,” in 2023 International Conference on Culture-Oriented Science and Technology (CoST), Xi’an, China: IEEE, pp. 257–261, 2023.
[105] Meyer, Michael, and Georg Kuschk. “Automotive radar dataset for deep learning based 3d object detection,” In 2019 16th european radar conference (EuRAD), pp. 129-132, 2019.
[106] L. Zheng, Z. Ma, X. Zhu, B. Tan, S. Li, K. Long et al., “TJ4DRadSet: A 4D radar dataset for autonomous driving,” in 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC), Macau, China: IEEE, pp. 493–498, 2022.
[107] J. Deng, W. Dong, R. Socher, L.-J. Li, Kai Li, and Li Fei-Fei, “ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL: IEEE, pp. 248–255, 2009.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 The Author(s)

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.