Skip to main content
Erschienen in: International Journal of Intelligent Transportation Systems Research 1/2024

26.01.2024

Enhancing Autonomous Driving By Exploiting Thermal Object Detection Through Feature Fusion

verfasst von: Moataz Eltahan, Khaled Elsayed

Erschienen in: International Journal of Intelligent Transportation Systems Research | Ausgabe 1/2024

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Autonomous driving does not imply self-driving vehicles; self-driving vehicles are the final evolution level of autonomous driving. Autonomous driving aims to add more benefits and features to the vehicle, including driver assistance technologies and automated driving systems. Object detection is a fundamental pillar in the autonomous vehicle industry. Recently, thermal imaging is increasingly gaining an essential role in autonomous driving. The use of a thermal camera in the sensor suite of an automotive vehicle has become more and more popular to increase reliability for robust and safe operation. Thermal imaging can detect the infrared radiation heat emitted by objects. Therefore, a thermal camera can outperform other sensors, such as LiDAR, radar, and RGB, in adverse lighting conditions. In this work, we study object detection in thermal imaging. We propose the Multi-Domain Feature Fusion Thermal Detector (MDFFTDet) framework that fuses the thermal weights with the RGB weights using only thermal images from the FLIR ADAS dataset. The proposed framework inherits the architecture of EfficientDet and further proposes a technique for the feature fusion of two different modalities: thermal and RGB. We use UNIT GAN for the image-to-image translation task to generate pseudo-RGB images from the input thermal images. The proposed framework could tackle the problem of the lack of large datasets in the thermal domain and improve detection accuracy. The experimental results on the FLIR ADAS dataset demonstrate that the proposed framework outperforms the detection accuracy of other state-of-the-art object detectors in thermal imagery, achieving this using much fewer parameters and FLOPS. Two variants of the framework, MDFFTDet-D0 and MDFFTDet-D2, are proposed, achieving mAP of 63.15% and 77.81%, respectively. On the other hand, EfficientDet-D0 and EfficientDet-D2, which are used as a baseline for the proposed frameworks, achieved mAP of 59.4% and 69.12%, respectively.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

ATZelectronics worldwide

ATZlectronics worldwide is up-to-speed on new trends and developments in automotive electronics on a scientific level with a high depth of information. 

Order your 30-days-trial for free and without any commitment.

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Weitere Produktempfehlungen anzeigen
Literatur
3.
Zurück zum Zitat Choi, J.D., Kim, M.Y.: A Sensor Fusion System with Thermal Infrared Camera and LiDAR for Autonomous Vehicles: Its Calibration and Application. Twelfth International Conference on Ubiquitous and Future Networks (ICUFN) 2021, 361–365 (2021)CrossRef Choi, J.D., Kim, M.Y.: A Sensor Fusion System with Thermal Infrared Camera and LiDAR for Autonomous Vehicles: Its Calibration and Application. Twelfth International Conference on Ubiquitous and Future Networks (ICUFN) 2021, 361–365 (2021)CrossRef
4.
Zurück zum Zitat Azam, S., Munir, F., Sheri, A.M., Ko, Y., Hussain, I., Jeon, M.: Data fusion of Lidar and Thermal Camera for Autonomous driving. Applied Industrial Optics 2019, OSA Technical Digest (Optica Publishing Group, 2019), paper T2A.5 (2019) Azam, S., Munir, F., Sheri, A.M., Ko, Y., Hussain, I., Jeon, M.: Data fusion of Lidar and Thermal Camera for Autonomous driving. Applied Industrial Optics 2019, OSA Technical Digest (Optica Publishing Group, 2019), paper T2A.5 (2019)
5.
Zurück zum Zitat Deng, J., Dong, W., Socher, R., Li, L., Li, K., Fei-Fei, L.: ImageNet: A large-scale hierarchical image database. IEEE Conference on Computer Vision and Pattern Recognition, 248–255 (2009) Deng, J., Dong, W., Socher, R., Li, L., Li, K., Fei-Fei, L.: ImageNet: A large-scale hierarchical image database. IEEE Conference on Computer Vision and Pattern Recognition, 248–255 (2009)
6.
Zurück zum Zitat Everingham, M., Gool, L.V., Williams, C.K., Winn, J.M., Zisserman, A.: The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vision 88, 303–338 (2009)CrossRef Everingham, M., Gool, L.V., Williams, C.K., Winn, J.M., Zisserman, A.: The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vision 88, 303–338 (2009)CrossRef
7.
Zurück zum Zitat Lin, T., Maire, M., Belongie, S.J., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C.L.: Microsoft COCO: Common Objects in Context. European Conference on Computer Vision (ECCV) . Lecture Notes in Computer Science, vol. 8693. Springer, Cham. https://doi.org/10.1007/978-3-319-10602-1_48 (2014) Lin, T., Maire, M., Belongie, S.J., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C.L.: Microsoft COCO: Common Objects in Context. European Conference on Computer Vision (ECCV) . Lecture Notes in Computer Science, vol. 8693. Springer, Cham. https://​doi.​org/​10.​1007/​978-3-319-10602-1_​48 (2014)
9.
Zurück zum Zitat Hwang, S., Park, J., Kim, N., Choi, Y., & So Kweon, I.: Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1037–1045) (2015) Hwang, S., Park, J., Kim, N., Choi, Y., & So Kweon, I.: Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1037–1045) (2015)
10.
Zurück zum Zitat Goodfellow, I.J.: NIPS 2016 Tutorial: Generative Adversarial Networks. ArXiv, abs/1701.00160. (2017) Goodfellow, I.J.: NIPS 2016 Tutorial: Generative Adversarial Networks. ArXiv, abs/1701.00160. (2017)
11.
Zurück zum Zitat Isola, P., Zhu, J., Zhou, T., Efros, A.A.: Image-to-Image Translation with Conditional Adversarial Networks. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017, 5967–5976 (2017) Isola, P., Zhu, J., Zhou, T., Efros, A.A.: Image-to-Image Translation with Conditional Adversarial Networks. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017, 5967–5976 (2017)
12.
Zurück zum Zitat Chu, C., Zhmoginov, A., & Sandler, M.: Cyclegan, a master of steganography. arXiv preprint arXiv:1712.02950 (2017) Chu, C., Zhmoginov, A., & Sandler, M.: Cyclegan, a master of steganography. arXiv preprint arXiv:1712.02950 (2017)
13.
Zurück zum Zitat Liu, M., Breuel, T.M., Kautz, J.: Unsupervised Image-to-Image Translation Networks. International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, USA, 700–708 (2017) Liu, M., Breuel, T.M., Kautz, J.: Unsupervised Image-to-Image Translation Networks. International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, USA, 700–708 (2017)
14.
Zurück zum Zitat Tan, M., Le, Q.V.: EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. ArXiv, abs/1905.11946 (2019) Tan, M., Le, Q.V.: EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. ArXiv, abs/1905.11946 (2019)
15.
Zurück zum Zitat Chen, Y., Li, W., Sakaridis, C., Dai, D., Gool, L.V.: Domain Adaptive Faster R-CNN for Object Detection in the Wild. IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018, 3339–3348 (2018) Chen, Y., Li, W., Sakaridis, C., Dai, D., Gool, L.V.: Domain Adaptive Faster R-CNN for Object Detection in the Wild. IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018, 3339–3348 (2018)
16.
Zurück zum Zitat Ghose, D., Desai, S.M., Bhattacharya, S., Chakraborty, D., Fiterau, M., Rahman, T.: Pedestrian Detection in Thermal Images Using Saliency Maps. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2019, 988–997 (2019) Ghose, D., Desai, S.M., Bhattacharya, S., Chakraborty, D., Fiterau, M., Rahman, T.: Pedestrian Detection in Thermal Images Using Saliency Maps. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2019, 988–997 (2019)
17.
Zurück zum Zitat Munir, F., Azam, S., Rafique, M.A., Sheri, A.M., Jeon, M.: Thermal Object Detection using Domain Adaptation through Style Consistency. ArXiv, abs/2006.00821 (2020) Munir, F., Azam, S., Rafique, M.A., Sheri, A.M., Jeon, M.: Thermal Object Detection using Domain Adaptation through Style Consistency. ArXiv, abs/2006.00821 (2020)
18.
Zurück zum Zitat Cao, Y., Zhou, T., Zhu, X., Su, Y.: Every Feature Counts: An Improved One-Stage Detector in Thermal Imagery. 2019 IEEE 5th International Conference on Computer and Communications (ICCC), 1965–1969 (2019) Cao, Y., Zhou, T., Zhu, X., Su, Y.: Every Feature Counts: An Improved One-Stage Detector in Thermal Imagery. 2019 IEEE 5th International Conference on Computer and Communications (ICCC), 1965–1969 (2019)
19.
Zurück zum Zitat Agrawal, K., Subramanian, A.: Enhancing Object Detection in Adverse Conditions using Thermal Imaging. ArXiv, abs/1909.13551 (2019) Agrawal, K., Subramanian, A.: Enhancing Object Detection in Adverse Conditions using Thermal Imaging. ArXiv, abs/1909.13551 (2019)
20.
Zurück zum Zitat Liu, N., Han, J.: PiCANet: Learning Pixel-wise Contextual Attention in ConvNets and Its Application in Saliency Detection. ArXiv, abs/1708.06433 (2017) Liu, N., Han, J.: PiCANet: Learning Pixel-wise Contextual Attention in ConvNets and Its Application in Saliency Detection. ArXiv, abs/1708.06433 (2017)
21.
Zurück zum Zitat Deng, Z., Hu, X., Zhu, L., Xu, X., Qin, J., Han, G., Heng, P.: R3Net: Recurrent Residual Refinement Network for Saliency Detection. International Joint Conference on Artificial Intelligence (IJCAI), 684–690 (2018) Deng, Z., Hu, X., Zhu, L., Xu, X., Qin, J., Han, G., Heng, P.: R3Net: Recurrent Residual Refinement Network for Saliency Detection. International Joint Conference on Artificial Intelligence (IJCAI), 684–690 (2018)
22.
Zurück zum Zitat Li, J., Liang, X., Wei, Y., Xu, T., Feng, J., Yan, S.: Perceptual Generative Adversarial Networks for Small Object Detection. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017, 1951–1959 (2017) Li, J., Liang, X., Wei, Y., Xu, T., Feng, J., Yan, S.: Perceptual Generative Adversarial Networks for Small Object Detection. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017, 1951–1959 (2017)
23.
Zurück zum Zitat Wang, W., Hong, W., Wang, F., Yu, J.: GAN-Knowledge Distillation for One-Stage Object Detection. IEEE Access 8, 60719–60727 (2020)CrossRef Wang, W., Hong, W., Wang, F., Yu, J.: GAN-Knowledge Distillation for One-Stage Object Detection. IEEE Access 8, 60719–60727 (2020)CrossRef
24.
Zurück zum Zitat Zhang, Y., Bai, Y., Ding, M., Ghanem, B.: Multi-task Generative Adversarial Network for Detecting Small Objects in the Wild. Int. J. Comput. Vision 128, 1810–1828 (2020)MathSciNetCrossRef Zhang, Y., Bai, Y., Ding, M., Ghanem, B.: Multi-task Generative Adversarial Network for Detecting Small Objects in the Wild. Int. J. Comput. Vision 128, 1810–1828 (2020)MathSciNetCrossRef
25.
Zurück zum Zitat Devaguptapu, C., Akolekar, N., Sharma, M.M., Balasubramanian, V.N.: Borrow From Anywhere: Pseudo Multi-Modal Object Detection in Thermal Imagery. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2019, 1029–1038 (2019) Devaguptapu, C., Akolekar, N., Sharma, M.M., Balasubramanian, V.N.: Borrow From Anywhere: Pseudo Multi-Modal Object Detection in Thermal Imagery. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2019, 1029–1038 (2019)
26.
Zurück zum Zitat Zhang, S., Wen, L., Bian, X., Lei, Z., Li, S.: Single-Shot Refinement Neural Network for Object Detection. IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018, 4203–4212 (2018) Zhang, S., Wen, L., Bian, X., Lei, Z., Li, S.: Single-Shot Refinement Neural Network for Object Detection. IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018, 4203–4212 (2018)
27.
Zurück zum Zitat Bin, J., Zhang, R., Du, S., Zheng, C. L., Liu, X., Blasch, E., Liu, Z.: Improved Object Detection in Thermal Imaging Through Context Enhancement and Information Fusion: A Case Study in Autonomous Driving. NeurIPS (Autonomous Driving) 2021, Machine Learning for Autonomous Driving workshops (2021) Bin, J., Zhang, R., Du, S., Zheng, C. L., Liu, X., Blasch, E., Liu, Z.: Improved Object Detection in Thermal Imaging Through Context Enhancement and Information Fusion: A Case Study in Autonomous Driving. NeurIPS (Autonomous Driving) 2021, Machine Learning for Autonomous Driving workshops (2021)
28.
Zurück zum Zitat Tu, Z., Ma, Y., Li, Z., Li, C., Xu, J., Liu, Y.: RGBT Salient Object Detection: A Large-scale Dataset and Benchmark. ArXiv, abs/2007.03262 (2022) Tu, Z., Ma, Y., Li, Z., Li, C., Xu, J., Liu, Y.: RGBT Salient Object Detection: A Large-scale Dataset and Benchmark. ArXiv, abs/2007.03262 (2022)
29.
Zurück zum Zitat Sun, Y., Cao, B., Zhu, P., & Hu, Q.: Drone-based RGB-Infrared Cross-Modality Vehicle Detection via Uncertainty-Aware Learning. IEEE Transactions on Circuits and Systems for Video Technology, 32, 6700-6713 (2022) Sun, Y., Cao, B., Zhu, P., & Hu, Q.: Drone-based RGB-Infrared Cross-Modality Vehicle Detection via Uncertainty-Aware Learning. IEEE Transactions on Circuits and Systems for Video Technology, 32, 6700-6713 (2022)
30.
Zurück zum Zitat Tan, M., Pang, R., Le, Q.V.: EfficientDet: Scalable and Efficient Object Detection. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020, 10778–10787 (2020) Tan, M., Pang, R., Le, Q.V.: EfficientDet: Scalable and Efficient Object Detection. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020, 10778–10787 (2020)
32.
Zurück zum Zitat Sandler, M., Howard, A.G., Zhu, M., Zhmoginov, A., Chen, L.: MobileNetV2: Inverted Residuals and Linear Bottlenecks. IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018, 4510–4520 (2018) Sandler, M., Howard, A.G., Zhu, M., Zhmoginov, A., Chen, L.: MobileNetV2: Inverted Residuals and Linear Bottlenecks. IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018, 4510–4520 (2018)
33.
Zurück zum Zitat Tan, M., Chen, B., Pang, R., Vasudevan, V., Le, Q.V.: MnasNet: Platform-Aware Neural Architecture Search for Mobile. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019, 2815–2823 (2019) Tan, M., Chen, B., Pang, R., Vasudevan, V., Le, Q.V.: MnasNet: Platform-Aware Neural Architecture Search for Mobile. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019, 2815–2823 (2019)
34.
Zurück zum Zitat Liu, M., Tuzel, O.: Coupled Generative Adversarial Networks. International Conference on Neural Information Processing Systems (NIPS'16). Curran Associates Inc., Red Hook, NY, USA, 469–477 (2016) Liu, M., Tuzel, O.: Coupled Generative Adversarial Networks. International Conference on Neural Information Processing Systems (NIPS'16). Curran Associates Inc., Red Hook, NY, USA, 469–477 (2016)
35.
Zurück zum Zitat Lin, T., Goyal, P., Girshick, R.B., He, K., Dollár, P.: Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 42, 318–327 (2020)CrossRef Lin, T., Goyal, P., Girshick, R.B., He, K., Dollár, P.: Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 42, 318–327 (2020)CrossRef
37.
38.
Zurück zum Zitat Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. Journal of big data 6(1), 1–48 (2019)CrossRef Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. Journal of big data 6(1), 1–48 (2019)CrossRef
39.
Zurück zum Zitat Zoph, B., Cubuk, E. D., Ghiasi, G., Lin, T. Y., Shlens, J., Le, Q. V.: Learning data augmentation strategies for object detection. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXVII 16 (pp. 566–583). Springer International Publishing (2020) Zoph, B., Cubuk, E. D., Ghiasi, G., Lin, T. Y., Shlens, J., Le, Q. V.: Learning data augmentation strategies for object detection. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXVII 16 (pp. 566–583). Springer International Publishing (2020)
40.
Zurück zum Zitat Zhong, Z., Zheng, L., Kang, G., Li, S., Yang, Y.: Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence 34, 07, 13001–13008) (2020) Zhong, Z., Zheng, L., Kang, G., Li, S., Yang, Y.: Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence 34, 07, 13001–13008) (2020)
Metadaten
Titel
Enhancing Autonomous Driving By Exploiting Thermal Object Detection Through Feature Fusion
verfasst von
Moataz Eltahan
Khaled Elsayed
Publikationsdatum
26.01.2024
Verlag
Springer US
Erschienen in
International Journal of Intelligent Transportation Systems Research / Ausgabe 1/2024
Print ISSN: 1348-8503
Elektronische ISSN: 1868-8659
DOI
https://doi.org/10.1007/s13177-024-00385-5

Weitere Artikel der Ausgabe 1/2024

International Journal of Intelligent Transportation Systems Research 1/2024 Zur Ausgabe

    Marktübersichten

    Die im Laufe eines Jahres in der „adhäsion“ veröffentlichten Marktübersichten helfen Anwendern verschiedenster Branchen, sich einen gezielten Überblick über Lieferantenangebote zu verschaffen.