Skip to main content

2021 | OriginalPaper | Buchkapitel

An Evaluation on Effectiveness of Deep Learning in Detecting Small Object Within a Large Image

verfasst von : Nazirah Hassan, Kong Wai Ming, Choo Keng Wah

Erschienen in: 17th International Conference on Biomedical Engineering

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Multiple Deep Learning (DL) algorithms have been developed recently and are shown to be achieving very high accuracy in object detection. However, challenges have been reported in detecting small objects within a large image (e.g. > 2000 by 2000 in resolution). Various methods have been proposed using different detection algorithms in order to detect small objects. However, these approaches require high computational resources and are not suitable for edge computing devices that are used for practical applications such as pedestrian traffic light detection. We explored two different methods of detection to evaluate which method is best at detecting small objects. The first method is a two—part procedure with the first step being image processing and the second step, a R-CNN based detection using Edge Boxes algorithm for the extraction of region proposals. The second method is solely Faster R-CNN Object Detection with Instance Segmentation, termed as Mask R-CNN. A total of 4000 streets images of Singapore with pedestrian traffic lights were used as training data. The dimensions of the images range from 1200 by 900 to 4000 by 3000. The small object to be detected is the green or red man within pedestrian traffic lights. We evaluated these methods based on training time required, detection time, accuracy as well as suitability for deployment in edge computing devices. From the results, it is shown that the HSV + R-CNN approach is preferred as it achieves an accuracy of 95.5% and can be deployed in edge devices.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Al-Qizwini, M., Barjasteh, I., AlQassab, H., Radha, H.: Deep learning algorithm for autonomous driving using googlenet. In: Intelligent Vehicles Symposium (IV), 2017 IEEE, pp. 89–96. IEEE (2017) Al-Qizwini, M., Barjasteh, I., AlQassab, H., Radha, H.: Deep learning algorithm for autonomous driving using googlenet. In: Intelligent Vehicles Symposium (IV), 2017 IEEE, pp. 89–96. IEEE (2017)
2.
Zurück zum Zitat Chen, C., Seff, A., Kornhauser, A., Xiao, J.: Deepdriving: Learning affordance for direct perception in autonomous driving. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2722–2730 (2015) Chen, C., Seff, A., Kornhauser, A., Xiao, J.: Deepdriving: Learning affordance for direct perception in autonomous driving. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2722–2730 (2015)
3.
Zurück zum Zitat Chen, G., Han, T. X., He, Z., Kays, R., Forrester, T.: Deep convolutional neural network based species recognition for wild animal monitoring. In: IEEE International Conference on Image Processing (ICIP), pp. 858–862 (2014) Chen, G., Han, T. X., He, Z., Kays, R., Forrester, T.: Deep convolutional neural network based species recognition for wild animal monitoring. In: IEEE International Conference on Image Processing (ICIP), pp. 858–862 (2014)
4.
Zurück zum Zitat Gomez Villa, A., Salazar, A., Vargas, F.: Towards automatic wild animal monitoring: Identification of animal species in camera-trap images using very deep convolutional neural networks. Ecological Informatics 41, 24–32 (2017)CrossRef Gomez Villa, A., Salazar, A., Vargas, F.: Towards automatic wild animal monitoring: Identification of animal species in camera-trap images using very deep convolutional neural networks. Ecological Informatics 41, 24–32 (2017)CrossRef
5.
Zurück zum Zitat Olliverre, N., Yang, G., Slabaugh, G., Reyes-Aldasoro, C.C., Alonso, E.: International Workshop on Simulation and Synthesis in Medical Imaging. Springer; Cham, Switzerland: 2018. Generating Magnetic Resonance Spectroscopy Imaging Data of Brain Tumours from Linear, Non-linear and Deep Learning Models, pp. 130–138 (2018) Olliverre, N., Yang, G., Slabaugh, G., Reyes-Aldasoro, C.C., Alonso, E.: International Workshop on Simulation and Synthesis in Medical Imaging. Springer; Cham, Switzerland: 2018. Generating Magnetic Resonance Spectroscopy Imaging Data of Brain Tumours from Linear, Non-linear and Deep Learning Models, pp. 130–138 (2018)
7.
Zurück zum Zitat Cheng, R., Wang, K., Yang, K., Long, N., Bai, J., Liu, D.: Real-time pedestrian crossing lights detection algorithm for the visually impaired. Multimedia Tools Appl. 77(16), 20651–20671 (2018) Cheng, R., Wang, K., Yang, K., Long, N., Bai, J., Liu, D.: Real-time pedestrian crossing lights detection algorithm for the visually impaired. Multimedia Tools Appl. 77(16), 20651–20671 (2018)
8.
Zurück zum Zitat de Charette, R., Nashashibi, F.: Traffic light recognition using image processing compared to learning processes. In: Proceedings of the 22nd International Con- ference on Intelligent Robots and Systems, IEEE, pp. 333–338 (2009) de Charette, R., Nashashibi, F.: Traffic light recognition using image processing compared to learning processes. In: Proceedings of the 22nd International Con- ference on Intelligent Robots and Systems, IEEE, pp. 333–338 (2009)
9.
Zurück zum Zitat Lu, Y., Lu, J., Zhang, S., Hall, P.: Traffic signal detection and classification in street views using an attention model: Computational Visual Media, vol. 4, No. 3, pp. 253–266 (2018) Lu, Y., Lu, J., Zhang, S., Hall, P.: Traffic signal detection and classification in street views using an attention model: Computational Visual Media, vol. 4, No. 3, pp. 253–266 (2018)
12.
Zurück zum Zitat Haralick, Robert, M., Linda, G.: Shapiro, Computer and Robot Vision, vol. I, Addison-Wesley, pp. 28–48 (1992) Haralick, Robert, M., Linda, G.: Shapiro, Computer and Robot Vision, vol. I, Addison-Wesley, pp. 28–48 (1992)
14.
Zurück zum Zitat Liu, W., Anguelov, D., Erhan, DE., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: SSD: Single Shot Multibox Detector. In: European conference on computer vision, pp. 21–37 (2016) Liu, W., Anguelov, D., Erhan, DE., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: SSD: Single Shot Multibox Detector. In: European conference on computer vision, pp. 21–37 (2016)
15.
Zurück zum Zitat Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779–788 (2016) Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779–788 (2016)
16.
Zurück zum Zitat Ren, S., He, K., Girshick, R. Sun, J.: Faster R-CNN: Towards real-time object detection with region proposal networks. In: Advances in neural information processing systems, pp. 91–99 (2015) Ren, S., He, K., Girshick, R. Sun, J.: Faster R-CNN: Towards real-time object detection with region proposal networks. In: Advances in neural information processing systems, pp. 91–99 (2015)
17.
Zurück zum Zitat He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, pp. 2980–2988, 22–29 Oct 2017 He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, pp. 2980–2988, 22–29 Oct 2017
18.
Zurück zum Zitat He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016, pp. 770–778 (2017) He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016, pp. 770–778 (2017)
19.
Zurück zum Zitat Lin, T., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, pp. 936–944 (2017) Lin, T., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, pp. 936–944 (2017)
Metadaten
Titel
An Evaluation on Effectiveness of Deep Learning in Detecting Small Object Within a Large Image
verfasst von
Nazirah Hassan
Kong Wai Ming
Choo Keng Wah
Copyright-Jahr
2021
DOI
https://doi.org/10.1007/978-3-030-62045-5_17

Neuer Inhalt