Skip to main content

2016 | OriginalPaper | Buchkapitel

A Large Contextual Dataset for Classification, Detection and Counting of Cars with Deep Learning

verfasst von : T. Nathan Mundhenk, Goran Konjevod, Wesam A. Sakla, Kofi Boakye

Erschienen in: Computer Vision – ECCV 2016

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

We have created a large diverse set of cars from overhead images (Data sets, annotations, networks and scripts are available from http://​gdo-datasci.​ucllnl.​org/​cowc/​), which are useful for training a deep learner to binary classify, detect and count them. The dataset and all related material will be made publically available. The set contains contextual matter to aid in identification of difficult targets. We demonstrate classification and detection on this dataset using a neural network we call ResCeption. This network combines residual learning with Inception-style layers and is used to count cars in one look. This is a new way to count objects rather than by localization or density estimation. It is fairly accurate, fast and easy to implement. Additionally, the counting method is not car or scene specific. It would be easy to train this method to count other kinds of objects and counting over new scenes requires no extra set up or assumptions about object locations.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Crawford, J.: Beyond supply and demand: making the invisible hand visible. In: Re-Work Deep Learning Summit, San Francisco, January 2016 Crawford, J.: Beyond supply and demand: making the invisible hand visible. In: Re-Work Deep Learning Summit, San Francisco, January 2016
2.
Zurück zum Zitat Tanner, F., Colder, B., Pullen, C., Heagy, D., Eppolito, M., Carlan, V., Oertel, C., Sallee, P.: Overhead imagery research data set: an annotated data library and tools to aid in the developement of computer vision algorithms. In: IEEE Applied Imagery Pattern Recognition Workshop (2009) Tanner, F., Colder, B., Pullen, C., Heagy, D., Eppolito, M., Carlan, V., Oertel, C., Sallee, P.: Overhead imagery research data set: an annotated data library and tools to aid in the developement of computer vision algorithms. In: IEEE Applied Imagery Pattern Recognition Workshop (2009)
3.
Zurück zum Zitat Razakarivony, S., Jurie, F.: Vehicle detection in aerial imegery: A small target detection benchmark. Journal of Visual Communication and Image Representation, December 2015. \(<\)hal-01122605v2\(>\) Razakarivony, S., Jurie, F.: Vehicle detection in aerial imegery: A small target detection benchmark. Journal of Visual Communication and Image Representation, December 2015. \(<\)hal-01122605v2\(>\)
10.
Zurück zum Zitat Chen, X., Xiang, S., Liu, C.L., Pan, C.H.: Vehicle detection in satellite images by parallel deep convolutional neural networks. In: Second IAPR Asian Conference on Pattern Recognition (2013) Chen, X., Xiang, S., Liu, C.L., Pan, C.H.: Vehicle detection in satellite images by parallel deep convolutional neural networks. In: Second IAPR Asian Conference on Pattern Recognition (2013)
11.
Zurück zum Zitat Moranduzzo, T., Melgani, F.: Automatic car counting method for unmanned aerial vehicle images. IEEE Trans. Geosci. Remote Sens. 52(3), 1635–1647 (2014)CrossRef Moranduzzo, T., Melgani, F.: Automatic car counting method for unmanned aerial vehicle images. IEEE Trans. Geosci. Remote Sens. 52(3), 1635–1647 (2014)CrossRef
12.
Zurück zum Zitat Holt, A.C., Seto, E.Y.W., Rivard, T., Gong, P.: Object-based detection and classification of vehicles from high-resolution aerial photography. Photogram. Eng. Remote Sens. 75(7), 871–880 (2009)CrossRef Holt, A.C., Seto, E.Y.W., Rivard, T., Gong, P.: Object-based detection and classification of vehicles from high-resolution aerial photography. Photogram. Eng. Remote Sens. 75(7), 871–880 (2009)CrossRef
13.
Zurück zum Zitat Kamenetsky, D., Sherrah, J.: Aerial car detection and urban understanding. In: IEEE Conference on Digital Image Computing: Techniques and Applications (DICTA) (2015) Kamenetsky, D., Sherrah, J.: Aerial car detection and urban understanding. In: IEEE Conference on Digital Image Computing: Techniques and Applications (DICTA) (2015)
14.
Zurück zum Zitat Zhang, C., Li, H., Wang, X., Yang, X.: Cross-scene crowd counting via deep convolutional neural networks. In: CVPR (2015) Zhang, C., Li, H., Wang, X., Yang, X.: Cross-scene crowd counting via deep convolutional neural networks. In: CVPR (2015)
15.
Zurück zum Zitat Arteta, C., Lempitsky, V., Noble, J.A., Zisserman, A.: Interactive object counting. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part III. LNCS, vol. 8691, pp. 504–518. Springer, Heidelberg (2014) Arteta, C., Lempitsky, V., Noble, J.A., Zisserman, A.: Interactive object counting. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part III. LNCS, vol. 8691, pp. 504–518. Springer, Heidelberg (2014)
16.
Zurück zum Zitat Lempitsky, V., Zisserman, A.: Learning to count objects in images. In: NIPS (2010) Lempitsky, V., Zisserman, A.: Learning to count objects in images. In: NIPS (2010)
17.
Zurück zum Zitat French, G., Fisher, M.H., Mackiewicz, M., Needle, C.L.: Convolutional neural network for counting fish in fisheries surveillance video. In: BMVC (2015) French, G., Fisher, M.H., Mackiewicz, M., Needle, C.L.: Convolutional neural network for counting fish in fisheries surveillance video. In: BMVC (2015)
18.
Zurück zum Zitat Idrees, H., Saleemi, I., Seibert, C., Shah, M.: Multi-source multi-scale counting in extremely dense crowd images. In: CVPR (2013) Idrees, H., Saleemi, I., Seibert, C., Shah, M.: Multi-source multi-scale counting in extremely dense crowd images. In: CVPR (2013)
19.
Zurück zum Zitat Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. arXiv preprint arXiv:1506.02640 (2015) Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. arXiv preprint arXiv:​1506.​02640 (2015)
20.
Zurück zum Zitat Segue, S., Pujol, O., Vitria, J.: Learning to count with deep object features. In: CVPR (2015) Segue, S., Pujol, O., Vitria, J.: Learning to count with deep object features. In: CVPR (2015)
21.
Zurück zum Zitat Wang, C., Zhang, H., Yang, L., Liu, S., Cao, X.: Deep people counting in extremely dense crowds. In: Proceedings of the 23rd Annual ACM Conference on Multimedia (2015) Wang, C., Zhang, H., Yang, L., Liu, S., Cao, X.: Deep people counting in extremely dense crowds. In: Proceedings of the 23rd Annual ACM Conference on Multimedia (2015)
22.
Zurück zum Zitat Kelly, J., Sukhatme, G.S.: Visual-inertial simultaneous localization, mapping and sensor-to-sensor self-calibration. In: CIRA (2009) Kelly, J., Sukhatme, G.S.: Visual-inertial simultaneous localization, mapping and sensor-to-sensor self-calibration. In: CIRA (2009)
23.
Zurück zum Zitat Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NIPS (2013) Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NIPS (2013)
24.
Zurück zum Zitat Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: CVPR (2015) Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: CVPR (2015)
25.
Zurück zum Zitat Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML (2015) Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML (2015)
26.
27.
Zurück zum Zitat Szegedy, C., Sergey Ioffe, V.V.: Inception-v4, inception-resnet and the impact of residual connections on learning. arXiv:1602.07261 (2016) Szegedy, C., Sergey Ioffe, V.V.: Inception-v4, inception-resnet and the impact of residual connections on learning. arXiv:​1602.​07261 (2016)
28.
Zurück zum Zitat Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., Guadarrama, S., Darrell, T.: Caffe: convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093 (2014) Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., Guadarrama, S., Darrell, T.: Caffe: convolutional architecture for fast feature embedding. arXiv preprint arXiv:​1408.​5093 (2014)
29.
Zurück zum Zitat Wu, Y., Yang, M.H., Lim, J.: Online object tracking: a benchmark. In: CVPR (2013) Wu, Y., Yang, M.H., Lim, J.: Online object tracking: a benchmark. In: CVPR (2013)
31.
Zurück zum Zitat Brust, C.A., Sickert, S., Simon, M., Rodner, E., Denzler, J.: Efficient convolutional patch networks for scene understanding. In: CVPR Scene Understanding Workshop (2015) Brust, C.A., Sickert, S., Simon, M., Rodner, E., Denzler, J.: Efficient convolutional patch networks for scene understanding. In: CVPR Scene Understanding Workshop (2015)
Metadaten
Titel
A Large Contextual Dataset for Classification, Detection and Counting of Cars with Deep Learning
verfasst von
T. Nathan Mundhenk
Goran Konjevod
Wesam A. Sakla
Kofi Boakye
Copyright-Jahr
2016
DOI
https://doi.org/10.1007/978-3-319-46487-9_48