Skip to main content

2017 | OriginalPaper | Buchkapitel

Online Fast Deep Learning Tracker Based on Deep Sparse Neural Networks

verfasst von : Xin Wang, Zhiqiang Hou, Wangsheng Yu, Zefenfen Jin

Erschienen in: Image and Graphics

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Deep learning can explore robust and powerful feature representations from data and has gained significant attention in visual tracking tasks. However, due to its high computational complexity and time-consuming training process, the most existing deep learning based trackers require an offline pre-training process on a large scale dataset, and have low tracking speeds. Therefore, aiming at these difficulties of the deep learning based trackers, we propose an online deep learning tracker based on Sparse Auto-Encoders (SAE) and Rectifier Linear Unit (ReLU). Combined ReLU with SAE, the deep neural networks (DNNs) obtain the sparsity similar to the DNNs with offline pre-training. The inherent sparsity make the deep model get rid of the complex pre-training process and can be used for online-only tracking well. Meanwhile, the technique of data augmentation is employed in the single positive sample to balance the quantities of positive and negative samples, which improve the stability of the model to some extent. Finally, in order to overcome the problem of randomness and drift of particle filter, we adopt a local dense sampling searching method to generate a local confidence map to locate the target’s position. Moreover, several corresponding update strategies are proposed to improve the robustness of the proposed tracker. Extensive experimental results show the effectiveness and robustness of the proposed tracker in challenging environment against state-of-the-art methods. Not only the proposed tracker leaves out the complicated and time-consuming pre-training process efficiently, but achieves an online fast and robust tracking.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Smeulders, A.W.M., Chu, D.M., Cucchiara, R., et al.: Visual tracking: an experimental survey. IEEE Trans. Pattern Anal. Mach. Intell. 36(7), 1442–1468 (2014)CrossRef Smeulders, A.W.M., Chu, D.M., Cucchiara, R., et al.: Visual tracking: an experimental survey. IEEE Trans. Pattern Anal. Mach. Intell. 36(7), 1442–1468 (2014)CrossRef
2.
Zurück zum Zitat Yilmaz, A., Javed, O., Shah, M.: Object tracking: a survey. ACM Comput. Surv. 38(4), 1–45 (2006)CrossRef Yilmaz, A., Javed, O., Shah, M.: Object tracking: a survey. ACM Comput. Surv. 38(4), 1–45 (2006)CrossRef
3.
Zurück zum Zitat Li, X., Hu, W.M., Shen, C.H., et al.: A survey of appearance models in visual object tracking. ACM Trans. Intell. Syst. Technol. 4(4), Article 58 (2013) Li, X., Hu, W.M., Shen, C.H., et al.: A survey of appearance models in visual object tracking. ACM Trans. Intell. Syst. Technol. 4(4), Article 58 (2013)
4.
Zurück zum Zitat Ross, D.A., Lim, J., Lin, R.S.: Incremental learning for robust visual tracking. Int. J. Comput. Vis. 77(1–3), 125–141 (2008)CrossRef Ross, D.A., Lim, J., Lin, R.S.: Incremental learning for robust visual tracking. Int. J. Comput. Vis. 77(1–3), 125–141 (2008)CrossRef
5.
Zurück zum Zitat Mei, X., Ling, H.: Robust visual tracking using l1 minimization. In: IEEE International Conference on Computer Vision, pp. 1436–1443. IEEE, Washington, D.C. (2009) Mei, X., Ling, H.: Robust visual tracking using l1 minimization. In: IEEE International Conference on Computer Vision, pp. 1436–1443. IEEE, Washington, D.C. (2009)
6.
Zurück zum Zitat Zhang, T.Z., Ghanem, B., Liu, S., et al.: Robust visual tracking via multi-task sparse learning. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2042–2049. IEEE, Washington, D.C. (2012) Zhang, T.Z., Ghanem, B., Liu, S., et al.: Robust visual tracking via multi-task sparse learning. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2042–2049. IEEE, Washington, D.C. (2012)
7.
Zurück zum Zitat Babenko, B., Yang, M.H., Belongie, S.: Robust object tracking with online multiple instance learning. IEEE Trans. Pattern Anal. Mach. Intell. 33(8), 1619–1632 (2011)CrossRef Babenko, B., Yang, M.H., Belongie, S.: Robust object tracking with online multiple instance learning. IEEE Trans. Pattern Anal. Mach. Intell. 33(8), 1619–1632 (2011)CrossRef
8.
Zurück zum Zitat Kalal, Z., Mikolajczyk, K., Matas, J.: Tracking-learning-detection. IEEE Trans. Pattern Anal. Mach. Intell. 34(7), 1409–1422 (2012)CrossRef Kalal, Z., Mikolajczyk, K., Matas, J.: Tracking-learning-detection. IEEE Trans. Pattern Anal. Mach. Intell. 34(7), 1409–1422 (2012)CrossRef
9.
Zurück zum Zitat Hare, S., Saffari, A., Torr, P.H.: Struck: structured output tracking with kernels. In: IEEE International Conference on Computer Vision, pp. 263–270. IEEE, Washington, D.C. (2011) Hare, S., Saffari, A., Torr, P.H.: Struck: structured output tracking with kernels. In: IEEE International Conference on Computer Vision, pp. 263–270. IEEE, Washington, D.C. (2011)
10.
Zurück zum Zitat Lecun, Y., Bengo, Y., Hinton, G.: Deep learning. Nature 521(7553), 436–444 (2015)CrossRef Lecun, Y., Bengo, Y., Hinton, G.: Deep learning. Nature 521(7553), 436–444 (2015)CrossRef
11.
Zurück zum Zitat Wang, N.Y., Yeung, D.: Learning a deep compact image representation for visual tracking. In: Advances in Neural Information Processing Systems, pp. 809–817. IMLS, Nevada (2013) Wang, N.Y., Yeung, D.: Learning a deep compact image representation for visual tracking. In: Advances in Neural Information Processing Systems, pp. 809–817. IMLS, Nevada (2013)
13.
Zurück zum Zitat Zhang, K.H., Liu, Q.S., Wu, Y., et al.: Robust visual tracking via convolutional networks. IEEE Trans. Image Process. 25(4), 1779–1792 (2015)MathSciNet Zhang, K.H., Liu, Q.S., Wu, Y., et al.: Robust visual tracking via convolutional networks. IEEE Trans. Image Process. 25(4), 1779–1792 (2015)MathSciNet
14.
Zurück zum Zitat Ma, C., Huang, J.B., Yang, X.K., et al.: Hierarchical convolutional features for visual tracking. In: IEEE International Conference on Computer Vision, pp. 3074–3082. IEEE, Washington, D.C. (2015) Ma, C., Huang, J.B., Yang, X.K., et al.: Hierarchical convolutional features for visual tracking. In: IEEE International Conference on Computer Vision, pp. 3074–3082. IEEE, Washington, D.C. (2015)
15.
Zurück zum Zitat Wu, Y., Lim, J., Yang, M.H.: Online object tracking: a benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1834–1848 (2015)CrossRef Wu, Y., Lim, J., Yang, M.H.: Online object tracking: a benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1834–1848 (2015)CrossRef
16.
Zurück zum Zitat Wang, X., Hou, Z., Yu, W., et al.: Robust visual tracking via multiscale deep sparse networks. Opt. Eng. 56(4), 043107 (2017)CrossRef Wang, X., Hou, Z., Yu, W., et al.: Robust visual tracking via multiscale deep sparse networks. Opt. Eng. 56(4), 043107 (2017)CrossRef
17.
Zurück zum Zitat Arpit, D., Zhou, Y., Ngo, H., et al.: Why regularized auto-encoders learn sparse representation? In: International Conference on Machine Learning, pp. 134–144. IMLS, Nevada (2015) Arpit, D., Zhou, Y., Ngo, H., et al.: Why regularized auto-encoders learn sparse representation? In: International Conference on Machine Learning, pp. 134–144. IMLS, Nevada (2015)
18.
Zurück zum Zitat Nair, V., Hinton, G.,: Rectified linear units improve restricted Boltzmann machines. In: International Conference on Machine Learning, pp. 807–814. IMLS, Nevada (2010) Nair, V., Hinton, G.,: Rectified linear units improve restricted Boltzmann machines. In: International Conference on Machine Learning, pp. 807–814. IMLS, Nevada (2010)
19.
Zurück zum Zitat Glorot, X., Bordes, A., Bengio, Y.: Deep sparse rectifier neural networks. In: International Conference on Artificial Intelligence and Statistics, pp. 315–323. Microtome, Brookline (2011) Glorot, X., Bordes, A., Bengio, Y.: Deep sparse rectifier neural networks. In: International Conference on Artificial Intelligence and Statistics, pp. 315–323. Microtome, Brookline (2011)
20.
Zurück zum Zitat Li, J., Zhang, T., Luo, W., et al.: Sparseness analysis in the pretraining of deep neural networks. IEEE Trans. Neural Netw. Learn. Syst. PP(99), 1–14 (2016) Li, J., Zhang, T., Luo, W., et al.: Sparseness analysis in the pretraining of deep neural networks. IEEE Trans. Neural Netw. Learn. Syst. PP(99), 1–14 (2016)
21.
Zurück zum Zitat Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using scale deep network. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2366–2374. IEEE, Washington, D.C. (2014) Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using scale deep network. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2366–2374. IEEE, Washington, D.C. (2014)
22.
Zurück zum Zitat Gao, C., Chen, F., Yu, J.G., et al.: Robust visual tracking using exemplar-based detectors. IEEE Trans. Circ. Syst. Video Technol. 27(2), 300–312 (2016)CrossRef Gao, C., Chen, F., Yu, J.G., et al.: Robust visual tracking using exemplar-based detectors. IEEE Trans. Circ. Syst. Video Technol. 27(2), 300–312 (2016)CrossRef
23.
Zurück zum Zitat Zhang, T.Z., Liu, S., Xu, C.S., et al.: Structural sparse tracking. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 150–158. IEEE, Washington, D.C. (2015) Zhang, T.Z., Liu, S., Xu, C.S., et al.: Structural sparse tracking. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 150–158. IEEE, Washington, D.C. (2015)
24.
Zurück zum Zitat Zhong, W., Lu, H., Yang, M.H.: Robust object tracking via sparsity-based collaborative model. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1838–1845. IEEE, Washington, D.C. (2012) Zhong, W., Lu, H., Yang, M.H.: Robust object tracking via sparsity-based collaborative model. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1838–1845. IEEE, Washington, D.C. (2012)
25.
Zurück zum Zitat Wang, G.F., Qin, X.Y., Zhong, F., et al.: Visual tracking via sparse and local linear coding. IEEE Trans. Image Process. 24(11), 3796–3809 (2015)MathSciNetCrossRef Wang, G.F., Qin, X.Y., Zhong, F., et al.: Visual tracking via sparse and local linear coding. IEEE Trans. Image Process. 24(11), 3796–3809 (2015)MathSciNetCrossRef
26.
Zurück zum Zitat Danelljan, M., Khan, F.S., Felsberg, M., et al.: Adaptive color attributes for real-time visual tracking. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1090–1097. IEEE, Washington, D.C. (2014) Danelljan, M., Khan, F.S., Felsberg, M., et al.: Adaptive color attributes for real-time visual tracking. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1090–1097. IEEE, Washington, D.C. (2014)
27.
Zurück zum Zitat Liu, F., Zhou, T., Yang, J., et al.: Visual tracking via nonnegative regularization multiple locality coding. In: IEEE International Conference on Computer Vision Workshop, pp. 912–920. IEEE, Washington, D.C. (2016) Liu, F., Zhou, T., Yang, J., et al.: Visual tracking via nonnegative regularization multiple locality coding. In: IEEE International Conference on Computer Vision Workshop, pp. 912–920. IEEE, Washington, D.C. (2016)
28.
Zurück zum Zitat Li, H., Li, Y., Porikli, F.: DeepTrack: learning discriminative feature representations by convolutional neural networks for visual tracking. In: British Machine Vision Conference, pp. 1–12 (2014) Li, H., Li, Y., Porikli, F.: DeepTrack: learning discriminative feature representations by convolutional neural networks for visual tracking. In: British Machine Vision Conference, pp. 1–12 (2014)
Metadaten
Titel
Online Fast Deep Learning Tracker Based on Deep Sparse Neural Networks
verfasst von
Xin Wang
Zhiqiang Hou
Wangsheng Yu
Zefenfen Jin
Copyright-Jahr
2017
DOI
https://doi.org/10.1007/978-3-319-71607-7_17