Skip to main content
Erschienen in: Cognitive Computation 1/2018

25.10.2017

Learning from Few Samples with Memory Network

verfasst von: Shufei Zhang, Kaizhu Huang, Rui Zhang, Amir Hussain

Erschienen in: Cognitive Computation | Ausgabe 1/2018

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Neural networks (NN) have achieved great successes in pattern recognition and machine learning. However, the success of a NN usually relies on the provision of a sufficiently large number of data samples as training data. When fed with a limited data set, a NN’s performance may be degraded significantly. In this paper, a novel NN structure is proposed called a memory network. It is inspired by the cognitive mechanism of human beings, which can learn effectively, even from limited data. Taking advantage of the memory from previous samples, the new model achieves a remarkable improvement in performance when trained using limited data. The memory network is demonstrated here using the multi-layer perceptron (MLP) as a base model. However, it would be straightforward to extend the idea to other neural networks, e.g., convolutional neural networks (CNN). In this paper, the memory network structure is detailed, the training algorithm is presented, and a series of experiments are conducted to validate the proposed framework. Experimental results show that the proposed model outperforms traditional MLP-based models as well as other competitive algorithms in response to two real benchmark data sets.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Ruck DW, Rogers SK, Kabrisky M, Oxley ME, Suter BW. The multilayer perceptron as an approximation to a bayes optimal discriminant function. IEEE Trans Neural Netw 1990;1(4):296–298.CrossRefPubMed Ruck DW, Rogers SK, Kabrisky M, Oxley ME, Suter BW. The multilayer perceptron as an approximation to a bayes optimal discriminant function. IEEE Trans Neural Netw 1990;1(4):296–298.CrossRefPubMed
2.
Zurück zum Zitat Zhang S, Huang K. Learning from few samples with memory network. International Conference on Neural Information Processing. Springer; 2016. p. 606–614. Zhang S, Huang K. Learning from few samples with memory network. International Conference on Neural Information Processing. Springer; 2016. p. 606–614.
3.
Zurück zum Zitat Gao F, Zhang Y, Wang J, Sun J, Yang E, Hussain A. Visual attention model based vehicle target detection in synthetic aperture radar images: a novel approach. Cogn Comput 2015;7(4):434–444.CrossRef Gao F, Zhang Y, Wang J, Sun J, Yang E, Hussain A. Visual attention model based vehicle target detection in synthetic aperture radar images: a novel approach. Cogn Comput 2015;7(4):434–444.CrossRef
4.
Zurück zum Zitat Girshick R. Fast r-cnn. Proceedings of the IEEE International Conference on Computer Vision; 2015. p. 1440–1448. Girshick R. Fast r-cnn. Proceedings of the IEEE International Conference on Computer Vision; 2015. p. 1440–1448.
5.
Zurück zum Zitat Iandola FN, Han S, Moskewicz M, Ashraf K, Dally WJ, Keutzer K. 2016. Squeezenet: alexnet-level accuracy with 50x fewer parameters and 0.5 mb model size. arXiv:1602.07360. Iandola FN, Han S, Moskewicz M, Ashraf K, Dally WJ, Keutzer K. 2016. Squeezenet: alexnet-level accuracy with 50x fewer parameters and 0.5 mb model size. arXiv:1602.​07360.
6.
Zurück zum Zitat Lawrence S, Giles CL, Tsoi AC, Back AD. Face recognition: a convolutional neural-network approach. IEEE Trans Neural Netw 1997;8(1):98–113.CrossRefPubMed Lawrence S, Giles CL, Tsoi AC, Back AD. Face recognition: a convolutional neural-network approach. IEEE Trans Neural Netw 1997;8(1):98–113.CrossRefPubMed
7.
Zurück zum Zitat Lyu C, Huang K, Liang H-N. A unified gradient regularization family for adversarial examples. 2015 IEEE International Conference on Data Mining (ICDM). IEEE; 2015. p. 301–309. Lyu C, Huang K, Liang H-N. A unified gradient regularization family for adversarial examples. 2015 IEEE International Conference on Data Mining (ICDM). IEEE; 2015. p. 301–309.
8.
Zurück zum Zitat Oquab M, Bottou L, Laptev I, Sivic J. Learning and transferring mid-level image representations using convolutional neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2014. p. 1717–1724. Oquab M, Bottou L, Laptev I, Sivic J. Learning and transferring mid-level image representations using convolutional neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2014. p. 1717–1724.
9.
Zurück zum Zitat Sun Y, Liang D, Wang X, Tang X. 2015. Deepid3: face recognition with very deep neural networks. arXiv:1502.00873. Sun Y, Liang D, Wang X, Tang X. 2015. Deepid3: face recognition with very deep neural networks. arXiv:1502.​00873.
10.
Zurück zum Zitat Rumelhart DE, Hinton GE, Williams RJ. Learning representations by back-propagating errors. Cogn Model 1988;5(3):1. Rumelhart DE, Hinton GE, Williams RJ. Learning representations by back-propagating errors. Cogn Model 1988;5(3):1.
11.
Zurück zum Zitat Cambria E, Hussain A, Vol. 1. Sentic computing: a common-sense-based framework for concept-level sentiment analysis. Berlin: Springer; 2015.CrossRef Cambria E, Hussain A, Vol. 1. Sentic computing: a common-sense-based framework for concept-level sentiment analysis. Berlin: Springer; 2015.CrossRef
12.
Zurück zum Zitat Chu H, Huang K, Zhang R, Hussian A. Sdrnf: generating scalable and discriminative random nonlinear features from data. Big Data Anal 2016;1(1):10.CrossRef Chu H, Huang K, Zhang R, Hussian A. Sdrnf: generating scalable and discriminative random nonlinear features from data. Big Data Anal 2016;1(1):10.CrossRef
13.
Zurück zum Zitat Malik Z, Hussain A, Wu J. 2015. Extracting online information from dual and multi-data streams (in press). Neural Computation and Applications. Malik Z, Hussain A, Wu J. 2015. Extracting online information from dual and multi-data streams (in press). Neural Computation and Applications.
14.
Zurück zum Zitat Huang K, Yang H, King I, Lyu MR. Machine learning: modeling data locally and globally. Berlin: Springer; 2008. Huang K, Yang H, King I, Lyu MR. Machine learning: modeling data locally and globally. Berlin: Springer; 2008.
15.
Zurück zum Zitat Huang K, Yang H, King I, Lyu MR. Local learning vs. global learning: an introduction to maxi-min margin machine. Support Vector Machines: Theory and Applications. Springer; 2005. p. 113–131. Huang K, Yang H, King I, Lyu MR. Local learning vs. global learning: an introduction to maxi-min margin machine. Support Vector Machines: Theory and Applications. Springer; 2005. p. 113–131.
16.
Zurück zum Zitat Yang H, Huang K, King I, Lyu MR. Maximum margin semi-supervised learning with irrelevant data. Neural Netw 2015;70:90–102.CrossRefPubMed Yang H, Huang K, King I, Lyu MR. Maximum margin semi-supervised learning with irrelevant data. Neural Netw 2015;70:90–102.CrossRefPubMed
17.
18.
Zurück zum Zitat Tao D, Li X, Wu X, Maybank SJ. Geometric mean for subspace selection. IEEE Trans Pattern Anal Mach Intell 2009;31(2):260–274.CrossRefPubMed Tao D, Li X, Wu X, Maybank SJ. Geometric mean for subspace selection. IEEE Trans Pattern Anal Mach Intell 2009;31(2):260–274.CrossRefPubMed
19.
Zurück zum Zitat Liu W, Zha ZJ, Wang Y, Lu K, Tao D. p-laplacian regularized sparse coding for human activity recognition. IEEE Trans Ind Electron 2016;63(8):5120–5129. Liu W, Zha ZJ, Wang Y, Lu K, Tao D. p-laplacian regularized sparse coding for human activity recognition. IEEE Trans Ind Electron 2016;63(8):5120–5129.
20.
Zurück zum Zitat Gourier N, Hall D, Crowley J. Estimating face orientation from robust detection of salient facial features. International conference on pattern recognition (ICPR); 2004. Gourier N, Hall D, Crowley J. Estimating face orientation from robust detection of salient facial features. International conference on pattern recognition (ICPR); 2004.
21.
Zurück zum Zitat Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE 1998;86(11):2278–2324.CrossRef Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE 1998;86(11):2278–2324.CrossRef
22.
Zurück zum Zitat Zhang X-Y, Huang K, Liu C-L. Pattern field classification with style normalized transformation. International joint conference on artificial intelligence (IJCAI); 2011. p. 1621– 1626. Zhang X-Y, Huang K, Liu C-L. Pattern field classification with style normalized transformation. International joint conference on artificial intelligence (IJCAI); 2011. p. 1621– 1626.
23.
Zurück zum Zitat Jolliffe I. Principal component analysis. Oxford: Wiley Online Library; 2002. Jolliffe I. Principal component analysis. Oxford: Wiley Online Library; 2002.
24.
Zurück zum Zitat Tenenbaum JB, Freeman WT. Separating style and content with bilinear models. Neural Comput 2000;12(6):1247–1283.CrossRefPubMed Tenenbaum JB, Freeman WT. Separating style and content with bilinear models. Neural Comput 2000;12(6):1247–1283.CrossRefPubMed
25.
Zurück zum Zitat Sarkar P, Nagy G. Style consistent classification of isogenous patterns. IEEE Trans Pattern Anal Mach Intell Jan 2005;27(1):88–98.CrossRefPubMed Sarkar P, Nagy G. Style consistent classification of isogenous patterns. IEEE Trans Pattern Anal Mach Intell Jan 2005;27(1):88–98.CrossRefPubMed
26.
Zurück zum Zitat Grother PJ. Nist special database 19 handprinted forms and characters database. Maryland: National Institute of Standards and Technology; 1995. Grother PJ. Nist special database 19 handprinted forms and characters database. Maryland: National Institute of Standards and Technology; 1995.
Metadaten
Titel
Learning from Few Samples with Memory Network
verfasst von
Shufei Zhang
Kaizhu Huang
Rui Zhang
Amir Hussain
Publikationsdatum
25.10.2017
Verlag
Springer US
Erschienen in
Cognitive Computation / Ausgabe 1/2018
Print ISSN: 1866-9956
Elektronische ISSN: 1866-9964
DOI
https://doi.org/10.1007/s12559-017-9507-z

Weitere Artikel der Ausgabe 1/2018

Cognitive Computation 1/2018 Zur Ausgabe

Premium Partner