Skip to main content

2018 | OriginalPaper | Buchkapitel

Stochasticity-Assisted Training in Artificial Neural Network

verfasst von : Adedamola Wuraola, Nitish Patel

Erschienen in: Neural Information Processing

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Strategically injected noise can speed up convergence during Neural Network training using backpropagation algorithm. Noise injection during Neural Network training have been proven empirically to improve convergence and generalizability. In this work, a new methodology proven to be efficient for speeding up learning convergence using weight noise in Single Layer Feed-forward Network (SLFN) architecture is presented. We present efficient and effective methods in which local minimum entrapment can be avoided. Our proposed controlled introduction of noise is based on 4 proven analytical and experimental methods. We show that criteria-based mini-batch noise injection to the weights during training often outperforms the noiseless weights as well as fixed noise introduction as seen in literature both in network generalization and convergence speed. The effectiveness of this methodology has been empirically shown as well as it achieving on an average 15%–25% improvement in convergence speed when compared to fixed and noiseless networks. The proposed method is evaluated on the MNIST dataset and other datasets from UCI repository. The comparative analysis confirms that the proposed method achieves superior performance regarding convergence speed.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Srivastava, N., Hinton, G.E., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958 (2014)MathSciNetMATH Srivastava, N., Hinton, G.E., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958 (2014)MathSciNetMATH
2.
Zurück zum Zitat Russell, R.: Pruning algorithms-a survey. IEEE Trans. Neural Netw. 4, 740–747 (1993)CrossRef Russell, R.: Pruning algorithms-a survey. IEEE Trans. Neural Netw. 4, 740–747 (1993)CrossRef
3.
Zurück zum Zitat Kartik, A., Osoba, O., Kosko, B.: Noise-enhanced convolutional neural networks. Neural Net. 78, 15–23 (2016)CrossRef Kartik, A., Osoba, O., Kosko, B.: Noise-enhanced convolutional neural networks. Neural Net. 78, 15–23 (2016)CrossRef
4.
Zurück zum Zitat Reed, R., Marks, R.J., Oh, S.: Similarities of error regularization, sigmoid gain scaling, target smoothing, and training with jitter. IEEE Trans. Neural Netw. 6(3), 529–538 (1995)CrossRef Reed, R., Marks, R.J., Oh, S.: Similarities of error regularization, sigmoid gain scaling, target smoothing, and training with jitter. IEEE Trans. Neural Netw. 6(3), 529–538 (1995)CrossRef
5.
Zurück zum Zitat Murray, A.F., Edwards, P.J.: Enhanced MLP performance and fault tolerance resulting from synaptic weight noise during training. IEEE Trans. Neural Netw. 5(5), 792–802 (1994)CrossRef Murray, A.F., Edwards, P.J.: Enhanced MLP performance and fault tolerance resulting from synaptic weight noise during training. IEEE Trans. Neural Netw. 5(5), 792–802 (1994)CrossRef
6.
Zurück zum Zitat Murray, A.F., Edwards, P.J.: Synaptic weight noise during multilayer perceptron training: fault tolerance and training improvements. IEEE Trans. Neural Netw. 4(4), 722–725 (1993)CrossRef Murray, A.F., Edwards, P.J.: Synaptic weight noise during multilayer perceptron training: fault tolerance and training improvements. IEEE Trans. Neural Netw. 4(4), 722–725 (1993)CrossRef
7.
Zurück zum Zitat Jim, K.C., Giles, C.L., Horne, B.G.: Synaptic noise in dynamically-driven recurrent neural networks: convergence and generalization (1998) Jim, K.C., Giles, C.L., Horne, B.G.: Synaptic noise in dynamically-driven recurrent neural networks: convergence and generalization (1998)
8.
Zurück zum Zitat Holmstrom, L., Koistinen, P.: Using additive noise in back-propagation training. IEEE Trans. Neural Netw. 3(1), 24–38 (1992)CrossRef Holmstrom, L., Koistinen, P.: Using additive noise in back-propagation training. IEEE Trans. Neural Netw. 3(1), 24–38 (1992)CrossRef
9.
Zurück zum Zitat Motaz, S., Kurita, T.: Effect of additive noise for multi-layered perceptron with autoencoders. IEICE Trans. Inf. Syst. 100(7), 1494–1504 (2017) Motaz, S., Kurita, T.: Effect of additive noise for multi-layered perceptron with autoencoders. IEICE Trans. Inf. Syst. 100(7), 1494–1504 (2017)
10.
Zurück zum Zitat Di, X., Yu, P.: Multiplicative noise channel in generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1165–1172 (2017) Di, X., Yu, P.: Multiplicative noise channel in generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1165–1172 (2017)
11.
Zurück zum Zitat Chandra, P., Singh, Y.: Regularization and feedforward artificial neural network training with noise. In: IEEE International Joint Conference on Neural Networks (IJCNN), vol. 3, pp. 2366–2371 (2003) Chandra, P., Singh, Y.: Regularization and feedforward artificial neural network training with noise. In: IEEE International Joint Conference on Neural Networks (IJCNN), vol. 3, pp. 2366–2371 (2003)
12.
Zurück zum Zitat An, G.: The effects of adding noise during backpropagation training on a generalization performance. Neural Comput. 8(3), 643–674 (1996)MathSciNetCrossRef An, G.: The effects of adding noise during backpropagation training on a generalization performance. Neural Comput. 8(3), 643–674 (1996)MathSciNetCrossRef
13.
Zurück zum Zitat Jim, K.C., Giles, C.L., Horne, B.G.: An analysis of noise in recurrent neural networks: convergence and generalization. IEEE Trans. Neural Netw. 7(6), 1424–1438 (1996)CrossRef Jim, K.C., Giles, C.L., Horne, B.G.: An analysis of noise in recurrent neural networks: convergence and generalization. IEEE Trans. Neural Netw. 7(6), 1424–1438 (1996)CrossRef
14.
Zurück zum Zitat Max, W., Teh, Y.W.: Bayesian learning via stochastic gradient Langevin dynamics. In: Proceedings of the 28th International Conference on Machine Learning (ICML), pp. 681–688 (2011) Max, W., Teh, Y.W.: Bayesian learning via stochastic gradient Langevin dynamics. In: Proceedings of the 28th International Conference on Machine Learning (ICML), pp. 681–688 (2011)
15.
Zurück zum Zitat Wang, C., Principe, J.C.: Training neural networks with additive noise in the desired signal. IEEE Trans. Neural Networks. 10(6), 1511–1517 (1999)CrossRef Wang, C., Principe, J.C.: Training neural networks with additive noise in the desired signal. IEEE Trans. Neural Networks. 10(6), 1511–1517 (1999)CrossRef
Metadaten
Titel
Stochasticity-Assisted Training in Artificial Neural Network
verfasst von
Adedamola Wuraola
Nitish Patel
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-030-04179-3_52