Skip to main content
Erschienen in: Neural Processing Letters 2/2022

30.10.2021

A Dynamically Stabilized Recurrent Neural Network

verfasst von: Samer Saab Jr., Yiwei Fu, Asok Ray, Michael Hauser

Erschienen in: Neural Processing Letters | Ausgabe 2/2022

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

This work proposes a novel recurrent neural network architecture, called the Dynamically Stabilized Recurrent Neural Network (DSRNN). The developed DSRNN includes learnable skip-connections across a specified number of time-steps, which allows for a state-space representation of the network’s hidden-state trajectory, and a regularization term is introduced in the loss function in the setting of Lyapunov stability theory. The regularizer enables the placement of eigenvalues of the (linearized) transfer function matrix to desired locations in the complex plane, thereby acting as an internal controller for the hidden-state trajectories. In this way, the DSRNN adjusts the weights of temporal skip-connections to achieve recurrent hidden-state stability, which mitigates the problems of vanishing and exploding gradients. The efficacy of the DSRNN is demonstrated on a forecasting task of a recorded double-pendulum experimental model. The results show that the DSRNN outperforms both the Long Short-Term Memory (LSTM) and vanilla recurrent neural networks, and the relative mean-squared error of the LSTM is reduced by up to \(\sim \)99.64%. The DSRNN also showed comparable results to the LSTM on a classification task of two Lorenz oscillator systems.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Schmidhuber J (2015) Deep learning in neural networks: an overview. Neural Netw 61:85–117CrossRef Schmidhuber J (2015) Deep learning in neural networks: an overview. Neural Netw 61:85–117CrossRef
2.
Zurück zum Zitat Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780CrossRef Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780CrossRef
3.
Zurück zum Zitat Gers FA, Schmidhuber J, Cummins F (1999) Learning to forget: continual prediction with lstm Gers FA, Schmidhuber J, Cummins F (1999) Learning to forget: continual prediction with lstm
4.
Zurück zum Zitat Greff K, Srivastava RK, Koutník J, Steunebrink BR, Schmidhuber J (2016) Lstm: a search space odyssey. IEEE Trans Neural Netw Learn Syst 28(10):2222–2232MathSciNetCrossRef Greff K, Srivastava RK, Koutník J, Steunebrink BR, Schmidhuber J (2016) Lstm: a search space odyssey. IEEE Trans Neural Netw Learn Syst 28(10):2222–2232MathSciNetCrossRef
5.
Zurück zum Zitat Lippi M, Montemurro MA, Degli Esposti M, Cristadoro G (2019) Natural language statistical features of lstm-generated texts. IEEE Trans Neural Netw Learn Syst 30(11):3326–3337CrossRef Lippi M, Montemurro MA, Degli Esposti M, Cristadoro G (2019) Natural language statistical features of lstm-generated texts. IEEE Trans Neural Netw Learn Syst 30(11):3326–3337CrossRef
6.
Zurück zum Zitat Yu X, Wu L, Xu C, Hu Y, Ma C (2019) A novel neural network for solving nonsmooth nonconvex optimization problems. IEEE Trans Neural Netw Learn Syst Yu X, Wu L, Xu C, Hu Y, Ma C (2019) A novel neural network for solving nonsmooth nonconvex optimization problems. IEEE Trans Neural Netw Learn Syst
7.
Zurück zum Zitat Qin S, Xue X (2014) A two-layer recurrent neural network for nonsmooth convex optimization problems. IEEE Trans Neural Netw Learn Syst 26(6):1149–1160MathSciNetCrossRef Qin S, Xue X (2014) A two-layer recurrent neural network for nonsmooth convex optimization problems. IEEE Trans Neural Netw Learn Syst 26(6):1149–1160MathSciNetCrossRef
8.
Zurück zum Zitat Che H, Wang J (2018) A two-timescale duplex neurodynamic approach to biconvex optimization. IEEE Trans Neural Netw Learn Syst 30(8):2503–2514MathSciNetCrossRef Che H, Wang J (2018) A two-timescale duplex neurodynamic approach to biconvex optimization. IEEE Trans Neural Netw Learn Syst 30(8):2503–2514MathSciNetCrossRef
9.
Zurück zum Zitat Chen K, Yao L, Zhang D, Wang X, Chang X, Nie F (2019) A semisupervised recurrent convolutional attention model for human activity recognition. IEEE Trans Neural Netw Learn Syst Chen K, Yao L, Zhang D, Wang X, Chang X, Nie F (2019) A semisupervised recurrent convolutional attention model for human activity recognition. IEEE Trans Neural Netw Learn Syst
10.
Zurück zum Zitat Lipton ZC, Kale D, Wetzel R (2016) Directly modeling missing data in sequences with rnns: improved classification of clinical time series. In: Machine learning for healthcare conference, pp 253–270 Lipton ZC, Kale D, Wetzel R (2016) Directly modeling missing data in sequences with rnns: improved classification of clinical time series. In: Machine learning for healthcare conference, pp 253–270
12.
Zurück zum Zitat Bao G, Peng Y, Zhou X, Gong S (2020) Region stability and stabilization of recurrent neural network with parameter disturbances. Neural Process Lett 52(3):2175–2188CrossRef Bao G, Peng Y, Zhou X, Gong S (2020) Region stability and stabilization of recurrent neural network with parameter disturbances. Neural Process Lett 52(3):2175–2188CrossRef
13.
Zurück zum Zitat Chandran R, Balasubramaniam P (2013) Delay dependent exponential stability for fuzzy recurrent neural networks with interval time-varying delay. Neural Process Lett 37(2):147–161CrossRef Chandran R, Balasubramaniam P (2013) Delay dependent exponential stability for fuzzy recurrent neural networks with interval time-varying delay. Neural Process Lett 37(2):147–161CrossRef
15.
Zurück zum Zitat Bengio Y, Boulanger-Lewandowski N, Pascanu R (2013) Advances in optimizing recurrent networks. In: 2013 IEEE international conference on acoustics, speech and signal processing, pp 8624–8628, IEEE Bengio Y, Boulanger-Lewandowski N, Pascanu R (2013) Advances in optimizing recurrent networks. In: 2013 IEEE international conference on acoustics, speech and signal processing, pp 8624–8628, IEEE
16.
Zurück zum Zitat Jaeger H, Lukoševičius M, Popovici D, Siewert U (2007) Optimization and applications of echo state networks with leaky-integrator neurons. Neural Netw 20(3):335–352CrossRef Jaeger H, Lukoševičius M, Popovici D, Siewert U (2007) Optimization and applications of echo state networks with leaky-integrator neurons. Neural Netw 20(3):335–352CrossRef
17.
Zurück zum Zitat Zhang Y, Chen G, Yu D, Yaco K, Khudanpur S, Glass J (2016) Highway long short-term memory rnns for distant speech recognition. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 5755–5759, IEEE Zhang Y, Chen G, Yu D, Yaco K, Khudanpur S, Glass J (2016) Highway long short-term memory rnns for distant speech recognition. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 5755–5759, IEEE
18.
Zurück zum Zitat Kim J, El-Khamy M, Lee J (2017) Residual lstm: design of a deep recurrent architecture for distant speech recognition. arXiv preprint, arXiv:1701.03360 Kim J, El-Khamy M, Lee J (2017) Residual lstm: design of a deep recurrent architecture for distant speech recognition. arXiv preprint, arXiv:​1701.​03360
19.
Zurück zum Zitat Haviv D, Rivkind A, Barak O (2019) Understanding and controlling memory in recurrent neural networks. In: International conference on machine learning, pp 2663–2671, PMLR Haviv D, Rivkind A, Barak O (2019) Understanding and controlling memory in recurrent neural networks. In: International conference on machine learning, pp 2663–2671, PMLR
20.
Zurück zum Zitat Kalman RE, Bertram JE (1960) Control system analysis and design via the “second method" of lyapunov: I—continuous-time systems. J Basic Eng 82(2):371–393MathSciNetCrossRef Kalman RE, Bertram JE (1960) Control system analysis and design via the “second method" of lyapunov: I—continuous-time systems. J Basic Eng 82(2):371–393MathSciNetCrossRef
21.
Zurück zum Zitat Hauser M, Gunn S, Saab S Jr, Ray A (2019) State-space representations of deep neural networks. Neural Comput 31(3):538–554MathSciNetCrossRef Hauser M, Gunn S, Saab S Jr, Ray A (2019) State-space representations of deep neural networks. Neural Comput 31(3):538–554MathSciNetCrossRef
22.
Zurück zum Zitat Tay Y, Luu AT, Hui SC (2018) Recurrently controlled recurrent networks. In: Bengio S, Wallach H, Larochelle H, Grauman K, Cesa-Bianchi N, Garnett R (eds) Advances in neural information processing systems 31, pp 4736–4748, Curran Associates, Inc Tay Y, Luu AT, Hui SC (2018) Recurrently controlled recurrent networks. In: Bengio S, Wallach H, Larochelle H, Grauman K, Cesa-Bianchi N, Garnett R (eds) Advances in neural information processing systems 31, pp 4736–4748, Curran Associates, Inc
23.
Zurück zum Zitat Asseman A, Kornuta T, Ozcan A (2018) Learning beyond simulated physics Asseman A, Kornuta T, Ozcan A (2018) Learning beyond simulated physics
24.
Zurück zum Zitat de Jesús Serrano-Pérez J, Fernández-Anaya G, Carrillo-Moreno S, Yu W (2021) New results for prediction of chaotic systems using deep recurrent neural networks. Neural Process Lett, pp 1–18 de Jesús Serrano-Pérez J, Fernández-Anaya G, Carrillo-Moreno S, Yu W (2021) New results for prediction of chaotic systems using deep recurrent neural networks. Neural Process Lett, pp 1–18
26.
Zurück zum Zitat Werbos PJ (1990) Backpropagation through time: what it does and how to do it. Proc IEEE 78(10):1550–1560CrossRef Werbos PJ (1990) Backpropagation through time: what it does and how to do it. Proc IEEE 78(10):1550–1560CrossRef
27.
Zurück zum Zitat Pascanu R, Mikolov T, Bengio Y (2013) On the difficulty of training recurrent neural networks. In: International conference on machine learning, pp 1310–1318 Pascanu R, Mikolov T, Bengio Y (2013) On the difficulty of training recurrent neural networks. In: International conference on machine learning, pp 1310–1318
28.
Zurück zum Zitat Ouyang X, Luo Y, Liu J, Liu Y, Bi J, Qiu S (2018) Period analysis of chaotic systems under finite precisions. In: 2018 26th International conference on systems engineering (ICSEng), pp 1–5, IEEE Ouyang X, Luo Y, Liu J, Liu Y, Bi J, Qiu S (2018) Period analysis of chaotic systems under finite precisions. In: 2018 26th International conference on systems engineering (ICSEng), pp 1–5, IEEE
29.
Zurück zum Zitat Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp 249–256 Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp 249–256
31.
Zurück zum Zitat Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M et al (2016) Tensorflow: a system for large-scale machine learning. OSDI 16:265–283 Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M et al (2016) Tensorflow: a system for large-scale machine learning. OSDI 16:265–283
Metadaten
Titel
A Dynamically Stabilized Recurrent Neural Network
verfasst von
Samer Saab Jr.
Yiwei Fu
Asok Ray
Michael Hauser
Publikationsdatum
30.10.2021
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 2/2022
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-021-10676-7

Weitere Artikel der Ausgabe 2/2022

Neural Processing Letters 2/2022 Zur Ausgabe

Neuer Inhalt