Skip to main content
Top

2018 | OriginalPaper | Chapter

Direct Training of Dynamic Observation Noise with UMarineNet

Authors : Stefan Oehmcke, Oliver Zielinski, Oliver Kramer

Published in: Artificial Neural Networks and Machine Learning – ICANN 2018

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Accurate uncertainty predictions are crucial to assess the reliability of a model, especially for neural networks. Part of this uncertainty is the observation noise, which is dynamic in our marine virtual sensor task. Typically, dynamic noise is not trained directly, but approximated through terms in the loss function. Unfortunately, this noise loss function needs to be scaled by a trade-off-parameter to achieve accurate uncertainties. In this paper we propose an upgrade to the existing architecture, which increases interpretability and introduces a novel direct training procedure for dynamic noise modelling. To that end, we train the point prediction model and the noise model separately. We present a new loss function that requires Monte Carlo runs of the model to directly train for the uncertainty prediction accuracy. In an experimental evaluation, we show that in most tested cases the uncertainty prediction is more accurate than the manually tuned trade-off-parameter. Because of the architectural changes we are able to analyze the importance of individual parts of the time series of our prediction.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Badewien, T.H., Zimmer, E., Bartholomä, A., Reuter, R.: Towards continuous long-term measurements of suspended particulate matter (SPM) in turbid coastal waters. Ocean Dyn. 59(2), 227–238 (2009)CrossRef Badewien, T.H., Zimmer, E., Bartholomä, A., Reuter, R.: Towards continuous long-term measurements of suspended particulate matter (SPM) in turbid coastal waters. Ocean Dyn. 59(2), 227–238 (2009)CrossRef
2.
go back to reference Balke, T., et al.: Experimental salt marsh islands: a model system for novel metacommunity experiments. Estuar. Coast. Shelf Sci. 198(Part A), 288–298 (2017)CrossRef Balke, T., et al.: Experimental salt marsh islands: a model system for novel metacommunity experiments. Estuar. Coast. Shelf Sci. 198(Part A), 288–298 (2017)CrossRef
3.
go back to reference Gal, Y.: Uncertainty in deep learning. Ph.D. thesis, University of Cambridge (2016) Gal, Y.: Uncertainty in deep learning. Ph.D. thesis, University of Cambridge (2016)
4.
go back to reference Gal, Y., Ghahramani, Z.: Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In: International Conference on Machine Learning (ICML), pp. 1050–1059 (2016) Gal, Y., Ghahramani, Z.: Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In: International Conference on Machine Learning (ICML), pp. 1050–1059 (2016)
5.
go back to reference Gal, Y., Ghahramani, Z.: A theoretically grounded application of dropout in recurrent NNs. In: Advances in Neural Information Processing Systems: Annual Conference on Neural Information Processing Systems (NIPS), pp. 1019–1027 (2016) Gal, Y., Ghahramani, Z.: A theoretically grounded application of dropout in recurrent NNs. In: Advances in Neural Information Processing Systems: Annual Conference on Neural Information Processing Systems (NIPS), pp. 1019–1027 (2016)
6.
go back to reference Graves, A., Fernández, S., Schmidhuber, J.: Bidirectional LSTM networks for improved phoneme classification and recognition. In: Duch, W., Kacprzyk, J., Oja, E., Zadrożny, S. (eds.) ICANN 2005. LNCS, vol. 3697, pp. 799–804. Springer, Heidelberg (2005). https://doi.org/10.1007/11550907_126CrossRef Graves, A., Fernández, S., Schmidhuber, J.: Bidirectional LSTM networks for improved phoneme classification and recognition. In: Duch, W., Kacprzyk, J., Oja, E., Zadrożny, S. (eds.) ICANN 2005. LNCS, vol. 3697, pp. 799–804. Springer, Heidelberg (2005). https://​doi.​org/​10.​1007/​11550907_​126CrossRef
7.
go back to reference He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778. IEEE (2016) He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778. IEEE (2016)
8.
go back to reference Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef
9.
go back to reference Huang, G., Liu, Z., van der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2261–2269. IEEE (2017) Huang, G., Liu, Z., van der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2261–2269. IEEE (2017)
10.
go back to reference Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: Squeezenet: Alexnet-level accuracy with 50x fewer parameters and 0.5mb model size. CoRR abs/1602.07360 (2016) Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: Squeezenet: Alexnet-level accuracy with 50x fewer parameters and 0.5mb model size. CoRR abs/1602.07360 (2016)
11.
go back to reference Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning (ICML), pp. 448–456. International Machine Learning Society (2015) Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning (ICML), pp. 448–456. International Machine Learning Society (2015)
12.
go back to reference Lin, M., Chen, Q., Yan, S.: Network in network. CoRR abs/1312.4400 (2013) Lin, M., Chen, Q., Yan, S.: Network in network. CoRR abs/1312.4400 (2013)
13.
go back to reference Murata, S., Masuda, W., Tomioka, S., Ogata, T., Sugano, S.: Mixing actual and predicted sensory states based on uncertainty estimation for flexible and robust robot behavior. In: International Conference on Artificial NNs - (ICANN), pp. 11–18 (2017)CrossRef Murata, S., Masuda, W., Tomioka, S., Ogata, T., Sugano, S.: Mixing actual and predicted sensory states based on uncertainty estimation for flexible and robust robot behavior. In: International Conference on Artificial NNs - (ICANN), pp. 11–18 (2017)CrossRef
14.
go back to reference Oehmcke, S., Zielinski, O., Kramer, O.: Input quality aware convolutional LSTM networks for virtual marine sensors. Neurocomputing 275, 2603–2615 (2017)CrossRef Oehmcke, S., Zielinski, O., Kramer, O.: Input quality aware convolutional LSTM networks for virtual marine sensors. Neurocomputing 275, 2603–2615 (2017)CrossRef
15.
go back to reference Oehmcke, S., Zielinski, O., Kramer, O.: Rnns and exponential PAA for virtual marine sensors. In: International Joint Conference on NNs (IJCNN), pp. 4459–4466. IEEE (2017) Oehmcke, S., Zielinski, O., Kramer, O.: Rnns and exponential PAA for virtual marine sensors. In: International Joint Conference on NNs (IJCNN), pp. 4459–4466. IEEE (2017)
16.
go back to reference Romeu, P., Zamora-Martínez, F., Botella-Rocamora, P., Pardo, J.: Time-series forecasting of indoor temperature using pre-trained deep neural networks. In: Mladenov, V., Koprinkova-Hristova, P., Palm, G., Villa, A.E.P., Appollini, B., Kasabov, N. (eds.) ICANN 2013. LNCS, vol. 8131, pp. 451–458. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-40728-4_57CrossRef Romeu, P., Zamora-Martínez, F., Botella-Rocamora, P., Pardo, J.: Time-series forecasting of indoor temperature using pre-trained deep neural networks. In: Mladenov, V., Koprinkova-Hristova, P., Palm, G., Villa, A.E.P., Appollini, B., Kasabov, N. (eds.) ICANN 2013. LNCS, vol. 8131, pp. 451–458. Springer, Heidelberg (2013). https://​doi.​org/​10.​1007/​978-3-642-40728-4_​57CrossRef
17.
go back to reference Srivastava, N., Hinton, G.E., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: A simple way to prevent nns from overfitting. J. Mach. Learn. Res. (JLMR) 15(1), 1929–1958 (2014)MATH Srivastava, N., Hinton, G.E., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: A simple way to prevent nns from overfitting. J. Mach. Learn. Res. (JLMR) 15(1), 1929–1958 (2014)MATH
Metadata
Title
Direct Training of Dynamic Observation Noise with UMarineNet
Authors
Stefan Oehmcke
Oliver Zielinski
Oliver Kramer
Copyright Year
2018
DOI
https://doi.org/10.1007/978-3-030-01418-6_13

Premium Partner