Skip to main content

2022 | OriginalPaper | Buchkapitel

Transfer Learning from Audio Domains a Valuable Tool for Structural Health Monitoring

verfasst von : Eleonora M. Tronci, Homayoon Beigi, Maria Q. Feng, Raimondo Betti

Erschienen in: Dynamics of Civil Structures, Volume 2

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Today, the application of artificial neural network tools to define models that mimic the dynamic behavior of structural systems is a wide-spread approach. A fundamental issue in developing these strategies for damage assessment in civil structures is represented by the unbalanced nature of the available databases, which commonly contain plenty of data coming from the structure under healthy operational conditions and very few samples from the system in unhealthy conditions since the structure would have failed by that time. Consequently, the learning task, carried on with standard deep learning approaches, becomes case-dependent and tends to be specialized for a particular structure and for a very limited number of damage scenarios.
This work presents a framework for damage classification in structural systems intended to overcome such limitations. In this methodology, the model is trained to gain knowledge in the learning task from a rich acoustic dataset (source domain), acquiring higher-level features characterizing vibration traits from a rich acoustic dataset. This knowledge is then transferred to a target domain, with much less training data, such as a structural system, in order to classify its structural condition.
The framework starts with constructing a time-delay neural network (TDNN) structure, trained on the VoxCeleb dataset, in the speech domain. The input of the network consists of Cepstral and pitch features extracted from the audio records. Higher-level features, the x-vectors, speaker embeddings, capturing neural outputs of the network’s intermediate layers, are derived and then used to train a probabilistic linear discriminant analysis (PLDA) model to provide a probabilistic discriminant model for speaker comparison. These features collect generic information regarding the source domain and characterize a classification process based on the frequency content of signals, which is not strictly dependent on the original acoustic domain. Because of the non-case-dependent nature of the x-vector embeddings (features), they can be used to train an alternative PLDA model to address a damage classification task, considering vibration measurements coming from a different system, a structural one which represents the target domain. The simulated data from the 12 degrees of freedom benchmark shear-building structure provided by the IASC-ASCE Structural Health Monitoring Group are studied to verify the proposed framework’s effectiveness.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Farrar, C.R., Worden, K.: Structural Health Monitoring: A Machine Learning Perspective. Wiley, Hoboken (2012)CrossRef Farrar, C.R., Worden, K.: Structural Health Monitoring: A Machine Learning Perspective. Wiley, Hoboken (2012)CrossRef
2.
Zurück zum Zitat Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2009)CrossRef Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2009)CrossRef
3.
Zurück zum Zitat Lu, J., Behbood, V., Hao, P., Zuo, H., Xue, S., Zhang, G.: Transfer learning using computational intelligence: a survey. Knowl.-Based Syst. 80, 14–23 (2015)CrossRef Lu, J., Behbood, V., Hao, P., Zuo, H., Xue, S., Zhang, G.: Transfer learning using computational intelligence: a survey. Knowl.-Based Syst. 80, 14–23 (2015)CrossRef
4.
Zurück zum Zitat Bengio, Y.: Deep learning of representations for unsupervised and transfer learning. In: Proceedings of ICML Workshop on Unsupervised and Transfer Learning, pp. 17–36 (2012) Bengio, Y.: Deep learning of representations for unsupervised and transfer learning. In: Proceedings of ICML Workshop on Unsupervised and Transfer Learning, pp. 17–36 (2012)
5.
Zurück zum Zitat Bengio, Y., Bastien, F., Bergeron, A., Boulanger-Lewandowski, N., Breuel, T., Chherawala, Y., Cisse, M., Côté, M., Erhan, D., Eustache, J., et al.: Deep learners benefit more from out-of-distribution examples. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 164–172 (2011) Bengio, Y., Bastien, F., Bergeron, A., Boulanger-Lewandowski, N., Breuel, T., Chherawala, Y., Cisse, M., Côté, M., Erhan, D., Eustache, J., et al.: Deep learners benefit more from out-of-distribution examples. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 164–172 (2011)
7.
Zurück zum Zitat Hinton, G.E., Osindero, S., Teh, Y.-W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527–1554 (2006)MathSciNetCrossRef Hinton, G.E., Osindero, S., Teh, Y.-W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527–1554 (2006)MathSciNetCrossRef
8.
Zurück zum Zitat Erhan, D., Bengio, Y., Courville, A., Manzagol, P.-A., Vincent, P., Bengio, S.: Why does unsupervised pre-training help deep discriminant learning? J. Mach. Learn. Res. 11, 625–660 (2009)MATH Erhan, D., Bengio, Y., Courville, A., Manzagol, P.-A., Vincent, P., Bengio, S.: Why does unsupervised pre-training help deep discriminant learning? J. Mach. Learn. Res. 11, 625–660 (2009)MATH
9.
Zurück zum Zitat Lee, H., Pham, P., Largman, Y., Ng, A.: Unsupervised feature learning for audio classification using convolutional deep belief networks. Adv. Neural Inf. Process. Syst. 22, 1096–1104 (2009) Lee, H., Pham, P., Largman, Y., Ng, A.: Unsupervised feature learning for audio classification using convolutional deep belief networks. Adv. Neural Inf. Process. Syst. 22, 1096–1104 (2009)
10.
Zurück zum Zitat Thomas, S., Seltzer, M.L., Church, K., Hermansky, H.: Deep neural network features and semi-supervised training for low resource speech recognition. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 6704–6708. IEEE, New York (2013) Thomas, S., Seltzer, M.L., Church, K., Hermansky, H.: Deep neural network features and semi-supervised training for low resource speech recognition. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 6704–6708. IEEE, New York (2013)
11.
Zurück zum Zitat Veselỳ, K., Karafiát, M., Grézl, F., Janda, M., Egorova, E.: The language-independent bottleneck features. In: 2012 IEEE Spoken Language Technology Workshop (SLT), pp. 336–341. IEEE, New York (2012) Veselỳ, K., Karafiát, M., Grézl, F., Janda, M., Egorova, E.: The language-independent bottleneck features. In: 2012 IEEE Spoken Language Technology Workshop (SLT), pp. 336–341. IEEE, New York (2012)
12.
Zurück zum Zitat Beigi, H.: Speaker recognition. In: Fundamentals of Speaker Recognition, pp. 543–559. Springer, New York (2011) Beigi, H.: Speaker recognition. In: Fundamentals of Speaker Recognition, pp. 543–559. Springer, New York (2011)
13.
Zurück zum Zitat Balsamo, L., Betti, R., Beigi, H.: A structural health monitoring strategy using cepstral features. J. Sound Vib. 333(19), 4526–4542 (2014)CrossRef Balsamo, L., Betti, R., Beigi, H.: A structural health monitoring strategy using cepstral features. J. Sound Vib. 333(19), 4526–4542 (2014)CrossRef
14.
Zurück zum Zitat Chung, J.S., Nagrani, A., Zisserman, A.: Voxceleb2: deep speaker recognition (2018). Preprint. arXiv:1806.05622 Chung, J.S., Nagrani, A., Zisserman, A.: Voxceleb2: deep speaker recognition (2018). Preprint. arXiv:1806.05622
15.
Zurück zum Zitat Johnson, E.A., Lam, H.-F., Katafygiotis, L.S., Beck, J.L.: Phase I IASC-ASCE structural health monitoring benchmark problem using simulated data. J. Eng. Mech. 130(1), 3–15 (2004)CrossRef Johnson, E.A., Lam, H.-F., Katafygiotis, L.S., Beck, J.L.: Phase I IASC-ASCE structural health monitoring benchmark problem using simulated data. J. Eng. Mech. 130(1), 3–15 (2004)CrossRef
16.
Zurück zum Zitat Snyder, D., Garcia-Romero, D., Sell, G., Povey, D., Khudanpur, S.: X-vectors: robust DNN embeddings for speaker recognition. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5329–5333. IEEE, New York (2018) Snyder, D., Garcia-Romero, D., Sell, G., Povey, D., Khudanpur, S.: X-vectors: robust DNN embeddings for speaker recognition. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5329–5333. IEEE, New York (2018)
17.
Zurück zum Zitat Ananthram, A., Saravanakumar, K.K., Huynh, J., Beigi, H.: Multi-modal emotion detection with transfer learning (2020). Preprint. arXiv:2011.07065 Ananthram, A., Saravanakumar, K.K., Huynh, J., Beigi, H.: Multi-modal emotion detection with transfer learning (2020). Preprint. arXiv:2011.07065
18.
Zurück zum Zitat Povey, D., Ghoshal, A., Boulianne, G., Burget, L., Glembek, O., Goel, N., Hannemann, M., Motlicek, P., Qian, Y., Schwarz, P., et al.: The Kaldi speech recognition toolkit. In: IEEE 2011 Workshop on Automatic Speech Recognition and Understanding. IEEE Signal Processing Society, Piscataway (2011) Povey, D., Ghoshal, A., Boulianne, G., Burget, L., Glembek, O., Goel, N., Hannemann, M., Motlicek, P., Qian, Y., Schwarz, P., et al.: The Kaldi speech recognition toolkit. In: IEEE 2011 Workshop on Automatic Speech Recognition and Understanding. IEEE Signal Processing Society, Piscataway (2011)
19.
Zurück zum Zitat Ioffe, S.: Probabilistic linear discriminant analysis. In: European Conference on Computer Vision, pp. 531–542. Springer, New York (2006) Ioffe, S.: Probabilistic linear discriminant analysis. In: European Conference on Computer Vision, pp. 531–542. Springer, New York (2006)
Metadaten
Titel
Transfer Learning from Audio Domains a Valuable Tool for Structural Health Monitoring
verfasst von
Eleonora M. Tronci
Homayoon Beigi
Maria Q. Feng
Raimondo Betti
Copyright-Jahr
2022
DOI
https://doi.org/10.1007/978-3-030-77143-0_11