Skip to main content

2023 | OriginalPaper | Buchkapitel

Out-of-Distribution Generalisation with Symmetry-Based Disentangled Representations

verfasst von : Loek Tonnaer, Mike Holenderski, Vlado Menkovski

Erschienen in: Advances in Intelligent Data Analysis XXI

Verlag: Springer Nature Switzerland

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Learning disentangled representations is suggested to help with generalisation in AI models. This is particularly obvious for combinatorial generalisation, the ability to combine familiar factors to produce new unseen combinations. Disentangling such factors should provide a clear method to generalise to novel combinations, but recent empirical studies suggest that this does not really happen in practice. Disentanglement methods typically assume i.i.d. training and test data, but for combinatorial generalisation we want to generalise towards factor combinations that can be considered out-of-distribution (OOD). There is a misalignment between the distribution of the observed data and the structure that is induced by the underlying factors.
A promising direction to address this misalignment is symmetry-based disentanglement, which is defined as disentangling symmetry transformations that induce a group structure underlying the data. Such a structure is independent of the (observed) distribution of the data and thus provides a sensible language to model OOD factor combinations as well. We investigate the combinatorial generalisation capabilities of a symmetry-based disentanglement model (LSBD-VAE) compared to traditional VAE-based disentanglement models. We observe that both types of models struggle with generalisation in more challenging settings, and that symmetry-based disentanglement appears to show no obvious improvement over traditional disentanglement. However, we also observe that even if LSBD-VAE assigns low likelihood to OOD combinations, the encoder may still generalise well by learning a meaningful mapping reflecting the underlying group structure.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)CrossRef Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)CrossRef
4.
Zurück zum Zitat Caselles-Dupré, H., Ortiz, M.G., Filliat, D.: Symmetry-based disentangled representation learning requires interaction with environments. In: Advances in Neural Information Processing Systems, pp. 4606–4615 (2019) Caselles-Dupré, H., Ortiz, M.G., Filliat, D.: Symmetry-based disentangled representation learning requires interaction with environments. In: Advances in Neural Information Processing Systems, pp. 4606–4615 (2019)
6.
Zurück zum Zitat Higgins, I., et al.: \(\beta \)-VAE: learning basic visual concepts with a constrained variational framework. In: International Conference on Learning Representations (2017) Higgins, I., et al.: \(\beta \)-VAE: learning basic visual concepts with a constrained variational framework. In: International Conference on Learning Representations (2017)
7.
Zurück zum Zitat Kim, H., Mnih, A.: Disentangling by factorising. In: International Conference on Machine Learning, pp. 2649–2658 (2018) Kim, H., Mnih, A.: Disentangling by factorising. In: International Conference on Machine Learning, pp. 2649–2658 (2018)
8.
Zurück zum Zitat Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. In: International Conference on Learning Representations (2014) Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. In: International Conference on Learning Representations (2014)
9.
Zurück zum Zitat Kumar, A., Sattigeri, P., Balakrishnan, A.: Variational inference of disentangled latent concepts from unlabeled observations. In: International Conference on Learning Representations (2018) Kumar, A., Sattigeri, P., Balakrishnan, A.: Variational inference of disentangled latent concepts from unlabeled observations. In: International Conference on Learning Representations (2018)
10.
Zurück zum Zitat Locatello, F., Bauer, S., Lucic, M., Gelly, S., Schölkopf, B., Bachem, O.: Challenging common assumptions in the unsupervised learning of disentangled representations. In: International Conference on Machine Learning (2019) Locatello, F., Bauer, S., Lucic, M., Gelly, S., Schölkopf, B., Bachem, O.: Challenging common assumptions in the unsupervised learning of disentangled representations. In: International Conference on Machine Learning (2019)
12.
Zurück zum Zitat Montero, M.L., Ludwig, C.J., Costa, R.P., Malhotra, G., Bowers, J.: The role of disentanglement in generalisation. In: International Conference on Learning Representations (2020) Montero, M.L., Ludwig, C.J., Costa, R.P., Malhotra, G., Bowers, J.: The role of disentanglement in generalisation. In: International Conference on Learning Representations (2020)
13.
Zurück zum Zitat Painter, M., Prugel-Bennett, A., Hare, J.: Linear disentangled representations and unsupervised action estimation. In: Advances in Neural Information Processing Systems, vol. 33 (2020) Painter, M., Prugel-Bennett, A., Hare, J.: Linear disentangled representations and unsupervised action estimation. In: Advances in Neural Information Processing Systems, vol. 33 (2020)
14.
Zurück zum Zitat Perez Rey, L.A., Menkovski, V., Portegies, J.: Diffusion variational autoencoders. In: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, pp. 2704–2710 (2020) Perez Rey, L.A., Menkovski, V., Portegies, J.: Diffusion variational autoencoders. In: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, pp. 2704–2710 (2020)
15.
Zurück zum Zitat Quessard, R., Barrett, T.D., Clements, W.R.: Learning group structure and disentangled representations of dynamical environments. In: Advances in Neural Information Processing Systems, vol. 33 (2020) Quessard, R., Barrett, T.D., Clements, W.R.: Learning group structure and disentangled representations of dynamical environments. In: Advances in Neural Information Processing Systems, vol. 33 (2020)
16.
Zurück zum Zitat Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. In: Proceedings of the 31st International Conference on International Conference on Machine Learning, vol. 32 (2014) Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. In: Proceedings of the 31st International Conference on International Conference on Machine Learning, vol. 32 (2014)
17.
Zurück zum Zitat Schott, L., et al.: Visual representation learning does not generalize strongly within the same domain. arXiv preprint arXiv:2107.08221 (2021) Schott, L., et al.: Visual representation learning does not generalize strongly within the same domain. arXiv preprint arXiv:​2107.​08221 (2021)
19.
Zurück zum Zitat Tonnaer, L., Pérez Rey, L.A., Menkovski, V., Holenderski, M., Portegies, J.W.: Quantifying and learning linear symmetry-based disentanglement. In: Proceedings of the 39th International Conference on Machine Learning (ICML) (2022). http://arxiv.org/abs/2011.06070 Tonnaer, L., Pérez Rey, L.A., Menkovski, V., Holenderski, M., Portegies, J.W.: Quantifying and learning linear symmetry-based disentanglement. In: Proceedings of the 39th International Conference on Machine Learning (ICML) (2022). http://​arxiv.​org/​abs/​2011.​06070
20.
Zurück zum Zitat Träuble, F., et al.: On disentangled representations learned from correlated data. In: International Conference on Machine Learning, pp. 10401–10412. PMLR (2021) Träuble, F., et al.: On disentangled representations learned from correlated data. In: International Conference on Machine Learning, pp. 10401–10412. PMLR (2021)
Metadaten
Titel
Out-of-Distribution Generalisation with Symmetry-Based Disentangled Representations
verfasst von
Loek Tonnaer
Mike Holenderski
Vlado Menkovski
Copyright-Jahr
2023
DOI
https://doi.org/10.1007/978-3-031-30047-9_34

Premium Partner