Skip to main content
Top

2021 | OriginalPaper | Chapter

Variational Hyper-encoding Networks

Authors : Phuoc Nguyen, Truyen Tran, Sunil Gupta, Santu Rana, Hieu-Chi Dam, Svetha Venkatesh

Published in: Machine Learning and Knowledge Discovery in Databases. Research Track

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

We propose a framework called HyperVAE for encoding distributions of distributions. When a target distribution is modeled by a VAE, its neural network parameters are sampled from a distribution in the model space modeled by a hyper-level VAE. We propose a variational inference framework to implicitly encode the parameter distributions into a low dimensional Gaussian distribution. Given a target distribution, we predict the posterior distribution of the latent code, then use a matrix-network decoder to generate a posterior distribution for the parameters. HyperVAE can encode the target parameters in full in contrast to common hyper-networks practices, which generate only the scale and bias vectors to modify the target-network parameters. Thus HyperVAE preserves information about the model for each task in the latent space. We derive the training objective for HyperVAE using the minimum description length (MDL) principle to reduce the complexity of HyperVAE. We evaluate HyperVAE in density estimation tasks, outlier detection and discovery of novel design classes, demonstrating its efficacy.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
1
This is not the same as zero-shot learning where label description is available.
 
2
We use \(\theta =(\theta _{p},\theta _{q})\) to denote the set of parameters for p and q.
 
3
We assume a Dirac delta distribution for \(\gamma \), i.e. a point estimate, in this study.
 
4
We abused the notation and use p to denote both a density and a probability mass function. Bits-back coding is applicable to continuous distributions [10].
 
5
We assumed a matrix multiplication takes O(1) time in GPU.
 
6
Batched matrix multiplication can be paralleled in GPU.
 
Literature
2.
5.
go back to reference Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 1126–1135 (2017). JMLR.org Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 1126–1135 (2017). JMLR.​org
6.
go back to reference Finn, C., Levine, S.: Meta-learning and universality: deep representations and gradient descent can approximate any learning algorithm. In: ICLR (2018) Finn, C., Levine, S.: Meta-learning and universality: deep representations and gradient descent can approximate any learning algorithm. In: ICLR (2018)
7.
go back to reference Gómez-Bombarelli, R., et al.: Automatic chemical design using a data-driven continuous representation of molecules. ACS Central Sci. 4(2), 268–276 (2018)CrossRef Gómez-Bombarelli, R., et al.: Automatic chemical design using a data-driven continuous representation of molecules. ACS Central Sci. 4(2), 268–276 (2018)CrossRef
8.
go back to reference Grant, E., Finn, C., Levine, S., Darrell, T., Griffiths, T.: Recasting gradient-based meta-learning as hierarchical Byes. arXiv preprint arXiv:1801.08930 (2018) Grant, E., Finn, C., Levine, S., Darrell, T., Griffiths, T.: Recasting gradient-based meta-learning as hierarchical Byes. arXiv preprint arXiv:​1801.​08930 (2018)
10.
go back to reference Hinton, G., Van Camp, D.: Keeping neural networks simple by minimizing the description length of the weights. In: Proceedings of the 6th Annual ACM Conference on Computational Learning Theory. Citeseer (1993) Hinton, G., Van Camp, D.: Keeping neural networks simple by minimizing the description length of the weights. In: Proceedings of the 6th Annual ACM Conference on Computational Learning Theory. Citeseer (1993)
12.
go back to reference Krueger, D., Huang, C.-W., Islam, R., Turner, R., Lacoste, A., Courville, A.: Bayesian hypernetworks. arXiv preprint arXiv:1710.04759 (2017) Krueger, D., Huang, C.-W., Islam, R., Turner, R., Lacoste, A., Courville, A.: Bayesian hypernetworks. arXiv preprint arXiv:​1710.​04759 (2017)
13.
go back to reference Le, H., Tran, T., Nguyen, T., Venkatesh, S.: Variational memory encoder-decoder. In: NeurIPS (2018) Le, H., Tran, T., Nguyen, T., Venkatesh, S.: Variational memory encoder-decoder. In: NeurIPS (2018)
14.
go back to reference Mishra, N., Rohaninejad, M., Chen, X., Abbeel, P.: A simple neural attentive meta-learner. In: ICLR 2018 (2018) Mishra, N., Rohaninejad, M., Chen, X., Abbeel, P.: A simple neural attentive meta-learner. In: ICLR 2018 (2018)
15.
go back to reference Nguyen, C., Li, Y., Bui, T.D., Turner, R.E.: Variational continual learning. In: ICLR (2018) Nguyen, C., Li, Y., Bui, T.D., Turner, R.E.: Variational continual learning. In: ICLR (2018)
16.
go back to reference Nguyen, P., Tran, T., Gupta, S., Rana, S., Barnett, M., Venkatesh, S.: Incomplete conditional density estimation for fast materials discovery. In: Proceedings of the 2019 SIAM International Conference on Data Mining, pp. 549–557. SIAM (2019) Nguyen, P., Tran, T., Gupta, S., Rana, S., Barnett, M., Venkatesh, S.: Incomplete conditional density estimation for fast materials discovery. In: Proceedings of the 2019 SIAM International Conference on Data Mining, pp. 549–557. SIAM (2019)
17.
go back to reference Rao, D., Visin, F., Rusu, A., Pascanu, R., Teh, Y.W., Hadsell, R.: Continual unsupervised representation learning. In: Advances in Neural Information Processing Systems, pp. 7645–7655 (2019) Rao, D., Visin, F., Rusu, A., Pascanu, R., Teh, Y.W., Hadsell, R.: Continual unsupervised representation learning. In: Advances in Neural Information Processing Systems, pp. 7645–7655 (2019)
18.
go back to reference Ratzlaff, N., Fuxin, L.: HyperGAN: a generative model for diverse. In: Performant Neural Networks, ICML (2019) Ratzlaff, N., Fuxin, L.: HyperGAN: a generative model for diverse. In: Performant Neural Networks, ICML (2019)
19.
go back to reference Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. arXiv preprint arXiv:1401.4082 (2014) Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. arXiv preprint arXiv:​1401.​4082 (2014)
20.
go back to reference Shahriari, B., Swersky, K., Wang, Z., Adams, R.P., De Freitas, N.: Taking the human out of the loop: a review of Bayesian optimization. Proc. IEEE 104(1), 148–175 (2016)CrossRef Shahriari, B., Swersky, K., Wang, Z., Adams, R.P., De Freitas, N.: Taking the human out of the loop: a review of Bayesian optimization. Proc. IEEE 104(1), 148–175 (2016)CrossRef
22.
go back to reference Townsend, J., Bird, T., Barber, D.: Practical lossless compression with latent variables using bits back coding. arXiv preprint arXiv:1901.04866 (2019) Townsend, J., Bird, T., Barber, D.: Practical lossless compression with latent variables using bits back coding. arXiv preprint arXiv:​1901.​04866 (2019)
23.
go back to reference Wang, K.-C., Vicol, P., Lucas, J., Gu, L., Grosse, R., Zemel, R.: Adversarial distillation of Bayesian neural network posteriors. In: International Conference on Machine Learning, pp. 5177–5186 (2018) Wang, K.-C., Vicol, P., Lucas, J., Gu, L., Grosse, R., Zemel, R.: Adversarial distillation of Bayesian neural network posteriors. In: International Conference on Machine Learning, pp. 5177–5186 (2018)
24.
go back to reference Yoon, J., Kim, T., Dia, O., Kim, S., Bengio, Y., Ahn, S.: Bayesian model-agnostic meta-learning. In: Advances in Neural Information Processing Systems, pp. 7332–7342 (2018) Yoon, J., Kim, T., Dia, O., Kim, S., Bengio, Y., Ahn, S.: Bayesian model-agnostic meta-learning. In: Advances in Neural Information Processing Systems, pp. 7332–7342 (2018)
Metadata
Title
Variational Hyper-encoding Networks
Authors
Phuoc Nguyen
Truyen Tran
Sunil Gupta
Santu Rana
Hieu-Chi Dam
Svetha Venkatesh
Copyright Year
2021
DOI
https://doi.org/10.1007/978-3-030-86520-7_7

Premium Partner