Papers
Topics
Authors
Recent
Search
2000 character limit reached

Happy People -- Image Synthesis as Black-Box Optimization Problem in the Discrete Latent Space of Deep Generative Models

Published 11 Jun 2023 in cs.CV | (2306.06684v1)

Abstract: In recent years, optimization in the learned latent space of deep generative models has been successfully applied to black-box optimization problems such as drug design, image generation or neural architecture search. Existing models thereby leverage the ability of neural models to learn the data distribution from a limited amount of samples such that new samples from the distribution can be drawn. In this work, we propose a novel image generative approach that optimizes the generated sample with respect to a continuously quantifiable property. While we anticipate absolutely no practically meaningful application for the proposed framework, it is theoretically principled and allows to quickly propose samples at the mere boundary of the training data distribution. Specifically, we propose to use tree-based ensemble models as mathematical programs over the discrete latent space of vector quantized VAEs, which can be globally solved. Subsequent weighted retraining on these queries allows to induce a distribution shift. In lack of a practically relevant problem, we consider a visually appealing application: the generation of happily smiling faces (where the training distribution only contains less happy people) - and show the principled behavior of our approach in terms of improved FID and higher smile degree over baseline approaches.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Leo Breiman. Random forests. Machine Learning, 45(1):5–32, 2001.
  2. A tutorial on bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning, 2010.
  3. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. CoRR, abs/1606.03657, 2016.
  4. Semi-supervised learning with context-conditional generative adversarial networks. CoRR, abs/1611.06430, 2016.
  5. Jerome H. Friedman. Greedy function approximation: A gradient boosting machine. The Annals of Statistics, 29(5):1189 – 1232, 2001.
  6. From variational to deterministic autoencoders. In 8th International Conference on Learning Representations (ICLR), Apr. 2020. *equal contribution.
  7. Automatic chemical design using a data-driven continuous representation of molecules. CoRR, abs/1610.02415, 2016.
  8. Deep residual learning for image recognition, 2015.
  9. Gans trained by a two time-scale update rule converge to a local nash equilibrium, 2018.
  10. Talk-to-edit: Fine-grained facial editing via dialog. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
  11. Junction tree variational autoencoder for molecular graph generation, 2019.
  12. Auto-encoding variational bayes, 2014.
  13. D. Kraft. A software package for sequential quadratic programming. Deutsche Forschungs- und Versuchsanstalt für Luft- und Raumfahrt Köln: Forschungsbericht. Wiss. Berichtswesen d. DFVLR, 1988.
  14. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV), December 2015.
  15. Structured variationally auto-encoded optimization. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 3267–3275. PMLR, 10–15 Jul 2018.
  16. Learning where to look – generative nas is surprisingly efficient. In ECCV, 2022.
  17. Carl Edward Rasmussen. Gaussian processes in machine learning. In Summer school on machine learning, pages 63–71. Springer, 2003.
  18. Taking the human out of the loop: A review of bayesian optimization. Proceedings of the IEEE, 104(1):148–175, 2016.
  19. Scalable bayesian optimization using deep neural networks. In Francis Bach and David Blei, editors, Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pages 2171–2180, Lille, France, 07–09 Jul 2015. PMLR.
  20. Entmoot: A framework for optimization over ensemble tree models. Computers & Chemical Engineering, 151:107343, Aug 2021.
  21. Sample-efficient optimization in the latent space of deep generative models via weighted retraining, 2020.
  22. Neural discrete representation learning, 2018.
  23. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res., 11:3371–3408, dec 2010.
  24. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Processing Letters, 23(10):1499–1503, Oct 2016.
Citations (1)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.