Papers
Topics
Authors
Recent
Search
2000 character limit reached

Deep Causal Generative Models with Property Control

Published 25 May 2024 in cs.LG and stat.ML | (2405.16219v1)

Abstract: Generating data with properties of interest by external users while following the right causation among its intrinsic factors is important yet has not been well addressed jointly. This is due to the long-lasting challenge of jointly identifying key latent variables, their causal relations, and their correlation with properties of interest, as well as how to leverage their discoveries toward causally controlled data generation. To address these challenges, we propose a novel deep generative framework called the Correlation-aware Causal Variational Auto-encoder (C2VAE). This framework simultaneously recovers the correlation and causal relationships between properties using disentangled latent vectors. Specifically, causality is captured by learning the causal graph on latent variables through a structural causal model, while correlation is learned via a novel correlation pooling algorithm. Extensive experiments demonstrate C2VAE's ability to accurately recover true causality and correlation, as well as its superiority in controllable data generation compared to baseline models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Causal regularization. arXiv preprint arXiv:1702.02604, 2017.
  2. Invertible residual networks. In International conference on machine learning, pages 573–582. PMLR, 2019.
  3. Regression transformer enables concurrent sequence regression and generation for molecular language modelling. Nature Machine Intelligence, 5(4):432–444, 2023.
  4. Diffusion self-guidance for controllable image generation. Advances in Neural Information Processing Systems, 36, 2024.
  5. Property controllable variational autoencoder via invertible mutual dependence. In International Conference on Learning Representations, 2020.
  6. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
  7. Amazon-m2: A multilingual multi-locale shopping session dataset for recommendation and text generation. Advances in Neural Information Processing Systems, 36, 2024.
  8. Variational autoencoders and nonlinear ica: A unifying framework. In International Conference on Artificial Intelligence and Statistics, pages 2207–2217. PMLR, 2020.
  9. Semi-supervised learning with deep generative models. Advances in neural information processing systems, 27, 2014.
  10. Learning latent subspaces in variational autoencoders. Advances in neural information processing systems, 31, 2018.
  11. Disentangling factors of variation using few labels. arXiv preprint arXiv:1905.01258, 2019.
  12. dsprites: Disentanglement testing sprites dataset. https://github.com/deepmind/dsprites-dataset/, 2017.
  13. Causal interpretation of self-attention in pre-trained transformers. Advances in Neural Information Processing Systems, 36, 2024.
  14. Weakly supervised disentangled generative causal representation learning. Journal of Machine Learning Research, 23(241):1–55, 2022.
  15. A linear non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(10), 2006.
  16. D’ya like dags? a survey on structure learning and causal discovery. ACM Computing Surveys, 55(4):1–36, 2022.
  17. Multi-objective deep data generation with correlated property control. Advances in Neural Information Processing Systems, 35:28889–28901, 2022.
  18. Controllable data generation by deep learning: A review. ACM Computing Surveys, 56(9):1–38, 2024.
  19. Geometric latent diffusion models for 3d molecule generation. In International Conference on Machine Learning, pages 38592–38610. PMLR, 2023.
  20. Causalvae: Disentangled representation learning via neural structural causal models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9593–9602, 2021.
  21. A survey of controllable text generation using transformer-based pre-trained language models. ACM Computing Surveys, 56(3):1–37, 2023.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 6 likes about this paper.