Papers
Topics
Authors
Recent
Search
2000 character limit reached

Phased Data Augmentation for Training a Likelihood-Based Generative Model with Limited Data

Published 22 May 2023 in cs.CV, cs.LG, and eess.IV | (2305.12681v2)

Abstract: Generative models excel in creating realistic images, yet their dependency on extensive datasets for training presents significant challenges, especially in domains where data collection is costly or challenging. Current data-efficient methods largely focus on GAN architectures, leaving a gap in training other types of generative models. Our study introduces "phased data augmentation" as a novel technique that addresses this gap by optimizing training in limited data scenarios without altering the inherent data distribution. By limiting the augmentation intensity throughout the learning phases, our method enhances the model's ability to learn from limited data, thus maintaining fidelity. Applied to a model integrating PixelCNNs with VQ-VAE-2, our approach demonstrates superior performance in both quantitative and qualitative evaluations across diverse datasets. This represents an important step forward in the efficient training of likelihood-based models, extending the usefulness of data augmentation techniques beyond just GANs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. N.T. Tran, V.H. Tran, N.B. Nguyen, T.K. Nguyen, and N.M. Cheung, “On data augmentation for gan training,” IEEE Transactions on Image Processing, vol.30, pp.1882–1897, 2021.
  2. Z. Wang, Z. Dai, B. Póczos, and J. Carbonell, “Characterizing and avoiding negative transfer,” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.11293–11302, 2019.
  3. S. Zhao, Z. Liu, J. Lin, J.Y. Zhu, and S. Han, “Differentiable augmentation for data-efficient gan training,” Advances in Neural Information Processing Systems, vol.33, pp.7559–7570, 2020.
  4. T. Karras, M. Aittala, J. Hellsten, S. Laine, J. Lehtinen, and T. Aila, “Training generative adversarial networks with limited data,” Advances in Neural Information Processing Systems, vol.33, pp.12104– 12114, 2020.
  5. A. Razavi, A. Van den Oord, and O. Vinyals, “Generating diverse high-fidelity images with vq-vae-2,” Advances in neural information processing systems, vol.32, 2019.
  6. I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” Advances in neural information processing systems, vol.27, 2014.
  7. A. Aggarwal, M. Mittal, and G. Battineni, “Generative adversarial network: An overview of theory and applications,” International Journal of Information Management Data Insights, vol.1, no.1, p.100004, 2021.
  8. J. Sohl-Dickstein, E. Weiss, N. Maheswaranathan, and S. Ganguli, “Deep unsupervised learning using nonequilibrium thermodynamics,” International conference on machine learning, pp.2256–2265, PMLR, 2015.
  9. X. Chen, N. Mishra, M. Rohaninejad, and P. Abbeel, “Pixelsnail: An improved autoregressive generative model,” International Conference on Machine Learning, pp.864–872, PMLR, 2018.
  10. T. Karras, S. Laine, and T. Aila, “A style-based generator architecture for generative adversarial networks,” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.4401– 4410, 2019.
  11. Y. Choi, Y. Uh, J. Yoo, and J.W. Ha, “Stargan v2: Diverse image synthesis for multiple domains,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2020.
  12. G. Parmar, R. Zhang, and J.Y. Zhu, “On aliased resizing and surprising subtleties in gan evaluation,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.11410– 11420, 2022.
  13. T. Karras, M. Aittala, S. Laine, E. Härkönen, J. Hellsten, J. Lehtinen, and T. Aila, “Alias-free generative adversarial networks,” Proc. NeurIPS, 2021.
  14. A. Adadi, “A survey on data-efficient algorithms in big data era,” Journal of Big Data, vol.8, no.1, pp.1–54, 2021.
  15. C. Shorten and T.M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” Journal of big data, vol.6, no.1, pp.1–48, 2019.
  16. A. Mikołajczyk and M. Grochowski, “Data augmentation for improving deep learning in image classification problem,” 2018 international interdisciplinary PhD workshop (IIPhDW), pp.117–122, IEEE, 2018.
  17. E.D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q.V. Le, “Autoaugment: Learning augmentation policies from data,” arXiv preprint arXiv:1805.09501, 2018.
  18. H. Jun, R. Child, M. Chen, J. Schulman, A. Ramesh, A. Radford, and I. Sutskever, “Distribution augmentation for generative modeling,” International Conference on Machine Learning, pp.5006–5019, PMLR, 2020.
  19. OpenAI, “ChatGPT (September 25 Version).” https://chat. openai.com/chat, 2023. [Large language model].

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.