Papers
Topics
Authors
Recent
Search
2000 character limit reached

Variational Transfer Learning using Cross-Domain Latent Modulation

Published 31 May 2022 in cs.LG, cs.AI, and cs.CV | (2205.15523v2)

Abstract: To successfully apply trained neural network models to new domains, powerful transfer learning solutions are essential. We propose to introduce a novel cross-domain latent modulation mechanism to a variational autoencoder framework so as to achieve effective transfer learning. Our key idea is to procure deep representations from one data domain and use it to influence the reparameterization of the latent variable of another domain. Specifically, deep representations of the source and target domains are first extracted by a unified inference model and aligned by employing gradient reversal. The learned deep representations are then cross-modulated to the latent encoding of the alternative domain, where consistency constraints are also applied. In the empirical validation that includes a number of transfer learning benchmark tasks for unsupervised domain adaptation and image-to-image translation, our model demonstrates competitive performance, which is also supported by evidence obtained from visualization.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on Knowledge and Data Engineering, vol. 22, no. 10, pp. 1345–1359, 2010.
  2. C. Tan, F. Sun, T. Kong, W. Zhang, C. Yang, and C. Liu, “A survey on deep transfer learning,” in International Conference on Artificial Neural Networks (ICANN), V. Kurková, Y. Manolopoulos, B. Hammer, L. S. Iliadis, and I. Maglogiannis, Eds., 2018, pp. 270–279.
  3. K. R. Weiss, T. M. Khoshgoftaar, and D. Wang, “A survey of transfer learning,” Journal of Big Data, vol. 3, no. 9, pp. 1–40, 2016.
  4. M.-Y. Y. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-image translation networks,” in Advances in Neural Information Processing Systems (NeurIPS), 2017, pp. 700–708.
  5. L. Wang, A. G. Schwing, and S. Lazebnik, “Diverse and accurate image description using a variational auto-encoder with an additive Gaussian encoding space,” in Advances in Neural Information Processing Systems (NeurIPS), 2017, pp. 5757–5767.
  6. E. Schonfeld, S. Ebrahimi, S. Sinha, T. Darrell, and Z. Akata, “Generalized zero-and few-shot learning via aligned variational autoencoders,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 8247–8255.
  7. D. P. Kingma, “Fast gradient-based inference with continuous latent variable models in auxiliary form,” arXiv:1306.0733 [cs.LG], 2013.
  8. F. d. Hollander, “Probability theory: The coupling method,” Leiden University, Lectures Notes-Mathematical Institute, p. 31, 2012.
  9. S. Mahajan, I. Gurevych, and S. Roth, “Latent normalizing flows for many-to-many cross-domain mappings,” in International Conference on Learning Representations (ICLR), 2020.
  10. J. M. Tomczak and M. Welling, “VAE with a vampprior,” in International Conference on Artificial Intelligence and Statistics (AISTATS), ser. PMLR, A. Storkey and F. Perez-Cruz, Eds., vol. 84, 2018, pp. 1214–1223.
  11. M. D. Hoffman and M. J. Johnson, “ELBO surgery: yet another way to carve up the variational evidence lower bound,” in Advances in Neural Information Processing Systems Workshop (NeurIPSW), 2016.
  12. J. Hou, J. D. Deng, S. Cranefield, and X. Ding, “Cross-domain latent modulation for variational transfer learning,” in IEEE Winter Conference on Applications of Computer Vision (WACV), 2021, pp. 3149–3158.
  13. A. B. L. Larsen, S. K. Sønderby, H. Larochelle, and O. Winther, “Autoencoding beyond pixels using a learned similarity metric,” in International Conference on Machine Learning (ICML), ser. PMLR, M. F. Balcan and K. Q. Weinberger, Eds., vol. 48, 2016, pp. 1558–1566.
  14. A. H. Liu, Y. C. Liu, Y. Y. Yeh, and Y. C. F. Wang, “A unified feature disentangler for multi-domain image translation and manipulation,” in Advances in Neural Information Processing Systems (NeurIPS), 2018, pp. 2590–2599.
  15. Z. M. Ziegler and A. M. Rush, “Latent normalizing flows for discrete sequences,” in International Conference on Machine Learning (ICML), ser. PMLR, R. S. Kamalika Chaudhuri, Ed., vol. 97, 2019, pp. 7673–7682.
  16. D. P. Kingma, T. Salimans, R. Jozefowicz, X. Chen, I. Sutskever, and M. Welling, “Improved variational inference with inverse autoregressive flow,” in Advances in Neural Information Processing Systems (NeurIPS), 2016, pp. 4743–4751.
  17. N. Dilokthanakul, P. A. M. Mediano, M. Garnelo, M. C. H. Lee, H. Salimbeni, K. Arulkumaran, and M. Shanahan, “Deep unsupervised clustering with Gaussian mixture variational autoencoders,” arXiv:1611.02648 [cs.LG], 2016.
  18. K. Sohn, X. Yan, and H. Lee, “Learning structured output representation using deep conditional generative models,” in Advances in Neural Information Processing Systems (NeurIPS), 2015, pp. 3483–3491.
  19. L. Yang, N. M. Cheung, J. Li, and J. Fang, “Deep clustering by Gaussian mixture variational autoencoders with graph embedding,” in IEEE International Conference on Computer Vision (ICCV), 2019, pp. 6439–6448.
  20. K. Bousmalis, G. Trigeorgis, N. Silberman, D. Krishnan, and D. Erhan, “Domain separation networks,” in Advances in Neural Information Processing Systems (NeurIPS), 2016, pp. 343–351.
  21. J. Zhang, Y. Huang, Y. Li, W. Zhao, and L. Zhang, “Multi-attribute transfer via disentangled representation,” in AAAI Conference on Artificial Intelligence (AAAI), vol. 33, 2019, pp. 9195–9202.
  22. Z. Feng, A. Zeng, X. Wang, D. Tao, C. Ke, and M. Song, “Dual swap disentangling,” in Advances in Neural Information Processing Systems (NeurIPS), 2018, pp. 5894–5904.
  23. H. Y. Lee, H. Y. Tseng, J. B. Huang, M. Singh, and M. H. Yang, “Diverse image-to-image translation via disentangled representations,” in European Conference on Computer Vision (ECCV), 2018, pp. 36–52.
  24. A. Gonzalez-Garcia, J. van de Weijer, and Y. Bengio, “Image-to-image translation for cross-domain disentanglement,” in Advances in Neural Information Processing Systems (NeurIPS), 2018, pp. 1294–1305.
  25. M. Naseer, S. H. Khan, H. Khan, F. S. Khan, and F. Porikli, “Cross-domain transferability of adversarial perturbations,” in Advances in Neural Information Processing Systems (NeurIPS), 2019.
  26. A. Noguchi and T. Harada, “Image generation from small datasets via batch statistics adaptation,” in IEEE International Conference on Computer Vision (ICCV), 2019, pp. 2750–2758.
  27. Y. Choi, M. Choi, M. Kim, J. W. Ha, S. Kim, and J. Choo, “StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8789–8797.
  28. K. Bousmalis, N. Silberman, D. Dohan, D. Erhan, and D. Krishnan, “Unsupervised pixel-level domain adaptation with generative adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 95–104.
  29. S. Sankaranarayanan, Y. Balaji, C. D. Castillo, and R. Chellappa, “Generate to adapt: Aligning domains using generative adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8503–8512.
  30. J. Hoffman, E. Tzeng, T. Park, J.-Y. Y. Zhu, P. Isola, K. Saenko, A. A. Efros, and T. Darrell, “CyCADA: Cycle-consistent adversarial domain adaptation,” in International Conference on Machine Learning (ICML), ser. PMLR, J. G. Dy and A. Krause, Eds., vol. 80, 2018, pp. 3162–3174.
  31. J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE International Conference on Computer Vision (ICCV), 2017, pp. 2242–2251.
  32. T. Kim, M. Cha, H. Kim, J. K. Lee, and J. Kim, “Learning to discover cross-domain relations with generative adversarial networks,” in International Conference on Machine Learning (ICML), ser. PMLR, Y. W. T. Doina Precup, Ed., vol. 70, 2017, pp. 1857–1865.
  33. D. P. Kingma and M. Welling, “Auto-encoding variational Bayes,” in International Conference on Learning Representations (ICLR), 2014.
  34. J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?” in Advances in Neural Information Processing Systems (NeurIPS), 2014, pp. 3320–3328.
  35. Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. S. Lempitsky, “Domain-adversarial training of neural networks,” Journal of Machine Learning Research, vol. 17, no. 59, pp. 1–35, 2016.
  36. E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 7167–7176.
  37. S. Ben-David, J. Blitzer, K. Crammer, and F. Pereira, “Analysis of representations for domain adaptation,” in Advances in Neural Information Processing Systems (NeurIPS), 2007, pp. 137–144.
  38. S. Ben-David, J. Blitzer, K. Crammer, A. Kulesza, F. Pereira, and J. W. Vaughan, “A theory of learning from different domains,” Machine Learning, vol. 79, pp. 151–175, 2010.
  39. M. Abadi et al., “TensorFlow: Large-scale machine learning on heterogeneous distributed systems,” arXiv:1603.04467 [cs.DC], 2016.
  40. Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998.
  41. Y. Le Cun, L. D. Jackel, B. Boser, J. S. Denker, H. P. Graf, I. Guyon, D. Henderson, R. E. Howard, and W. Hubbard, “Handwritten digit recognition: Applications of neural network chips and automatic learning,” IEEE Communications Magazine, vol. 27, no. 11, pp. 41–46, 1989.
  42. H. Xiao, K. Rasul, and R. Vollgraf, “Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms,” arXiv:1708.07747 [cs.LG], 2017.
  43. S. Hinterstoisser, C. Cagniart, S. Ilic, P. Sturm, N. Navab, P. Fua, and V. Lepetit, “Gradient response maps for real-time detection of texture-less objects,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 5, pp. 876–888, 2012.
  44. P. Wohlhart and V. Lepetit, “Learning descriptors for object recognition and 3d pose estimation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 3109–3118.
  45. A. Yu and K. Grauman, “Semantic jitter: Dense supervision for visual comparisons via synthetic images,” in IEEE International Conference on Computer Vision (ICCV), 2017, pp. 5570–5579.
  46. Z. Liu, P. Luo, X. Wang, and X. Tang, “Deep learning face attributes in the wild,” in IEEE International Conference on Computer Vision (ICCV), 2015, pp. 3730–3738.
  47. Y. C. Liu, Y. Y. Yeh, T. C. Fu, S. D. Wang, W. C. Chiu, and Y. C. F. Wang, “Detach and adapt: Learning cross-domain disentangled deep representation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8867–8876.
  48. J. F. Canny, “A computational approach to edge detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 8, no. 6, pp. 679–698, 1986.
  49. M. Long, Z. Cao, J. Wang, and M. I. Jordan, “Conditional adversarial domain adaptation,” in Advances in Neural Information Processing Systems (NeurIPS), 2018, pp. 1640–1650.
  50. Y. Pan, T. Yao, Y. Li, Y. Wang, C. Ngo, and T. Mei, “Transferrable prototypical networks for unsupervised domain adaptation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 2239–2247.
  51. Y. Zhang, B. Deng, H. Tang, L. Zhang, and K. Jia, “Unsupervised multi-class domain adaptation: Theory, algorithms, and practice,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 5, pp. 2775–2792, 2020.
  52. A. Sharma, T. Kalluri, and M. Chandraker, “Instance level affinity-based transfer for unsupervised domain adaptation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 5361–5371.
  53. J. Hou, X. Ding, J. D. Deng, and S. Cranefield, “Deep adversarial transition learning using cross-grafted generative stacks,” Neural Networks, vol. 149, pp. 172–183, 2022.
  54. L. van der Maaten and G. Hinton, “Visualizing data using t-SNE,” Journal of Machine Learning Research, vol. 9, pp. 2579–2605, 2008.
  55. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: From error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004.
  56. M. Long, Y. Cao, J. Wang, and M. I. Jordan, “Learning transferable features with deep adaptation networks,” in International Conference on Machine Learning (ICML), ser. PMLR, F. Bach and D. Blei, Eds., vol. 37, 2015, pp. 97–105.
  57. X. Peng, Z. Huang, Y. Zhu, and K. Saenko, “Federated adversarial domain adaptation,” in International Conference on Learning Representations (ICLR), 2020.

Summary

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.