Papers
Topics
Authors
Recent
Search
2000 character limit reached

Leave No One Behind: Online Self-Supervised Self-Distillation for Sequential Recommendation

Published 22 Mar 2024 in cs.IR and cs.LG | (2404.07219v2)

Abstract: Sequential recommendation methods play a pivotal role in modern recommendation systems. A key challenge lies in accurately modeling user preferences in the face of data sparsity. To tackle this challenge, recent methods leverage contrastive learning (CL) to derive self-supervision signals by maximizing the mutual information of two augmented views of the original user behavior sequence. Despite their effectiveness, CL-based methods encounter a limitation in fully exploiting self-supervision signals for users with limited behavior data, as users with extensive behaviors naturally offer more information. To address this problem, we introduce a novel learning paradigm, named Online Self-Supervised Self-distillation for Sequential Recommendation ($S4$Rec), effectively bridging the gap between self-supervised learning and self-distillation methods. Specifically, we employ online clustering to proficiently group users by their distinct latent intents. Additionally, an adversarial learning strategy is utilized to ensure that the clustering procedure is not affected by the behavior length factor. Subsequently, we employ self-distillation to facilitate the transfer of knowledge from users with extensive behaviors (teachers) to users with limited behaviors (students). Experiments conducted on four real-world datasets validate the effectiveness of the proposed method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Self-labelling via simultaneous clustering and representation learning. In ICLR.
  2. Unsupervised learning of visual features by contrasting cluster assignments. NeurIPS 33 (2020), 9912–9924.
  3. Wei-Chi Chen and Wei-Ta Chu. 2023. SSSD: Self-Supervised Self Distillation. In WACV. 2769–2776.
  4. Improving end-to-end sequential recommendations with intent-aware diversification. In CIKM. 175–184.
  5. Intent Contrastive Learning for Sequential Recommendation. In WWW. 2172–2182.
  6. Marco Cuturi. 2013. Sinkhorn distances: Lightspeed computation of optimal transport. NeurIPS 26 (2013).
  7. Yaroslav Ganin and Victor S. Lempitsky. 2015. Unsupervised Domain Adaptation by Backpropagation. In ICML, Vol. 37. 1180–1189.
  8. SimCSE: Simple Contrastive Learning of Sentence Embeddings. In EMNLP. 6894–6910.
  9. Generative Adversarial Nets. In NeurIPS. 2672–2680.
  10. Ruining He and Julian McAuley. 2016. Fusing similarity models with markov chains for sparse sequential recommendation. In ICDM. 191–200.
  11. Balázs Hidasi and Alexandros Karatzoglou. 2018. Recurrent neural networks with top-k gains for session-based recommendations. In CIKM. 843–852.
  12. Session-based recommendations with recurrent neural networks. ICLR (2015).
  13. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015).
  14. Improving sequential recommendation with knowledge-enhanced memory networks. In SIGIR. 505–514.
  15. Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In ICDM. IEEE, 197–206.
  16. Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
  17. Intention-aware Sequential Recommendation with Structured Intent Transition : (Extended Abstract). In ICDE. 3759–3760.
  18. Multi-Intention Oriented Contrastive Learning for Sequential Recommendation. In WSDM. 411–419.
  19. Dual Contrastive Network for Sequential Recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’22). Association for Computing Machinery, New York, NY, USA, 2686–2691. https://doi.org/10.1145/3477495.3531918
  20. Contrastive Self-supervised Sequential Recommendation with Robust Augmentation. CoRR abs/2108.06479 (2021).
  21. Augmenting sequential recommendation with pseudo-prior items via reversely pre-training transformer. In SIGIR. 1608–1612.
  22. Disentangled Self-Supervision in Sequential Recommenders. In KDD. 483–491.
  23. Image-Based Recommendations on Styles and Substitutes. In SIGIR. 43–52.
  24. BPR: Bayesian Personalized Ranking from Implicit Feedback. In UAI. 452–461.
  25. Factorizing personalized markov chains for next-basket recommendation. In WWW. 811–820.
  26. BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer. In CIKM. 1441–1450.
  27. Sparse-interest network for sequential recommendation. In WSDM. 598–606.
  28. Jiaxi Tang and Ke Wang. 2018. Personalized top-n sequential recommendation via convolutional sequence embedding. In WSDM. 565–573.
  29. Representation Learning with Contrastive Predictive Coding. CoRR (2018). http://arxiv.org/abs/1807.03748
  30. Attention is All you Need. In NeurIPS. 5998–6008.
  31. Feng Wang and Huaping Liu. 2021. Understanding the Behaviour of Contrastive Loss. In CVPR. 2495–2504.
  32. Tongzhou Wang and Phillip Isola. 2020. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In ICML. 9929–9939.
  33. Self-Supervised Graph Co-Training for Session-based Recommendation. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management (CIKM ’21). Association for Computing Machinery, New York, NY, USA, 2180–2190. https://doi.org/10.1145/3459637.3482388
  34. Contrastive Learning for Sequential Recommendation. In ICDE. 1259–1273.
  35. Snapshot Distillation: Teacher-Student Optimization in One Generation. In CVPR. 2859–2868.
  36. Are graph augmentations necessary? simple graph contrastive learning for recommendation. In SIGIR. 1294–1303.
  37. A simple convolutional generative network for next item recommendation. In WSDM. 582–590.
  38. Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation. In ICCV. 3712–3721.
  39. S3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization. In CIKM. 1893–1902.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 2 likes about this paper.