Papers
Topics
Authors
Recent
Search
2000 character limit reached

CIRP: Cross-Item Relational Pre-training for Multimodal Product Bundling

Published 2 Apr 2024 in cs.IR and cs.MM | (2404.01735v1)

Abstract: Product bundling has been a prevailing marketing strategy that is beneficial in the online shopping scenario. Effective product bundling methods depend on high-quality item representations, which need to capture both the individual items' semantics and cross-item relations. However, previous item representation learning methods, either feature fusion or graph learning, suffer from inadequate cross-modal alignment and struggle to capture the cross-item relations for cold-start items. Multimodal pre-train models could be the potential solutions given their promising performance on various multimodal downstream tasks. However, the cross-item relations have been under-explored in the current multimodal pre-train models. To bridge this gap, we propose a novel and simple framework Cross-Item Relational Pre-training (CIRP) for item representation learning in product bundling. Specifically, we employ a multimodal encoder to generate image and text representations. Then we leverage both the cross-item contrastive loss (CIC) and individual item's image-text contrastive loss (ITC) as the pre-train objectives. Our method seeks to integrate cross-item relation modeling capability into the multimodal encoder, while preserving the in-depth aligned multimodal semantics. Therefore, even for cold-start items that have no relations, their representations are still relation-aware. Furthermore, to eliminate the potential noise and reduce the computational cost, we harness a relation pruning module to remove the noisy and redundant relations. We apply the item representations extracted by CIRP to the product bundling model ItemKNN, and experiments on three e-commerce datasets demonstrate that CIRP outperforms various leading representation learning methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Personalized Bundle List Recommendation. In WWW. ACM, 60–71.
  2. Bundle Recommendation and Generation With Graph Neural Networks. IEEE Trans. Knowl. Data Eng. 35, 3 (2023), 2326–2340.
  3. Build Your Own Bundle - A Neural Combinatorial Optimization Method. In ACM MM. ACM, 2625–2633.
  4. Personalized Bundle Recommendation in Online Games. In CIKM. ACM, 2381–2388.
  5. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT (1). Association for Computational Linguistics, 4171–4186.
  6. Leveraging Two Types of Global Graph for Sequential Fashion Recommendation. In ICMR. ACM, 73–81.
  7. Personalized fashion outfit generation with user coordination preference learning. Information Processing & Management 60, 5 (2023), 103434.
  8. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR. OpenReview.net.
  9. Enhancing Item-level Bundle Representation for Bundle Recommendation. CoRR abs/2311.16892 (2023).
  10. Zero-shot Item-based Recommendation via Multi-task Product Knowledge Graph Pre-Training. In CIKM. ACM, 483–493.
  11. Multimodal Compatibility Modeling via Exploring the Consistent and Complementary Correlations. In ACM MM. ACM, 2299–2307.
  12. Deep Residual Learning for Image Recognition. In CVPR. IEEE Computer Society, 770–778.
  13. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. In SIGIR. ACM, 639–648.
  14. Consistency-Aware Recommendation for User-Generated Item List Continuation. In WSDM. ACM, 250–258.
  15. Session-based Recommendations with Recurrent Neural Networks. In ICLR (Poster).
  16. GPT-GNN: Generative Pre-Training of Graph Neural Networks. In KDD. ACM, 1857–1867.
  17. KG-FLIP: Knowledge-guided Fashion-domain Language-Image Pre-training for E-commerce. In ACL (industry). Association for Computational Linguistics, 81–88.
  18. Wang-Cheng Kang and Julian J. McAuley. 2018. Self-Attentive Sequential Recommendation. In ICDM. IEEE Computer Society, 197–206.
  19. Thomas N. Kipf and Max Welling. 2017a. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR (Poster). OpenReview.net.
  20. Thomas N. Kipf and Max Welling. 2017b. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR (Poster). OpenReview.net.
  21. BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models. In ICML (Proceedings of Machine Learning Research, Vol. 202). PMLR, 19730–19742.
  22. BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation. In ICML (Proceedings of Machine Learning Research, Vol. 162). PMLR, 12888–12900.
  23. Align before Fuse: Vision and Language Representation Learning with Momentum Distillation. In NeurIPS. 9694–9705.
  24. MealRec: A Meal Recommendation Dataset. CoRR abs/2205.12133 (2022).
  25. Personalized trip recommendation for tourists based on user interests, points of interest visit durations and visit recency. Knowl. Inf. Syst. 54, 2 (2018), 375–406.
  26. Improved Baselines with Visual Instruction Tuning. CoRR abs/2310.03744 (2023).
  27. Visual Instruction Tuning. CoRR abs/2304.08485 (2023).
  28. Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017).
  29. MultiCBR: Multi-view Contrastive Learning for Bundle Recommendation. CoRR abs/2311.16751 (2023).
  30. CrossCBR: Cross-view Contrastive Learning for Bundle Recommendation. In KDD. ACM, 1233–1241.
  31. Leveraging Multimodal Features and Item-level User Feedback for Bundle Construction. In WSDM. ACM, 510–519.
  32. Contrastive Language-Image Pre-Training with Knowledge Graphs. In NeurIPS.
  33. Learning Transferable Visual Models From Natural Language Supervision. In ICML (Proceedings of Machine Learning Research, Vol. 139). PMLR, 8748–8763.
  34. Distillation-Enhanced Graph Masked Autoencoders for Bundle Recommendation. In SIGIR. ACM, 1660–1669.
  35. BPR: Bayesian personalized ranking from implicit feedback. arXiv preprint arXiv:1205.2618 (2012).
  36. Item-based collaborative filtering recommendation algorithms. In WWW. ACM, 285–295.
  37. Modality-Oriented Graph Learning Toward Outfit Compatibility Modeling. IEEE Trans. Multim. 25 (2023), 856–867.
  38. NeuroStylist: Neural Compatibility Modeling for Clothing Matching. In ACM MM. ACM, 753–761.
  39. Revisiting Bundle Recommendation for Intent-aware Product Bundling. ACM Transactions on Recommender Systems (2024).
  40. Revisiting Bundle Recommendation: Datasets, Tasks, Challenges and Opportunities for Intent-aware Product Bundling. In SIGIR. ACM, 2900–2911.
  41. Jiaxi Tang and Ke Wang. 2018. Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding. In WSDM. ACM, 565–573.
  42. Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks. CoRR abs/2208.10442 (2022).
  43. Self-supervised Heterogeneous Graph Neural Network with Co-contrastive Learning. In KDD. ACM, 1726–1736.
  44. FashionKLIP: Enhancing E-Commerce Image-Text Retrieval with Fashion Multi-Modal Conceptual Knowledge Graph. In ACL (industry). Association for Computational Linguistics, 149–158.
  45. Towards Personalized Bundle Creative Generation with Contrastive Non-Autoregressive Decoding. In SIGIR. ACM, 2634–2638.
  46. Strategy-aware Bundle Recommender System. In SIGIR. ACM, 1198–1207.
  47. Self-supervised Graph Learning for Recommendation. In SIGIR. ACM, 726–735.
  48. Multi-View Intent Disentangle Graph Networks for Bundle Recommendation. In AAAI. AAAI Press, 4379–4387.
  49. MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. CoRR abs/2304.10592 (2023).
  50. Knowledge Perceived Multi-modal Pretraining in E-commerce. In ACM Multimedia. ACM, 2744–2752.
Citations (2)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.