CIRP: Cross-Item Relational Pre-training for Multimodal Product Bundling
Abstract: Product bundling has been a prevailing marketing strategy that is beneficial in the online shopping scenario. Effective product bundling methods depend on high-quality item representations, which need to capture both the individual items' semantics and cross-item relations. However, previous item representation learning methods, either feature fusion or graph learning, suffer from inadequate cross-modal alignment and struggle to capture the cross-item relations for cold-start items. Multimodal pre-train models could be the potential solutions given their promising performance on various multimodal downstream tasks. However, the cross-item relations have been under-explored in the current multimodal pre-train models. To bridge this gap, we propose a novel and simple framework Cross-Item Relational Pre-training (CIRP) for item representation learning in product bundling. Specifically, we employ a multimodal encoder to generate image and text representations. Then we leverage both the cross-item contrastive loss (CIC) and individual item's image-text contrastive loss (ITC) as the pre-train objectives. Our method seeks to integrate cross-item relation modeling capability into the multimodal encoder, while preserving the in-depth aligned multimodal semantics. Therefore, even for cold-start items that have no relations, their representations are still relation-aware. Furthermore, to eliminate the potential noise and reduce the computational cost, we harness a relation pruning module to remove the noisy and redundant relations. We apply the item representations extracted by CIRP to the product bundling model ItemKNN, and experiments on three e-commerce datasets demonstrate that CIRP outperforms various leading representation learning methods.
- Personalized Bundle List Recommendation. In WWW. ACM, 60–71.
- Bundle Recommendation and Generation With Graph Neural Networks. IEEE Trans. Knowl. Data Eng. 35, 3 (2023), 2326–2340.
- Build Your Own Bundle - A Neural Combinatorial Optimization Method. In ACM MM. ACM, 2625–2633.
- Personalized Bundle Recommendation in Online Games. In CIKM. ACM, 2381–2388.
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT (1). Association for Computational Linguistics, 4171–4186.
- Leveraging Two Types of Global Graph for Sequential Fashion Recommendation. In ICMR. ACM, 73–81.
- Personalized fashion outfit generation with user coordination preference learning. Information Processing & Management 60, 5 (2023), 103434.
- An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR. OpenReview.net.
- Enhancing Item-level Bundle Representation for Bundle Recommendation. CoRR abs/2311.16892 (2023).
- Zero-shot Item-based Recommendation via Multi-task Product Knowledge Graph Pre-Training. In CIKM. ACM, 483–493.
- Multimodal Compatibility Modeling via Exploring the Consistent and Complementary Correlations. In ACM MM. ACM, 2299–2307.
- Deep Residual Learning for Image Recognition. In CVPR. IEEE Computer Society, 770–778.
- LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. In SIGIR. ACM, 639–648.
- Consistency-Aware Recommendation for User-Generated Item List Continuation. In WSDM. ACM, 250–258.
- Session-based Recommendations with Recurrent Neural Networks. In ICLR (Poster).
- GPT-GNN: Generative Pre-Training of Graph Neural Networks. In KDD. ACM, 1857–1867.
- KG-FLIP: Knowledge-guided Fashion-domain Language-Image Pre-training for E-commerce. In ACL (industry). Association for Computational Linguistics, 81–88.
- Wang-Cheng Kang and Julian J. McAuley. 2018. Self-Attentive Sequential Recommendation. In ICDM. IEEE Computer Society, 197–206.
- Thomas N. Kipf and Max Welling. 2017a. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR (Poster). OpenReview.net.
- Thomas N. Kipf and Max Welling. 2017b. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR (Poster). OpenReview.net.
- BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models. In ICML (Proceedings of Machine Learning Research, Vol. 202). PMLR, 19730–19742.
- BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation. In ICML (Proceedings of Machine Learning Research, Vol. 162). PMLR, 12888–12900.
- Align before Fuse: Vision and Language Representation Learning with Momentum Distillation. In NeurIPS. 9694–9705.
- MealRec: A Meal Recommendation Dataset. CoRR abs/2205.12133 (2022).
- Personalized trip recommendation for tourists based on user interests, points of interest visit durations and visit recency. Knowl. Inf. Syst. 54, 2 (2018), 375–406.
- Improved Baselines with Visual Instruction Tuning. CoRR abs/2310.03744 (2023).
- Visual Instruction Tuning. CoRR abs/2304.08485 (2023).
- Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017).
- MultiCBR: Multi-view Contrastive Learning for Bundle Recommendation. CoRR abs/2311.16751 (2023).
- CrossCBR: Cross-view Contrastive Learning for Bundle Recommendation. In KDD. ACM, 1233–1241.
- Leveraging Multimodal Features and Item-level User Feedback for Bundle Construction. In WSDM. ACM, 510–519.
- Contrastive Language-Image Pre-Training with Knowledge Graphs. In NeurIPS.
- Learning Transferable Visual Models From Natural Language Supervision. In ICML (Proceedings of Machine Learning Research, Vol. 139). PMLR, 8748–8763.
- Distillation-Enhanced Graph Masked Autoencoders for Bundle Recommendation. In SIGIR. ACM, 1660–1669.
- BPR: Bayesian personalized ranking from implicit feedback. arXiv preprint arXiv:1205.2618 (2012).
- Item-based collaborative filtering recommendation algorithms. In WWW. ACM, 285–295.
- Modality-Oriented Graph Learning Toward Outfit Compatibility Modeling. IEEE Trans. Multim. 25 (2023), 856–867.
- NeuroStylist: Neural Compatibility Modeling for Clothing Matching. In ACM MM. ACM, 753–761.
- Revisiting Bundle Recommendation for Intent-aware Product Bundling. ACM Transactions on Recommender Systems (2024).
- Revisiting Bundle Recommendation: Datasets, Tasks, Challenges and Opportunities for Intent-aware Product Bundling. In SIGIR. ACM, 2900–2911.
- Jiaxi Tang and Ke Wang. 2018. Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding. In WSDM. ACM, 565–573.
- Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks. CoRR abs/2208.10442 (2022).
- Self-supervised Heterogeneous Graph Neural Network with Co-contrastive Learning. In KDD. ACM, 1726–1736.
- FashionKLIP: Enhancing E-Commerce Image-Text Retrieval with Fashion Multi-Modal Conceptual Knowledge Graph. In ACL (industry). Association for Computational Linguistics, 149–158.
- Towards Personalized Bundle Creative Generation with Contrastive Non-Autoregressive Decoding. In SIGIR. ACM, 2634–2638.
- Strategy-aware Bundle Recommender System. In SIGIR. ACM, 1198–1207.
- Self-supervised Graph Learning for Recommendation. In SIGIR. ACM, 726–735.
- Multi-View Intent Disentangle Graph Networks for Bundle Recommendation. In AAAI. AAAI Press, 4379–4387.
- MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. CoRR abs/2304.10592 (2023).
- Knowledge Perceived Multi-modal Pretraining in E-commerce. In ACM Multimedia. ACM, 2744–2752.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.