- The paper introduces G-Refer, a novel framework that combines hybrid graph retrieval (path and node level) with Large Language Models to generate personalized and explainable recommendations by effectively integrating structural and semantic collaborative filtering information.
- G-Refer employs a Knowledge Pruning strategy to filter irrelevant data for efficiency and utilizes Retrieval-Augmented Fine-Tuning (RAFT) with an adapter-free approach to integrate retrieved graph knowledge effectively into the LLM during training.
- Experimental evaluations show G-Refer achieves superior performance, especially in recall, on datasets like Yelp and Google-reviews, confirming the benefits of its hybrid retrieval and knowledge pruning components.
The paper "G-Refer: Graph Retrieval-Augmented LLM for Explainable Recommendation" addresses the challenge of providing personalized and interpretable explanations in recommendation systems by integrating collaborations between LLMs and user-item interaction graphs. Traditional methods often face challenges in effectively utilizing the complex structures of graphs to extract Collaborative Filtering (CF) information for generating explanations due to modality discrepancies between structured graph data and natural language.
The proposed G-Refer framework aims to overcome these hurdles using a novel pipeline consisting of three core components:
- Hybrid Graph Retrieval Mechanism: G-Refer introduces a comprehensive retrieval system combining multi-granularity approaches, involving both path-level and node-level retrievers. The path-level retriever utilizes Graph Neural Networks (GNNs) to identify influential paths associated with user-item interactions, effectively capturing structural CF signals crucial for explaining recommendations. Conversely, the node-level retriever leverages semantic similarities between nodes to capture semantic CF information in user and item profiles. Each of these captured signals is then translated into human-readable text using graph translation techniques, which is essential for enhancing the interpretability of LLM-generated explanations.
- Knowledge Pruning Strategy: Recognizing that not all training samples require additional CF signals for effective explanations, knowledge pruning is implemented to eliminate less relevant data, thereby prioritizing training samples that necessitate external knowledge. This filtering process enhances the model’s ability to capitalize on CF information efficiently, while concurrently reducing computational costs due to a smaller training dataset.
- Retrieval-Augmented Fine-Tuning (RAFT): This phase employs lightweight fine-tuning of LLMs with an adapter-free strategy, focusing on the integration of retrieved graph knowledge at the training stage. By fine-tuning with diverse, real-time retrieved CF information, G-Refer bridges the modality gap, enhancing the LLM’s capacity to understand and effectively utilize structured information in generating accurate and personalized recommendations. Specifically, RAFT ensures that LLMs can leverage its internal understanding alongside external CF signals, maximizing explanation completeness while minimizing context misinterpretation or hallucination.
The experimental evaluation of G-Refer demonstrates its superior performance compared to contemporary state-of-the-art methods. The model outperforms others in key performance measures, especially in producing explanations with higher recall, indicating enhanced utilization and incorporation of retrieved CF information. Such methodological advancements make G-Refer particularly effective in datasets like Yelp and Google-reviews where both structural and semantic CF cues are pivotal.
Additionally, the ablation studies emphasize the importance of each component, specifically showcasing the complementary benefits of integrating node and path-level information, as well as the indispensable role of efficient knowledge pruning. Moreover, the scaling tests based on variants of LLMs further reveal that substantial improvements can be achieved even without significantly increasing computational parameters, marking an advancement towards computationally efficient graph-enhanced recommendation systems.
Overall, G-Refer represents a strategic integration of hybrid graph retrieval with LLMs in the field of recommendation systems, facilitating not just effective recommendations but also transparency that is vital for user trust and system reliability. As the framework can be extended across various domains requiring explainability, this study provides meaningful insights into future directions of graph-based and language-model-driven recommendation systems.