Papers
Topics
Authors
Recent
Search
2000 character limit reached

Cold-Start Personalization Approaches

Updated 21 February 2026
  • Cold-start personalization is the task of delivering individualized recommendations for new users or items that lack sufficient interaction data.
  • Key methodologies include profile-based initialization, contextual modeling, meta-learning, and active preference elicitation to estimate latent user preferences.
  • Empirical evaluations show that adaptive meta-learning and prompt tuning can improve recommendation accuracy by 5-15% and achieve sub-300ms adaptation latency in production settings.

Cold-start personalization refers to the task of providing personalized recommendations, responses, or predictions for new users or items with little or no historical interaction data. This scenario is ubiquitous in recommender systems, conversational AI, adaptive learning, and preference modeling, with significant implications for user engagement and system performance. Research from foundational collaborative filtering to recent LLM personalization highlights a diverse array of approaches, theoretical formulations, and practical solutions to address the cold-start problem for both users and items.

1. Formal Problem Definition and Variants

Cold-start personalization arises when a system must deliver individualized outputs for entities (users or items) that lack sufficient historical data for standard collaborative, sequential, or deep models. The principal challenge is inferring a high-dimensional latent preference vector or behavioral signature from strictly limited direct evidence, which undermines the statistical power of conventional models.

Typical variants include:

  • New User (User Cold-Start): No or extremely few interactions for an incoming user. Collaborative filtering shadows such users with a null or unreliable embedding.
  • New Item (Item Cold-Start): No or very little interaction data for a freshly introduced item; classic CF degrades due to an undefined or uniform item embedding.
  • Continuous Cold Start (CoCoS): Even “known” users/items may periodically become "cold" due to sporadic activity or temporal drift (Kiseleva et al., 2016).
  • Few-shot Cold Start: The system must adapt to user/item heterogeneity with only a handful of interactions—often formalized as few-shot meta-learning (Lee et al., 2019, Kim et al., 2023, Zhao et al., 22 Jul 2025).
  • Domain-Shift/New Domain: Transfer to a new domain or context where interaction histories are not directly portable (Sato, 18 Feb 2025).

The core objective is to minimize prediction, ranking, or classification loss with respect to a (potentially latent) personalized target, subject to either zero or severely limited in-domain evidence.

2. Methodological Taxonomy: Approaches and Core Architectures

A wide spectrum of methodological paradigms has been developed for cold-start personalization. These can be systematically grouped as follows:

A. Profile- and Attribute-Based Initialization

  • User/item profiles leveraging demographic, side, or content information are mapped to latent spaces via explicit regression or embedding models. Clusters in the embedding space enable semi-personalized initial recommendations before sufficient behavioral data accrues (Briand et al., 2021, Uyangoda et al., 2019).
  • Feature-weighted aggregation over known item or user attributes can yield robust similarity measures for early-stage recommendations (Uyangoda et al., 2019).

B. Contextual and Session-Based Modeling

  • Contextual profiles are induced via clustering or deep encoding of implicit features (device, time, session metadata), with context-assigned models providing robust segment-level personalization in the absence of longitudinal user trajectories (Kiseleva et al., 2016).

C. Meta-Learning and Few-Shot Personalization

D. Generative and Zero-Shot Embedding via Side Information

  • Dual autoencoder architectures synthesize pseudo-behavioral embeddings for new users directly from attribute embeddings, which are consumed by a downstream model-agnostic recommendation tower (Feng et al., 2021).

E. Active Preference Elicitation and Query Optimization

F. Learning from Online Feedback and Reward Modeling

  • In cold-start LLM personalization, test-time online preference feedback enables reward function learning (dueling bandits for efficient feedback solicitation), steering generation via learned reward models without updating backbone LLM parameters (Qu et al., 29 Sep 2025).

G. Content-Based Approaches and Prompt Tuning for Items/Users

  • For item cold-start, prompt tuning architectures adapt “frozen” content embeddings with lightweight, trainable deltas, maintaining semantic structure while allowing task-specific adaptation (Pembek et al., 25 Jul 2025).
  • Positive feedback (pinnacle) user-IDs and prompt-encoded adapter networks encode cold item characteristics in non-content domains (Jiang et al., 2024).
  • Visual preference modeling with multi-task computer vision and feature-weighted clustering supports cold-start scenario in fashion recommendation (Verma et al., 2020).

3. Mathematical Formulations and Algorithmic Frameworks

Representative mathematical and algorithmic innovations underpinning cold-start personalization include:

Category Key Mathematical/Algorithmic Principle Sample Reference
Meta-learning (MAML) θu=θαθL(Su)\theta'_u = \theta - \alpha \nabla_\theta \mathcal{L}(S_u); optimize outer-loss over query set QuQ_u (Lee et al., 2019, Kim et al., 2023, Yu et al., 2020, Zhao et al., 22 Jul 2025)
Region Elicitation User encoding as confidence region Rt={uRd:uμt2rt}R_t = \{u\in\mathbb{R}^d : \|u-\mu_t\|_2 \leq r_t\} (Nguyen et al., 2024)
Reward Bandits y=argmaxy[logpθ(yx)+ωr(y;ϕT)]y^* = \arg\max_y[ \log p_\theta(y|x) + \omega r(y; \phi_T) ] (Qu et al., 29 Sep 2025)
Distribution Alignment Empirical minimization of W1(μTDT,μS)W_1(\mu_T^{\mathcal{D}_T}, \mu_S) or MMD (Sato, 18 Feb 2025)
Prompt Embeddings (LLM) Z=[P;X]R(k+t)×dZ = [P; X] \in \mathbb{R}^{(k+t)\times d}, meta-optimized over tasks (Zhao et al., 22 Jul 2025)
Bayesian World Model p(θDt)p(θ;ϕ)i=1tp(aiθ,qi)p(\theta | D_t) \propto p(\theta; \phi) \prod_{i=1}^t p(a_i | \theta, q_i) (Bose et al., 16 Feb 2026)
Adaptive Loss Task-state encoder gϕg_\phi outputs weights for loss per rating (Kim et al., 2023)

These approaches unify fast adaptation, region shrinking, probabilistic belief updating, and parameter-efficient prompt learning for both user and item cold-start personalization.

4. Empirical Evaluation and Benchmarks

Quantitative evaluation of cold-start solutions spans synthetic and production-scale settings, including:

  • Standard Public Datasets: MovieLens (1M, 100K, 20M), Amazon, Yelp, BookCrossing, public music (Deezer, NetEase), and e-commerce logs.
  • Metrics: Root Mean Square Error (RMSE), Mean Absolute Error (MAE), normalized Discounted Cumulative Gain (nDCG@K), Hit Rate (HR@K), Mean Reciprocal Rank (MRR), AUC, GAUC.
  • A/B Tests: Improvements in click-through rate (CTR) and product engagement (e.g., watch time) in live systems (Dai et al., 2021, Briand et al., 2021, Feng et al., 2021, Jiang et al., 2024).
  • Cold-Start Simulation Protocols: Few-shot splits (~1–5 support examples), zero-shot user/item assignment, or artificial test domain partitioning for new users/items.

Key empirical findings:

5. Practical Implementations and Industrial Deployment

Production-grade cold-start systems blend modular components for reliability, interpretability, and scalability:

  • Hybridization: Attribute-to-latent mapping, clustering, and cluster-level recommendation lists are fused for semi-personalized cold-start music recommendation (Briand et al., 2021).
  • Plug-in Adaptivity: Techniques such as POSO (Personalized Cold Start Modules) wrap existing deep architectures (MLP, MMoE, attention) with lightweight, user-group-specialized gating, delivering large business uplifts at negligible computational overhead (Dai et al., 2021).
  • Prompt Tuning at Scale: Item-specific prompt modules with pinnacle feedback embeddings yield significant CTR, play time, and like gains on billion-user short-video platforms (Jiang et al., 2024).
  • Robustness: Continuous retraining, modular separation of NER/AP/AF (conversation systems), and simulation for failure modes ensure cold-start solutions generalize over non-stationary online distributions (Kong et al., 2023).

6. Open Challenges, Extensions, and Theoretical Foundations

Cold-start personalization remains at the frontier of machine learning for several reasons:

  • Data Imbalance: Overcoming major/minor user bias, long-tail distributions, and the "submergence of personalization" problem requires fine-grained adaptive learning rates and task-state regularization (Yu et al., 2020, Kim et al., 2023, Dai et al., 2021).
  • Feedback Efficiency: Optimal elicitation with limited user interaction budget mandates principled value-of-information and mutual information-based question selection (Nguyen et al., 2024, Bose et al., 16 Feb 2026).
  • Scalability: Parameter-efficient methods (e.g., few-shot prompt adaptation, Bayesian world-model inference in 10K parameters vs. RL’s 8B) enable practical deployment in edge, on-device, and privacy-sensitive environments (Zhao et al., 22 Jul 2025, Bose et al., 16 Feb 2026).
  • Domain Shift and Transfer: Cross-domain generalization remains a major challenge; domain shift can induce up to 12% accuracy drop in prompt-based meta-learners (Zhao et al., 22 Jul 2025).
  • Interaction Modalities: Extension to richer feedback (multi-point, natural language, bandit losses), multi-modal inputs (vision, text, audio), and longitudinal adaptation over drifting user/item preference distributions are active topics (Qu et al., 29 Sep 2025, Kim et al., 2024, Verma et al., 2020).
  • End-User Agency: User-driven preference transfer algorithms (Pretender) turn the cold-start problem into an empirical distribution alignment task, with tight single-shot regret bounds (Sato, 18 Feb 2025).

Ongoing research is converging toward modular, sample-efficient, and privacy-preserving cold-start personalization that unifies rapid adaptation, robust uncertainty estimation, and explicit modeling of individual and collective structure in user/item interactions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Cold-Start Personalization.