The Power of Stories: Narrative Priming Shapes How LLM Agents Collaborate and Compete
Abstract: According to Yuval Noah Harari, large-scale human cooperation is driven by shared narratives that encode common beliefs and values. This study explores whether such narratives can similarly nudge LLM agents toward collaboration. We use a finitely repeated public goods game in which LLM agents choose either cooperative or egoistic spending strategies. We prime agents with stories highlighting teamwork to different degrees and test how this influences negotiation outcomes. Our experiments explore four questions:(1) How do narratives influence negotiation behavior? (2) What differs when agents share the same story versus different ones? (3) What happens when the agent numbers grow? (4) Are agents resilient against self-serving negotiators? We find that story-based priming significantly affects negotiation strategies and success rates. Common stories improve collaboration, benefiting each agent. By contrast, priming agents with different stories reverses this effect, and those agents primed toward self-interest prevail. We hypothesize that these results carry implications for multi-agent system design and AI alignment.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Knowledge Gaps
Knowledge gaps, limitations, and open questions
Below is a single, concrete list of what remains missing, uncertain, or unexplored in the paper—articulated so future researchers can act on it.
- Story selection was ad hoc and not systematically controlled; future work should construct a balanced, factorial story set varying plot structure, moral framing, emotional valence, cultural origin, readability, and length to isolate causal features.
- Potential bag-of-words/lexical confounds were not ruled out; design minimally different narratives (e.g., permuted syntax, synonym-swapped, matched-length controls) to test whether effects are driven by specific teamwork-related vocabulary versus deeper semantics.
- The causal mechanism of narrative priming is unclear; use mechanistic interpretability (e.g., representation probing, causal tracing, attention/MLP activation analysis) to identify how narratives shift internal states, value representations, or decision policies.
- The role of RLHF and instruction-tuning is unknown; compare base vs. instruction-tuned vs. RLHF variants across model families to determine whether alignment training amplifies or dampens narrative effects.
- Cross-model generalization was not tested; replicate on diverse LLMs (e.g., GPT-4-family, Mistral, Llama-3.3, smaller models) and different parameter scales to assess robustness and scaling laws of narrative priming.
- Quantization/hardware effects may confound behavior; rerun experiments on full-precision models and varied hardware to confirm that FP8 quantization and deployment specifics do not drive outcomes.
- Temperature and sampling strategy effects are underexplored; run a systematic sweep over temperature, top-p/top-k, and deterministic decoding to quantify sensitivity of cooperation to sampling noise.
- The experimental game parameters were fixed (
m=1.5,T=10,R=5); perform a parameter sweep over multiplierm, endowmentT, horizonR, and discounting to identify thresholds and regimes where narratives matter most. - Observation structure was limited to aggregate contributions; test designs with individual-level observability, reputation tracking, and punish/reward mechanisms to evaluate whether narrative priming interacts with richer social signals.
- Agents had no explicit communication channel; introduce message exchange or negotiation protocols to study whether shared narratives synergize with language-mediated coordination (and whether adversarial narratives exploit communication).
- Strategy identification was not performed; extract and classify agent strategies (e.g., tit-for-tat, threshold, win-stay-lose-shift) from outputs to map narrative prompts to recognizable human/game-theoretic behaviors.
- The collaboration metric focused on contributions; include complementary metrics (e.g., individual and group payoff, fairness/variance, inequality, exploitability, resilience to defection) to capture multiple dimensions of cooperation quality.
- Heterogeneous groups mixed cooperative, control, and self-interest stories; test heterogeneous-but-all-cooperative story sets to disentangle heterogeneity from antagonistic narratives.
- Group composition effects were not analyzed; systematically vary the fraction and identity of selfish vs. cooperative narratives to find tipping points for cooperation collapse and recovery.
- Scaling beyond
N=32and network topology effects were not studied; examine larger populations and different interaction graphs (fully connected vs. sparse networks) to assess scalability and structural constraints. - Temporal persistence and decay of priming were not measured; run longer horizons, multiple consecutive games, and re-priming interventions to test whether effects fade, compound, or require reinforcement.
- Transfer across tasks remains unknown; evaluate whether narratives that help in public goods also help in other social dilemmas (Prisoner’s Dilemma, Ultimatum, bargaining) and multi-issue negotiation benchmarks.
- Narrative placement and prompt design could confound results; compare persona-style priming, moral framing, goal-setting, and story placement (beginning vs. end of prompt) to estimate the most effective and safest priming formats.
- Comprehension and uptake of narratives were not validated; add manipulation checks (e.g., asking agents to summarize or restate the story’s moral) to ensure the model actually internalizes the narrative.
- Cultural and linguistic generalization was not tested; run cross-lingual experiments and culturally diverse narratives to evaluate whether effects depend on language, cultural familiarity, or training corpus biases.
- Adversarial or malicious narratives were not explored; design and evaluate deceptive, divisive, or manipulation-focused stories to quantify risks and develop defenses (e.g., priming hygiene, prompt sanitization).
- Robustness experiments used a single “bad apple” condition; broaden adversarial tests (vary number, behavior patterns, strategic defectors) and measure how narratives affect resilience and recovery mechanisms.
- Statistical reporting in the main text was limited; include formal hypothesis tests, effect sizes, and preregistered analysis plans to strengthen claims and avoid overinterpretation.
- Real-world applicability is speculative; test narrative priming in practical multi-agent tasks (collaborative coding, scheduling, resource allocation with constraints) to bridge the gap from toy games to operational systems.
- Ethical and safety implications of narrative steering were not empirically assessed; perform risk analyses on manipulation, undue influence, and misuse (e.g., weaponized narratives) and propose governance safeguards.
Collections
Sign up for free to add this paper to one or more collections.