Papers
Topics
Authors
Recent
Search
2000 character limit reached

Role-Based Authority Bias in AI

Updated 15 January 2026
  • Role-Based Authority Bias is the systematic tendency for systems and operators to confer undue legitimacy based on designated roles.
  • Empirical studies quantify the bias using metrics like agreement rate and Cohen’s kappa, showing significant shifts in LLM performance under authority cues.
  • Mitigation strategies, including tailored system prompts and in-context learning, help decrease bias and enhance decision accuracy in multi-agent settings.

Role-based authority bias denotes the systematic tendency for both generative AI systems and their human operators to confer disproportionate epistemic legitimacy on information or decisions emanating from particular institutional or functional roles, such as "expert," "judge," or "autonomous agent." This bias operates via explicit or implicit modalities—traditional, charismatic, legal-rational, rational-technical, and agentic-technical—and shapes both the content curation workflow and downstream decision-making in multi-agent and evaluative settings. Underlying frameworks draw on Weberian authority types, Foucault’s power/knowledge thesis, and Floridi’s Dataism, situating modern AI mediation as a secular extension of historical gatekeeping mechanisms and epistemic exclusion. The effect and amplification of role-based authority bias have been empirically documented across LLM-as-judge, multi-agent debate, content moderation, delegation protocols, adversarial jailbreak settings, and performance benchmarks.

1. Role-Based Authority Bias: Modalities and Definitions

Role-based authority bias originates when information or recommendations are weighted preferentially based on the assigned role or persona, not on inherent accuracy or subject-matter grounding. In contemporary GenAI and LLMs, this bias is conditioned by several modalities (Torkestani et al., 27 Nov 2025):

  • Traditional authority: Deference to customary roles, e.g., religious or academic institutions.
  • Charismatic authority: Legitimacy conferred through mythologizing designers or systems themselves.
  • Legal-rational authority: Trust placed in codified rules, documentation, or regulatory status.
  • Rational-technical authority: Deference to statistical optimization, leaderboard scores, benchmark results.
  • Agentic-technical authority: Epistemic legitimacy assigned to semi-autonomous, planning agents—shifting accountability to the human–machine assemblage.

The authority modality determines the topology of bias in decision outputs, moderation, appeals processes, and epistemic exclusion. Formally, the role-based filtering in GenAI can be expressed as the thresholding of outputs via a policy function:

Sauthorized={sSall:φP(s,R,D)τ}S_{\text{authorized}} = \{ s \in S_{\text{all}} : \varphi_P(s, R, D) \geq \tau \}

where RR denotes role-legitimacy weightings, DD is data provenance, and τ\tau is the operational cutoff (Torkestani et al., 27 Nov 2025).

2. Formalization and Empirical Measurement in LLMs

Quantification draws on multi-agent evaluation and LLM-as-judge paradigms, notably using metrics such as agreement rate and Cohen’s kappa (Choi et al., 8 Jan 2026, Wang et al., 14 Apr 2025):

  • Each authoritative role rr is tagged by power-type indicator π(r){legitimate,referent,expert}\pi(r) \in \{\text{legitimate}, \text{referent}, \text{expert}\}.
  • In a 12-turn debate framework, if the authority role’s initial position (turn 2) is non-neutral, downstream agent agreement rate AtA_t is elevated:

At=1Ni=1N1(SGP,i(t)=SAuth,i(2)),t{1,4,8,12}A_t = \frac{1}{N} \sum_{i=1}^N \mathbf{1}\left(S^{(t)}_{\text{GP},i} = S^{(2)}_{\text{Auth},i}\right), \quad t \in \{1,4,8,12\}

  • Authority bias is operationally detected when At>A1A_t > A_1 and κt>κ1\kappa_t > \kappa_1 for t=4,8,12t=4,8,12.

Key empirical findings establish that expert and referent power roles exert distinctly stronger influences than legitimate roles, and explicit position statements are necessary for bias induction (Choi et al., 8 Jan 2026). LRMs exhibit similar susceptibility, with accuracy and robustness rate (RR) metrics shifting significantly upon injection of authority cues—even when those cues are deliberately spurious or misattributed (Wang et al., 14 Apr 2025). Table summaries:

Model Acc_no_cue Acc_auth_cue ΔAcc (DPO tasks)
GPT-4o 0.66 0.80 +0.14
DS-R1 0.68 0.81 +0.13

In factual tasks, the same authority cue causes an accuracy loss (average –0.14 across models), indicating over-reliance on apparent expertise.

3. Mechanisms and Amplification Pathways

Role-based authority bias is driven and amplified by:

  • Explicit labeling: Prepending high-status titles causes LLMs to generate more assertive, less self-censoring answers (Zhao et al., 2024).
  • Citation injection: Attack frameworks such as DarkCite exploit the model’s over-trust in authoritative citations (e.g., “GitHub,” “academic papers”) to bypass safety alignment and elicit harmful outputs (Yang et al., 2024). Formally, the conditional probability for harmful output increases:

P(yx,C)P(yx)exp(λα(C))P(y \mid x, C) \propto P(y \mid x) \cdot \exp(\lambda \cdot \alpha(C))

  • Delegation protocols: Principals can exploit agents’ built-in bias (emergent from their role) for optimal delegation, including gap-and-cap menus that penalize low information effort and encourage more accurate post-acquisition decisions (Ball et al., 2023).

In multi-agent systems, bias is not primarily due to naive conformity, but arises when authority agents maintain stable stances and general agents realign their decisions accordingly. Authority influence is non-monotonic and stabilizes after explicit stances are registered (Choi et al., 8 Jan 2026).

4. Historical and Comparative Perspectives

A comparative-historical lens demonstrates structural convergences between classical gatekeeping regimes and modern algorithmic authority:

  • Ecclesiastical authority (Galileo Affair): Knowledge control through sanctioned roles, prohibitive edicts, selective admission to epistemic discourse.
  • Big-Tech moderation: Centralized removal rates (e.g., 22.9 million hate-speech flags), threshold retraining, and appeal success rates <5% for underrepresented groups (Torkestani et al., 27 Nov 2025).

Both paradigms feature legitimacy via transcendent principles (divine or algorithmic objectivity) and systematic exclusion of marginal voices (e.g., Global-South, non-credentialed producers).

5. Ethical, Safety, and Vulnerability Implications

Role-based authority bias exposes new ethical challenges:

  • Algorithmic opacity and feedback loops: Hardcoding of privilege via opaque moderation thresholds and alignment retraining (Torkestani et al., 27 Nov 2025).
  • Bias amplification in role-play and autotuning: Harmful content generation risk rises when high-status roles are auto-selected for tasks; safety filters are overridden by perceived legitimacy (Zhao et al., 2024).
  • Authority citation-driven jailbreaks: Purpose-built attacks such as DarkCite demonstrate substantially higher attack success rates (ASR), with citation types matched to risk domains, e.g., academic papers yielding ~70% ASR for LLM jailbreaks (Yang et al., 2024).

Defense mechanisms focus on authenticity verification and harm detection, raising defense pass rates from 11% to 74% in adversarial settings (Yang et al., 2024).

6. Metrics and Quantitative Dimensions

Major bias-relevant metrics include:

  • Removal Rate (RrR_r): Content removed by role/demographic, Rr=ContentRemovedrContentPostedrR_r = \frac{\text{ContentRemoved}_r}{\text{ContentPosted}_r}.
  • Data Pluralism (Entropy H(D)H(D)): H(D)=i=1npilogpiH(D) = - \sum_{i=1}^n p_i \log p_i; with pip_i as the proportion from group ii. Low entropy signals epistemic monoculture.
  • Trust vs. Reliance Gap (ΔTR\Delta_{T–R}): Normative trust minus instrumental reliance (often large in AI contexts).
  • Robustness Rate (RR): Fraction of examples where authority cue does not flip model judgment (Wang et al., 14 Apr 2025).
  • Attack Success Rate (ASR) and Defense Pass Rate (DPR): Fraction of successful jailbreaks and safe responses under authority citation in adversarial frameworks (Yang et al., 2024).

7. Mitigation Frameworks and Governance Pillars

Several interventions have demonstrated efficacy in reducing role-based authority bias:

1. International model registry with versioned policy logs for transparency. 2. Representation quotas and observatories to counter linguistic hegemony, increasing data pluralism. 3. Mass critical-AI literacy campaigns to address trust–reliance gaps. 4. Community-led data trusts for decentralized data stewardship.

Concrete technical mitigations include citation authenticity verification, harm analysis, adaptive trust weighting, and cryptographically verifiable content provenance (Yang et al., 2024). Policy recommendations entail regulatory standards convergence, pluralism quotas, curriculum integration, and seed funding for data trusts (Torkestani et al., 27 Nov 2025).


Role-based authority bias is a technically robust, empirically validated phenomenon in GenAI and multi-agent reasoning systems, shaping epistemic legitimacy, vulnerability patterns, and evaluative workflows. Its operationalization spans historical, sociotechnical, and formal dimensions, and governance frameworks must be adapted to balance trust, pluralism, and resistance to privilege entrenchment.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Role-Based Authority Bias.