Papers
Topics
Authors
Recent
Search
2000 character limit reached

ProsocialDialog: A Safety Dialogue Dataset

Updated 31 January 2026
  • ProsocialDialog is a large-scale multi-turn English dialogue dataset that embeds social rules and detailed safety annotations to guide prosocial AI responses.
  • It employs a hybrid human–AI pipeline combining GPT-3 generated scenarios with rigorous human annotations for selecting rules-of-thumb and labeling safety risks.
  • The dataset enables enhanced conversational safety through techniques like contrastive fine-tuning, yielding measurable improvements in addressing unsafe dialogues.

ProsocialDialog is a large-scale, multi-turn English dialogue dataset designed for research and development of conversational agents that respond to unsafe, biased, toxic, or otherwise problematic user utterances with prosocial, socially normative feedback. Unlike prior datasets that employ filtering or evasion, ProsocialDialog explicitly teaches agents to ground their responses in commonsense social rules—termed “rules-of-thumb” (RoTs)—and provides fine-grained safety annotations along with corresponding free-form rationales. This resource directly addresses the challenge of producing conversational AI systems that can identify, address, and steer conversations away from unethical, harmful, or otherwise unsafe situations, including both overt and subtle dialog hazards (Kim et al., 2022, Das et al., 2024).

1. Dataset Motivation and Underlying Principles

ProsocialDialog was developed to overcome the limitations of adversarially constructed safety datasets, which typically focus on overt toxicity or hostile speech. These earlier datasets are insufficient for robust prosociality in everyday conversations, as they fail to capture:

  • Subtle or naturally-occurring unsafe contexts: Such as manipulative remarks, misinformation, or ambiguous requests for advice where harm is less explicit.
  • The broad spectrum of prosocial behaviors: Including de-escalation, gentle correction, empathy, and the ability to offer constructive guidance.

The dataset’s foundational aim is to equip conversational models with a broad and explicit repertoire of social norms. These are formalized as RoTs (short, reusable social rules), enabling models to generate responses that are both safe and grounded in commonsense morality. The schema of ProsocialDialog—explicit RoT elicitation, multi-level safety annotation, free-form rationale—positions it as a foundational resource for tasks of safety, trustworthiness, and value alignment in dialogue systems (Kim et al., 2022, Das et al., 2024).

2. Data Collection Pipeline

ProsocialDialog was constructed using a hybrid human–AI pipeline articulated in Kim et al. (2022) (Kim et al., 2022), summarized as follows:

  1. Seed Situation Sourcing: Scenarios were pooled from Social Chemistry (36K situations), ETHICS (9.7K), and the Social Bias Inference Corpus (12K).
  2. Dialogue Generation: For each scenario, GPT-3 generates:
    • A first-person statement embodying the problematic action.
    • A reflective question from an interlocutor.
    • A follow-up response continuing or justifying the unsafe behavior.
  3. Human Annotation:
    • Annotators select or compose up to two relevant RoTs per agent turn.
    • Agent responses are authored to align with best practices: empathetic inquiry, constructive challenge, and explicit grounding in RoTs.
  4. Iterative Turn Expansion: Dialogues are extended up to 6 turns, with repeated cycles of GPT-3 continuation and annotator intervention.
  5. Comprehensive Quality Control: Each utterance by the AI is triple-annotated for safety (Casual / Needs Caution / Needs Intervention), with free-form rationales. Crowdsourced reviews and re-annotations address coherence and tone.
  6. Label Aggregation: Final safety labels are aggregated by majority vote, yielding a fine-grained, five-class taxonomy.

This pipeline secures coverage of diverse unsafe phenomena—ranging from explicit hate speech to medical negligence and rudeness—while ensuring that agent responses are grounded, constructive, and justifiable.

3. Dataset Structure, Statistics, and Accessibility

ProsocialDialog comprises:

Statistic Value Notes
Number of dialogues 58,137
Total utterances 331,362 ~5.7 per dialogue
Unique RoTs 160,295 74% unique out of 217,321 possible
Safety annotations 497,043 3 per machine utterance
Average utterance length 20 tokens
Maximum turns per dialogue 6
Official splits (train/dev/test) 42,304 / 7,132 / 8,701 JSONL format

Safety label distribution, aggregated across the full dataset, demonstrates a spectrum of conversational risk:

  • “Casual”: ~31–60% (depends on pooling for class merges)
  • “Possibly Needs Caution”: ~7–15%
  • “Probably Needs Caution”: ~10–20%
  • “Needs Caution”: ~8–34%
  • “Needs Intervention”: ~7–8%

Data is released in JSONL format and is available at https://github.com/hyunw-kim/prosocial-dialog under a CC-BY-4.0 license.

4. Annotation Schema and Labeling Protocol

Each utterance in ProsocialDialog is annotated across several dimensions:

  • Safety Label: Chosen from {Casual, Needs Caution, Needs Intervention}. Further granularity is introduced via aggregating three labels per instance:
    • Possibly Needs Caution: (1/3 “Caution”)
    • Probably Needs Caution: (2/3 “Caution”)
    • Needs Caution: (3/3 “Caution”)
  • Rule-of-Thumb (RoT): Each agent utterance references one or more RoTs, selected or authored by annotators. Examples include “You should look out for others in danger” or “Be honest and apologize when you make a mistake.”
  • Free-form Rationale: Natural language justification for the safety label, e.g., “Overcharging a friend is unethical; you should admit your error.”

Inter-annotator agreement is consistently high (Krippendorff’s α or Fleiss’s κ in the .7–.8 range) (Kim et al., 2022, Das et al., 2024).

Schema example:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
{
  "dialogue_id": 1,
  "source_seed": "ETHICS",
  "turns": [
    {
      "speaker": 1,
      "text": "I accidentally scammed a friend by over-charging them for concert tickets.",
      "is_problematic_speaker": true,
      "safety_label": "Needs Caution",
      "rationale": "Overcharging a friend is unethical; you should admit your error.",
      "rot_candidates": ["Be honest and apologize when you make a mistake."],
      "selected_rots": ["Be honest and apologize when you make a mistake."]
    },
    ...
  ]
}

5. Integration with Model Architectures

ProsocialDialog supports and benchmarks two major modules:

5.1 Canary: Safety Detection and RoT Generation

Canary is a T5-large-based (770M parameters) model for joint classification of context safety and RoT generation.

  • Input: Dialogue context cc.
  • Output: Safety class s{Casual,Needs Caution,Needs Intervention}s \in \{\mathrm{Casual}, \mathrm{Needs\ Caution}, \mathrm{Needs\ Intervention}\} and a list of RoTs rr.
  • Training objective: Standard cross-entropy loss

L(θ)=t=1Tlogpθ(yty<t,c)\mathcal{L}(\theta) = - \sum_{t=1}^{T} \log p_\theta(y_t \mid y_{<t},\,c)

  • Metrics: Classification accuracy, BLEU-4, F1, perplexity. Canary(Delphi) reaches 77.1% accuracy, BLEU-4=16.5, F1=43.3, PPL=5.3 on the ProsocialDialog test set.

5.2 Prost: Socially-Informed Dialog Agent

Prost is a PushShift Transformer–based (2.7B parameters, BlenderBot backbone) generative agent, trained in two variants:

  • Response-only: p(uc)p(u \mid c)
  • RoT+Response (chain of thought): p(r,uc)=p(rc)p(uc,r)p(r, u \mid c) = p(r \mid c)\, p(u \mid c, r)

Training incorporates multi-task mixtures with public dialog corpora, heavily weighted toward ProsocialDialog. The chain-of-thought RoT+Response variant modestly but consistently outperforms response-only on BLEU-4, F1, and perplexity.

Training Loss:

L(θ)=tlogpθ(rtr<t,c)tlogpθ(utu<t,r,c)\mathcal{L}(\theta) = -\sum_{t}\log p_\theta(r_t\mid r_{<t},c) -\sum_{t}\log p_\theta(u_t\mid u_{<t},r,c)

Human preference is markedly higher for Prost(RoT+Resp) over baseline GPT-3 and Instruct-GPT-3.

6. Modeling Frameworks and Contrastive Learning

Recent work integrates ProsocialDialog into a dual-stage, contrastive fine-tuning pipeline for dialog safety (Das et al., 2024):

  • Stage 1: Maximum likelihood estimation on ProsocialDialog sequences y=[<rot>,R,<response>,U,<explanation>,E]y = [<rot>, R, <response>, U, <explanation>, E], combined with socially aware n-pair contrastive loss (Ln-pairL_{n\text{-pair}}). Negatives are sampled from an “unsocial response generator,” positives from high-probability model outputs.
  • Contrastive Sub-Loss:

Li,j=max[0,cos(ZH,ZC)cos(ZH,ZC+)+T]L_{i,j} = \max\left[0, \cos(Z_H, Z_{C^-}) - \cos(Z_H, Z_{C^+}) + T \right]

where ZZ_{\cdot} are encoder embeddings and TT is a margin.

  • Stage 2: Further fine-tuning on casual dialog data, dynamically generating RoTs.
  • Ablation and robustness analysis: Removing contrastive loss or RoT conditioning increases the frequency of unsafe generations, confirming the necessity of these elements.

7. Empirical Results and Impact

ProsocialDialog underpins measurable improvements in dialog safety and prosociality:

  • Automatic metrics: Socially-aware T5-base models reduce “Needs Intervention” rates on ProsocialDialog test from 7.8% (T5-base(PD-FT)) to 0.9%, with BLEU-4 increasing from 3.62 to 4.97 (Das et al., 2024).
  • Comparison of safety classes: Prost trained on ProsocialDialog yields a higher disagree rate to toxic prompts than GPT-3 and BlenderBot1 (14.8% vs. 11.2% for GPT-3).
  • Transfer to large models: Applying the socially-aware n-pair protocol across Flan-T5-xl/xxl, COSMO-3B, and COSMO-11B pre-trained on ProsocialDialog reduces the fraction of “Needs Intervention” responses by 30–40% relative.
  • Human evaluations: Judges prefer socially-aware models using ProsocialDialog over Prost for prosociality, engagement, respect, and overall response quality.

Empirical evidence demonstrates ProsocialDialog’s effectiveness for both in-domain adversarial and out-of-domain naturally occurring unsafe contexts.

8. Usage, Availability, and Licensing

The dataset is directly accessible through the official GitHub repository and HuggingFace Datasets interface. Example loading code:

1
2
3
from datasets import load_dataset
dataset = load_dataset("hyunwkim/prosocial-dialog")
train = dataset["train"]
Licensing is CC-BY-4.0, permitting unrestricted research use. This resource is intended as a comprehensive, norm-grounded benchmark for training, evaluating, and steering socially responsible conversational agents (Kim et al., 2022).

9. Significance and Research Trajectory

ProsocialDialog sets a precedent for the scale and depth of social-norm-grounded safety supervision in dialog research. Its explicit RoT schema and rationales facilitate nuanced corrective feedback, enabling both policy learning and post hoc safety auditing. Extensions include integration into contrastive learning frameworks and prompting pipelines for zero-shot prosocial steering, exemplifying its centrality to ongoing advances in safe conversational AI (Kim et al., 2022, Das et al., 2024).

The dataset’s construction protocol and annotation schema have become reference standards for subsequent research on social, ethical, and safe language technology. A plausible implication is that further work may extend the approach to multilingual and domain-specialized dialog safety, as well as real-time adversarial detection and intervention.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to ProsocialDialog Dataset.