Papers
Topics
Authors
Recent
Search
2000 character limit reached

Political DEBATE: Efficient Zero-shot and Few-shot Classifiers for Political Text

Published 3 Sep 2024 in cs.CL | (2409.02078v1)

Abstract: Social scientists quickly adopted LLMs due to their ability to annotate documents without supervised training, an ability known as zero-shot learning. However, due to their compute demands, cost, and often proprietary nature, these models are often at odds with replication and open science standards. This paper introduces the Political DEBATE (DeBERTa Algorithm for Textual Entailment) LLMs for zero-shot and few-shot classification of political documents. These models are not only as good, or better than, state-of-the art LLMs at zero and few-shot classification, but are orders of magnitude more efficient and completely open source. By training the models on a simple random sample of 10-25 documents, they can outperform supervised classifiers trained on hundreds or thousands of documents and state-of-the-art generative models with complex, engineered prompts. Additionally, we release the PolNLI dataset used to train these models -- a corpus of over 200,000 political documents with highly accurate labels across over 800 classification tasks.

Citations (2)

Summary

  • The paper introduces a robust framework by compiling diverse datasets and leveraging innovative zero-shot and few-shot classifiers for political stance detection.
  • It outlines detailed methodologies, including binary classification prompts and optimized training parameters, to enhance model performance.
  • The research provides a foundation for real-world applications in social media monitoring and political analysis, setting new standards for future AI studies.

A Comprehensive Study on Data Sources and Methodologies for Political Stance Detection

The paper in question provides an extensive evaluation of data sources and methodologies utilized for political stance detection. The authors have compiled and annotated a variety of datasets sourced from several credible references. This collection is aimed at aiding in the development and performance benchmarking of models that can discern political stances and classify events related to political and social unrest.

Data Sources

A primary contribution of this work is the diverse array of datasets it introduces and details. This compilation includes the following datasets:

  • Multi-target Stance Detection: Includes tweets specifically annotated for stance towards multiple politicians.
  • PoliBERTweet Training: Features tweets about prominent political figures, Trump and Biden, labeled for stance.
  • Polistance Affect: A newly introduced dataset containing tweets labeled for stance towards over 20 members of Congress.
  • Polistance Quote Tweets: Another new dataset, focusing on quote tweets labeled for stance towards members of Congress.
  • Newsletter Sentences: Consists of sentences extracted from newsletters, labeled similarly for stance towards members of Congress.
  • Political Tweets: Housed on the Huggingface Hub, this dataset includes tweets from senators and representatives annotated for political issue stances.
  • ADL Heat Map Dataset: Describes antisemitic incidents with categorical and typological labels.
  • State of the Union Speeches: Contains sentence-wise topic coded excerpts from the State of the Union speeches.
  • Democratic and Republican Party Platforms: Coded sentences by topic from party platforms.
  • The Supreme Court Database: Offers summaries of court cases labeled by legal topics.
  • Argument Quality Ranking: Focuses on crowd-sourced arguments for various political propositions, annotated for quality.
  • Global Warming Media Stance: News leads labeled on whether they portray global warming as a threat.
  • Claim Stance Datasets: Features claims from Wikipedia across multiple topics, annotated for stance and topic.
  • ACLED and SCAD: Capture descriptions and summaries of violent events and political demonstrations, categorized by event type.
  • Measuring Hate Speech: Annotates instances of hate speech and counter hate speech based on crowd-sourced labels.
  • Anthropic Persuasion: Features arguments generated by AI models Claude 2 and 3, filtered for political topics.
  • Polarizing Rhetoric Tweets: Annotated for the use of polarizing rhetoric.
  • Bill Summaries and Political or Not: Contain summaries and labels of bills, and a mixed set of news articles and samples from other datasets.

Given the breadth and diversity of these data sources, the paper provides a solid foundation for research in political stance detection and related domains. Each dataset is tailored to specific facets of political communication, ensuring that models trained on these data will be robust and generalizable.

Methodologies

The paper further outlines the LLM prompt designs and training parameters used for model training and evaluation:

  • Prompts: For GPT-4 and GPT-4o, the authors employed a binary classification prompt and a prompt for hypothesis augmentation. These are designed to streamline and standardize the input to the model to ensure consistency in classification tasks.
  • Training Parameters: Both base and large models utilized a linear learning rate scheduler, specific learning rates, batch sizes, and epochs tailored to optimize the models’ performance. For instance, the base model employed a learning rate of 2e-5 and the large model 9e-6, reflecting a well-considered trade-off between convergence speed and stability.

Implications and Future Work

The implications of this compilation are extensive. Practically, the datasets enable the training of highly detailed political stance detection models, which can be utilized for various applications, including social media monitoring, political forecasting, and scholarly analysis of political communication.

Theoretically, the methodologies and data annotations set a new standard for research in this field. By aligning various datasets under a common framework, the paper facilitates replication and further exploration of political sentiment analysis and stance detection via advanced LLMs.

Future developments may build on this work by integrating real-time data from dynamic sources like social media or live broadcasts, enhancing the temporal relevance of stance detection models. Moreover, the fine-tuning of LLMs on this data could lead to more specialized models capable of nuanced political analysis, potentially extending into multilingual and multicultural contexts.

In conclusion, the paper provides a comprehensive foundation for advancing political stance detection through a rich collection of datasets and a robust methodological framework. This work is poised to drive future research and practical applications in political data analysis and artificial intelligence.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 6 tweets with 69 likes about this paper.