Papers
Topics
Authors
Recent
Search
2000 character limit reached

Leveraging Large Language Models for Causal Discovery: a Constraint-based, Argumentation-driven Approach

Published 18 Feb 2026 in cs.AI | (2602.16481v1)

Abstract: Causal discovery seeks to uncover causal relations from data, typically represented as causal graphs, and is essential for predicting the effects of interventions. While expert knowledge is required to construct principled causal graphs, many statistical methods have been proposed to leverage observational data with varying formal guarantees. Causal Assumption-based Argumentation (ABA) is a framework that uses symbolic reasoning to ensure correspondence between input constraints and output graphs, while offering a principled way to combine data and expertise. We explore the use of LLMs as imperfect experts for Causal ABA, eliciting semantic structural priors from variable names and descriptions and integrating them with conditional-independence evidence. Experiments on standard benchmarks and semantically grounded synthetic graphs demonstrate state-of-the-art performance, and we additionally introduce an evaluation protocol to mitigate memorisation bias when assessing LLMs for causal discovery.

Summary

  • The paper proposes a novel approach that leverages LLM-generated constraints to improve causal graph inference.
  • It integrates structured semantic priors with data-derived conditional independence tests within the Causal ABA framework.
  • Evaluation on synthetic datasets shows superior performance in normalized SHD and F1-score compared to traditional methods.

Leveraging LLMs for Causal Discovery: A Constraint-Based, Argumentation-Driven Approach

This paper proposes a novel methodology for causal discovery by integrating LLMs into Causal Assumption-based Argumentation (Causal ABA). By using LLMs as imperfect experts, it aims to enhance the discovery process through structured semantic priors while maintaining the rigor of symbolic reasoning. This approach addresses the challenge of combining semantic insights with data-derived evidence, thereby advancing both theoretical and practical aspects of causal graph inference.

Integrating LLMs with Causal ABA

The integration of LLMs into Causal ABA involves a pipeline that elicits causal constraints from variable metadata, such as names and optional descriptions, by prompting LLMs to generate pairwise causal statements. Figure 1

Figure 1: LLM Integration Pipeline: LLMs generate causal statements from variable metadata, which are structured into assumptions for Causal ABA to infer causal graphs.

The generated constraints, filtered through a consensus mechanism, are introduced into Causal ABA, complementing data-derived conditional independence constraints. This dual-source constraint integration aims to maximize causal inference accuracy by combining the rich prior knowledge encoded in LLMs with empirical data evidence.

Evaluation Protocol and Results

Evaluation of this approach is conducted using both standard benchmarks and novel synthetic datasets from CauseNet, a web-extracted knowledge graph. The synthetic evaluation protocol ensures robustness against memorization biases by grounding random DAGs in real-world concepts and selecting the most semantically coherent subgraphs through heuristic scoring. Figure 2

Figure 2: Synthetic Evaluation Protocol: DAGs are grounded in CauseNet, and heuristic scores select the most coherent subgraphs for robust LLM evaluation.

Comparisons with traditional algorithms show that the LLM-augmented Causal ABA outperforms competitors in normalised Structural Hamming Distance (SHD) and F1-score across synthetic datasets. Figure 3

Figure 3: Normalised SHD and F1-score comparisons of LLM-augmented Causal ABA with baselines, indicating superior performance on synthetic datasets.

Interaction of Constraints and Data

The quality of LLM-derived constraints is pivotal. Heatmaps illustrate the correlation between the quality of these constraints, derived from LLMs, and improvements in graph reconstruction accuracy after integration with Causal ABA. Figure 4

Figure 4: Heatmap showing the relationship between LLM-derived and data-derived constraint quality and graph reconstruction accuracy.

Results indicate that when high-quality LLM constraints align with strong data-derived evidence, structural improvements are most substantial. This synergy highlights the value of structured integration of semantic and statistical signals.

Conclusion

This novel framework represents an effective approach to hybrid causal discovery, showcasing significant improvements in causal graph inference when semantic insights from LLMs are judiciously integrated with data-derived constraints. Future work could focus on extending this approach to encompass knowledge from scientific texts, exploring adaptive confidence calibration for asserting LLM insights, and further diversifying LLM capabilities through more nuanced prompt designs.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 1 like about this paper.