Papers
Topics
Authors
Recent
Search
2000 character limit reached

Why Isn't Relational Learning Taking Over the World?

Published 17 Jul 2025 in cs.AI, cs.DB, and cs.LG | (2507.13558v1)

Abstract: AI seems to be taking over the world with systems that model pixels, words, and phonemes. The world is arguably made up, not of pixels, words, and phonemes but of entities (objects, things, including events) with properties and relations among them. Surely we should model these, not the perception or description of them. You might suspect that concentrating on modeling words and pixels is because all of the (valuable) data in the world is in terms of text and images. If you look into almost any company you will find their most valuable data is in spreadsheets, databases and other relational formats. These are not the form that are studied in introductory machine learning, but are full of product numbers, student numbers, transaction numbers and other identifiers that can't be interpreted naively as numbers. The field that studies this sort of data has various names including relational learning, statistical relational AI, and many others. This paper explains why relational learning is not taking over the world -- except in a few cases with restricted relations -- and what needs to be done to bring it to it's rightful prominence.

Summary

  • The paper critically examines why relational learning remains niche despite abundant real-world relational data.
  • It identifies significant challenges in data representation, evaluation metrics, and aggregation mechanisms that impede practical applications.
  • The study advocates for new benchmarks and lightweight, expressive models that align evaluation with real-world decision-making tasks.

Why Isn't Relational Learning Taking Over the World?

David Poole's "Why Isn't Relational Learning Taking Over the World?" presents a critical analysis of the current state and limitations of relational learning, contrasting its theoretical promise with its limited practical adoption. The paper systematically examines why, despite the prevalence of relational data in real-world settings, relational learning has not achieved the same level of impact as models focused on perceptual data such as pixels and words.

Central Arguments

The paper begins by highlighting a disconnect between the nature of real-world data and the focus of mainstream machine learning. While much of the recent progress in AI has centered on perceptual data (text, images, audio), most valuable organizational data exists in relational formats—spreadsheets, databases, and knowledge graphs—composed of entities, properties, and relations. The author argues that modeling these entities and their interrelations is more aligned with the structure of the world and the requirements of many practical applications.

Despite this, relational learning remains a niche area, with widespread adoption limited to specific domains (e.g., collaborative filtering, protein structure prediction, traffic modeling). The paper explores several reasons for this, including challenges in data representation, evaluation, and the handling of missing or incomplete information.

Data Representation and Knowledge Graphs

The paper provides a detailed discussion of how relational data is structured, using knowledge graphs as a canonical example. It describes the process of converting tabular data into triples (subject, verb, object), the use of reification to represent complex relations, and the limitations of current benchmarks such as FB15k and WN18. The analysis reveals that standard datasets often fail to capture the complexity and sparsity of real-world knowledge graphs, where the vast majority of entities are associated with very few facts.

A key empirical observation is that in Wikidata, approximately 99% of entities are the subject of fewer than 10 triples, indicating a long-tail distribution that is not reflected in commonly used benchmarks. This has significant implications for the generalizability and utility of relational learning models trained on these datasets.

Evaluation Challenges

The paper critiques standard evaluation methodologies in relational learning, particularly the reliance on ranking metrics such as mean reciprocal rank (MRR) and hit-at-k. These metrics, while convenient, are shown to be poorly aligned with real-world decision-making tasks. For example, predicting whether a random triple is true in Wikidata is trivial due to the overwhelming sparsity of the data. Furthermore, the lack of negative examples and the open-world assumption in most knowledge graphs complicate the estimation of meaningful probabilities.

The author emphasizes that evaluation should be closely tied to downstream tasks, whether autonomous or human-in-the-loop decision making. Current practices, such as generating negative examples by corrupting triples, introduce artificiality and do not reflect the true distribution of missing or false information in real databases.

Theoretical and Practical Limitations

Several theoretical and practical limitations of current relational learning approaches are identified:

  • Aggregation: Existing models struggle with aggregation, especially when the number of related entities varies widely or is unbounded. Most models either assume independence (e.g., sum, noisy-or) or collapse evidence (e.g., max, mean, attention), both of which are problematic in practice and theory.
  • Embeddings: The use of fixed-size embeddings for entities of vastly different complexity (e.g., the USA vs. a specific sports relationship) is highlighted as a conceptual flaw.
  • Generalization: There is a lack of models that can generalize across domains or to new populations, as most benchmarks and evaluation protocols do not test for this.
  • Missing Data: The open-world nature of most knowledge graphs means that missing data is not missing at random, and current probabilistic models for missingness are either too heavyweight or insufficiently expressive.

Implications and Future Directions

The paper calls for a shift in focus towards real, public datasets that matter to stakeholders, such as environmental or health records, despite the challenges of access and privacy. It advocates for evaluation protocols that are tailored to actual decision-making tasks and for models that can handle the complexities of missing data, aggregation, and heterogeneous information.

The author suggests that realizing the full potential of relational learning will require:

  • Development of lightweight, expressive models for missing data.
  • Improved aggregation mechanisms that can handle both independence and dependence among related entities.
  • Embedding strategies that scale with the complexity of entities.
  • Integration of meta-information and ontologies to enable reasoning across heterogeneous datasets.
  • Closer collaboration with domain experts to identify meaningful prediction tasks.

Numerical Results and Claims

The paper notes that state-of-the-art models achieve hit-at-10 rates of approximately 55% on FB15k-237, a widely used benchmark. However, it questions the practical utility of such results, given the triviality or impossibility of many queries and the lack of alignment with real-world tasks. The claim is made that current evaluation metrics and benchmarks do not provide a meaningful measure of progress towards the goals of relational learning.

Conclusion

"Why Isn't Relational Learning Taking Over the World?" provides a comprehensive and critical perspective on the field, identifying both the promise and the persistent obstacles facing relational learning. The analysis underscores the need for new datasets, evaluation protocols, and modeling approaches that are better aligned with the structure of real-world data and the requirements of practical applications. The paper's insights have significant implications for future research, suggesting that progress in relational learning will depend on addressing foundational issues in data representation, evaluation, and model design, as well as fostering interdisciplinary collaboration.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 3 likes about this paper.

alphaXiv