Papers
Topics
Authors
Recent
Search
2000 character limit reached

Heimdall: test-time scaling on the generative verification

Published 14 Apr 2025 in cs.AI | (2504.10337v2)

Abstract: An AI system can create and maintain knowledge only to the extent that it can verify that knowledge itself. Recent work on long Chain-of-Thought reasoning has demonstrated great potential of LLMs on solving competitive problems, but their verification ability remains to be weak and not sufficiently investigated. In this paper, we propose Heimdall, the long CoT verification LLM that can accurately judge the correctness of solutions. With pure reinforcement learning, we boost the verification accuracy from 62.5% to 94.5% on competitive math problems. By scaling with repeated sampling, the accuracy further increases to 97.5%. Through human evaluation, Heimdall demonstrates impressive generalization capabilities, successfully detecting most issues in challenging math proofs, the type of which is not included during training. Furthermore, we propose Pessimistic Verification to extend the functionality of Heimdall to scaling up the problem solving. It calls Heimdall to judge the solutions from a solver model and based on the pessimistic principle, selects the most likely correct solution with the least uncertainty. Taking DeepSeek-R1-Distill-Qwen-32B as the solver model, Pessimistic Verification improves the solution accuracy on AIME2025 from 54.2% to 70.0% with 16x compute budget and to 83.3% with more compute budget. With the stronger solver Gemini 2.5 Pro, the score reaches 93.0%. Finally, we prototype an automatic knowledge discovery system, a ternary system where one poses questions, another provides solutions, and the third verifies the solutions. Using the data synthesis work NuminaMath for the first two components, Heimdall effectively identifies problematic records within the dataset and reveals that nearly half of the data is flawed, which interestingly aligns with the recent ablation studies from NuminaMath.

Summary

  • The paper presents Heimdall, which employs reinforcement learning via PPO to enhance verification accuracy, achieving up to 97.5% with repeated sampling.
  • It utilizes step-by-step chain-of-thought reasoning along with Majority Voting for robust automatic verification of mathematical solutions.
  • Experimental results demonstrate a significant accuracy boost from baseline models, underscoring its potential in improving AI reliability and dataset integrity.

Heimdall: Test-Time Scaling on the Generative Verification

Introduction

The paper "Heimdall: test-time scaling on the generative verification" (2504.10337) introduces Heimdall, a novel LLM trained for verification during the test-time scaling process of generative tasks. Verification remains a critical component within AI systems, enabling these systems to maintain and create knowledge effectively. Despite the extensive research on problem-solving capabilities with LLMs enhanced by CoT reasoning, the verification dimension lacks comprehensive exploration. Heimdall targets this gap by specifically improving the verification accuracy using reinforcement learning (RL), reaching a remarkable increase in verification accuracy on mathematical problems.

Methodology

Heimdall employs Proximal Policy Optimization (PPO) for reinforcement learning to enhance verification accuracy. The RL setup involves filtering specific problems to ensure effective data processing, focusing on contrasting cases, which expedites the learning process. The training method emphasizes solutions where Heimdall verifies solution correctness via step-by-step CoT reasoning. A specially designed prompt guides the model through verifying solution correctness, using forward and backward checking techniques, as illustrated in multiple verification scenarios. Figure 1

Figure 1: Scaling of Heimdall. Left: the verification accuracy scales with response length; Middle: accuracy scales with repeated sampling; Right: problem solving accuracy scales with the number of solutions.

The algorithm incorporates repeated sampling and Majority Voting for judgment augmentation, achieving significant improvements in verification accuracy. Through an iterative process, Heimdall effectively scales its verification ability with problem-solving scenarios, demonstrating superior performance over existing baseline models.

Results and Implications

Heimdall achieves impressive results, advancing verification accuracy from 62.5% to 94.5%, and further to 97.5% with repeated sampling. The implementation of Pessimistic Verification optimizes problem-solving by minimizing uncertainty and adjusting for compute constraints, enhancing solution accuracy significantly. Figure 2

Figure 2: Accuracy and response length during RL training. Normal RL converges faster after filtering problematic datasets, shown in left and right plots respectively.

Heimdall's verification framework optimally collaborates with solver models, drastically improving accuracy—as evidenced from the uplift from DeepSeek-R1-Distill-Qwen-32B’s baseline results from 54.2% to 83.3% with an extended compute budget. Moreover, when incorporated with more advanced solver models such as Gemini 2.5 Pro, Heimdall reaches 93% accuracy, aligning with current state-of-the-art techniques.

The paper further showcases Heimdall’s utility in an automatic knowledge discovery system where it successfully identifies erroneous data within a synthetic dataset. This highlights Heimdall’s potential in ensuring dataset integrity and its broader implications in sustainable AI training paradigms. Figure 3

Figure 3: The inference-time scaling of verification on AIME datasets showcasing accuracy enhancement with verification multiplicity.

Discussion and Future Work

The extension of Heimdall into domains such as coding and proof verification is a promising future endeavor. Current results demonstrate Heimdall’s excellent generalization capability; however, integrating more diverse datasets could further reinforce its adaptability. Additionally, the creation of high-fidelity synthetic datasets remains a substantial challenge; the application of Heimdall in this domain could provide a pathway to improved data quality through rigorous verification mechanisms.

Simultaneously, the exploration of autonomous knowledge generation processes poses new challenges, potentially benefitting from an AI-driven inquisitive capability harmonized with Heimdall’s verification acumen. These explorations may well align with breakthroughs in both theoretical and applied AI research, reinforcing the robustness of autonomous systems driven by LLMs.

Conclusion

Heimdall substantiates a methodological leap in generative verification, leveraging reinforcement learning to refine CoT reasoning capabilities for competitive tasks. With innovations like Pessimistic Verification, it fosters synergistic collaboration between problem-solving solvers and verification, positioning verification as an instrumental component in enhancing AI’s problem-solving repertoire. The paper reinforces the significance of verification in embedding reliability within autonomous AI systems, paving pathways for future exploration in transparent and verifiable AI processes.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 8 likes about this paper.