Papers
Topics
Authors
Recent
Search
2000 character limit reached

Iterative Document Reranking (IDRQA)

Updated 10 February 2026
  • IDRQA is an iterative, adaptive framework that refines candidate document rankings through repeated cycles of retrieval and reranking to enhance open-domain and multi-hop QA.
  • It leverages uncertainty modeling with Bayesian updates and graph-based message passing to selectively allocate computation and reduce excessive LLM calls.
  • Empirical evaluations show that IDRQA outperforms fixed-computation baselines on benchmarks like HotpotQA and SQuAD Open by achieving superior accuracy-efficiency trade-offs.

Iterative Document Reranking for Question Answering (IDRQA) encompasses a class of adaptive, multi-stage pipelines for open-domain and multi-hop information retrieval and reasoning. These pipelines incorporate repeated cycles of retrieval and reranking—often in concert with document-level uncertainty modeling and dynamic allocation of computation—to maximize retrieval accuracy and downstream reasoning efficiency under compute or inference cost constraints. IDRQA systems are distinguished by their iterative nature: they continually refine candidate rankings based on evolving evidence, uncertainty, or multi-document interactions, terminating upon satisfaction of answerability or confidence criteria.

1. Conceptual Principles

IDRQA formalizes retrieval pipelines as a sequence of interleaved retrieval, reranking, and reasoning stages optimized for both accuracy and efficiency. Unlike static rerankers that execute a single listwise scoring pass over a subset of candidates, or fixed-computation schemes (e.g., a set number of sliding window passes), IDRQA leverages adaptive, uncertainty-aware procedures to selectively allocate reranker computation where its marginal utility is highest. This typically involves probabilistic modeling of relevance estimates and the use of Bayesian or graph-based mechanisms to focus reranker effort on ambiguous or informationally-rich regions of the ranking, iterating until targeted convergence conditions are met (Yoon et al., 24 May 2025, &&&1&&&, Nie et al., 2020).

2. Framework Architectures

IDRQA pipelines vary in their exact architectures but generally conform to a three-stage design:

  1. Initial Retrieval: An initial candidate pool is assembled from a large corpus using sparse (BM25, TF–IDF) or dense (e.g., DPR, qwen3-embedding-8b) retrievers. Candidate documents are truncated to a manageable context size (e.g., 512 tokens).
  2. Iterative Reranking: Subsets (often the top-dd candidates by raw retrieval score) are repeatedly subjected to listwise reranking via LLMs or graph-based neural rerankers. Reranking may use:
    • Bayesian Adaptive Reranking: Probabilistic beliefs (e.g., Gaussian posteriors of relevance) are updated via TrueSkill-style message passing based on LLM or listwise feedback (Yoon et al., 24 May 2025).
    • Graph-Based Multi-Document Reranking: Candidates are nodes in a named-entity graph, with information aggregated via Graph Attention Networks to permit multi-document fusion and filtering (Nie et al., 2020).
    • Static Listwise Reranking: Shortlists are reranked in a single pass, with the top-kk passed downstream (Sharifymoghaddam et al., 20 Jan 2026).
  3. Downstream Reasoning: The top reranked subset is provided to a chain-of-thought reasoning agent or a span-based answer extractor. If answerability is not achieved (e.g., no span exceeds a no-answer probability threshold), retrieval and reranking steps are repeated or the query is updated using extracted clues.

A stylized flow: Query qq \rightarrow Retrieve {D1,,DK}\{D_1,\ldots,D_K\} \rightarrow Rerank \rightarrow Top-kk \rightarrow Reason \rightarrow [Iterate or Halt].

3. Adaptive Reranking Algorithms

Modern IDRQA methods implement adaptive computation using explicit uncertainty quantification. Notable examples include:

  • AcuRank (Yoon et al., 24 May 2025): Maintains Gaussian beliefs xiN(μi,σi2+β2)x_i\sim\mathcal N(\mu_i,\sigma_i^2+\beta^2) over document relevance. At each iteration, it computes the probability sis_i that each document lies above the top-kk cutoff:

P(xi>t)=1Φ(tμiσi2+β2)P(x_i > t) = 1 - \Phi\left(\frac{t - \mu_i}{\sqrt{\sigma_i^2+\beta^2}}\right)

Documents with si(ϵ,1ϵ)s_i \in (\epsilon,1-\epsilon) (uncertainty band) are selected for batch reranking. LLM reranker outputs are mapped to “multiplayer match” outcomes, updating beliefs via closed-form TrueSkill updates:

μi=μi+σi2cλ,σi2=σi2(1σi2cν)\mu'_i = \mu_i + \frac{\sigma_i^2}{c}\lambda\,,\quad \sigma_i'^2 = \sigma_i^2\left(1 - \frac{\sigma_i^2}{c}\nu\right)

where cc aggregates batch variances. Iteration continues until few uncertain documents remain (C<τ|C|<\tau), the top-kk stabilizes, or a budget is exhausted.

  • Graph-based Reranking (Nie et al., 2020): Constructs an entity-graph of paragraphs, using attention-based message passing and entity pooling to propagate signals, enabling collective document reranking and filtering informed by global context.

Empirically, these adaptive schemes provide superior accuracy-efficiency trade-offs, with AcuRank yielding 20–50% fewer LLM calls than fixed-window or tournament baselines, and the graph-based method improving both supporting-paragraph recall and answer accuracy in multi-hop QA settings.

4. Experimental Results and Cost-Effectiveness

Table-based evaluations across benchmarks (TREC-DL, BEIR, HotpotQA, BrowseComp-Plus) consistently demonstrate the utility of IDRQA. A summary of key findings:

Method Calls (avg) NDCG@10 Answer EM Key Result Context
SlidingWindow-1 8.8 54.3 BM25-100 + RankZephyr (Yoon et al., 24 May 2025)
AcuRank-9 9 54.6 Matches call budget, ↑ NDCG
AcuRank (full) 19.7 55.5 Exceeds SW-3 at lower cost
IDRQA (HotpotQA) 62.5 Test EM, surpasses Multi-hop Dense Retrieval (Nie et al., 2020)
IDRQA (SQuAD Open) 56.6 EM; Dense Passage Retriever: 36.7

The Effective Token Cost (ETC) metric (Sharifymoghaddam et al., 20 Jan 2026),

ETC=Inputnc+αInputc+βOutputt,\mathrm{ETC} = \mathrm{Input}_{nc} + \alpha\cdot\mathrm{Input}_c + \beta\cdot \mathrm{Output}_t,

captures practical trade-offs between reranking depth dd, reasoning output, and caching. Empirical studies confirm that moderate reranking (d=10d=10–$20$) consistently boosts end-to-end accuracy more efficiently than allocating extra budget to long-form reasoning.

Beyond d=20d=20, accuracy gains diminish relative to cost. For example, oss-120b with d=20d=20 and medium reasoning matches high-reasoning accuracy while reducing ETC by over 50%.

5. Adaptive Stopping and Pipeline Dynamics

IDRQA pipelines typically terminate based on dynamic criteria:

  • Confidence-based halting: Iteration stops when the top-kk set is highly probable per document posteriors (e.g., all sis_i outside the uncertainty band) (Yoon et al., 24 May 2025).
  • Answerability threshold: In QA pipelines, reranking terminates when the answer span predicted by the reader exceeds the “no answer” confidence, i.e., Pans>PnaP_{ans} > P_{na} (Nie et al., 2020).
  • Budget exhaustion: Allows anytime behavior and explicit tuning of computational cost.

Studies show that in most cases (∼85%), IDRQA converges within two hops or reranking rounds, with more complex or uncertain queries consuming a larger computational budget. This ensures efficiency and supports cost-sensitive deployment in production settings.

6. Comparative Analyses with Fixed-Computation Baselines

IDRQA systematically outperforms fixed-computation baselines by balancing effort against instance difficulty and rank uncertainty:

  • Sliding Window: Makes fixed passes over the document list, incurring wasteful computation on “easy” instances while insufficiently resolving harder ones.
  • Tournament and TrueSkill-Static: Run a set number of rounds or operate on fixed candidate subsets, leading to inefficiency and suboptimal accuracy particularly for long-tailed query difficulty distributions.
  • IDRQA adaptive approaches: Focus LLM calls on ambiguous segments and adapt computation to query complexity, consistently producing superior Pareto frontiers in the accuracy–efficiency plane (Yoon et al., 24 May 2025, Sharifymoghaddam et al., 20 Jan 2026).

7. Extensions and Empirical Insights

Noteworthy extensions include:

  • Graph-based multi-hop question answering: IDRQA’s iterative retrieval and reranking with graph-structured aggregation unifies single- and multi-hop QA, yielding state-of-the-art performance on benchmarks such as HotpotQA and NQ Open (Nie et al., 2020). Ablations confirm the necessity of both graph-based reranking and iterative operation for maximal paragraph recall and supporting-fact F1.
  • Cost-effective deep search agents: Incorporation of IDRQA as a preliminary stage in deep research agents allows substantial reductions in output token cost and end-to-end reasoning expense with minimal (or no) accuracy loss. Accurate calibration and recall are attained at substantially lower effective token cost when targeting moderate reranking depths (d=10d=10 or $20$) (Sharifymoghaddam et al., 20 Jan 2026).
  • Correlation with query difficulty: Harder queries (by WIG uncertainty) naturally allocate more reranker calls, while trivial ones terminate quickly, confirming IDRQA’s computation allocation rationale (Yoon et al., 24 May 2025).

In summary, IDRQA provides a principled and generalizable iterative reranking strategy, enabling both accuracy gains and computational savings across retrieval-based and multi-hop QA systems (Yoon et al., 24 May 2025, Sharifymoghaddam et al., 20 Jan 2026, Nie et al., 2020).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Iterative Document Reranking (IDRQA).