Papers
Topics
Authors
Recent
Search
2000 character limit reached

Doc-Researcher: A Unified System for Multimodal Document Parsing and Deep Research

Published 24 Oct 2025 in cs.IR and cs.CL | (2510.21603v1)

Abstract: Deep Research systems have revolutionized how LLMs solve complex questions through iterative reasoning and evidence gathering. However, current systems remain fundamentally constrained to textual web data, overlooking the vast knowledge embedded in multimodal documents Processing such documents demands sophisticated parsing to preserve visual semantics (figures, tables, charts, and equations), intelligent chunking to maintain structural coherence, and adaptive retrieval across modalities, which are capabilities absent in existing systems. In response, we present Doc-Researcher, a unified system that bridges this gap through three integrated components: (i) deep multimodal parsing that preserves layout structure and visual semantics while creating multi-granular representations from chunk to document level, (ii) systematic retrieval architecture supporting text-only, vision-only, and hybrid paradigms with dynamic granularity selection, and (iii) iterative multi-agent workflows that decompose complex queries, progressively accumulate evidence, and synthesize comprehensive answers across documents and modalities. To enable rigorous evaluation, we introduce M4DocBench, the first benchmark for Multi-modal, Multi-hop, Multi-document, and Multi-turn deep research. Featuring 158 expert-annotated questions with complete evidence chains across 304 documents, M4DocBench tests capabilities that existing benchmarks cannot assess. Experiments demonstrate that Doc-Researcher achieves 50.6% accuracy, 3.4xbetter than state-of-the-art baselines, validating that effective document research requires not just better retrieval, but fundamentally deep parsing that preserve multimodal integrity and support iterative research. Our work establishes a new paradigm for conducting deep research on multimodal document collections.

Summary

  • The paper introduces a unified system that leverages multimodal parsing and iterative workflows to extract and synthesize evidence from complex documents.
  • It employs dynamic, hybrid retrieval strategies and layout-aware parsing techniques, achieving a 50.6% accuracy that significantly outperforms baseline methods.
  • The system establishes the M4DocBench benchmark, setting a new standard for evaluating multimodal, multi-hop document research capabilities.

Doc-Researcher: A Unified System for Multimodal Document Parsing and Deep Research

Introduction

The "Doc-Researcher" paper presents a unified system aimed at bridging the gap between current deep research systems and multimodal document processing. Traditional systems have been largely limited to textual web data and overlook the multimodal nature of documents such as scientific papers, technical reports, and financial documents that contain critical information embedded in figures, tables, charts, and equations. Doc-Researcher integrates multimodal document parsing with advanced multi-agent deep research capabilities to address this limitation, facilitating complex query decomposition and iterative evidence synthesis across diverse document collections. Figure 1

Figure 1: An typical use case of multimodal doc deep research. The user asks a multi-hop question in the context of multi-turn conversations, where ground-truth evidence spans across multiple documents and modalities. The demo conversations, evidence, and answer are shortened for simplicity.

System Architecture

Doc-Researcher is structured into three main components:

  1. Deep Multimodal Parsing: This component preserves layout structure and visual semantics via MinerU for layout-aware document analysis and intelligent chunking strategies.
  2. Systematic Retrieval Architecture: Doc-Researcher supports text-only, vision-only, and hybrid retrieval paradigms with dynamic granularity selection which adapts retrieval strategies to query characteristics (Figure 2).
  3. Iterative Multi-Agent Workflows: These workflows decompose complex queries, progressively accumulate evidence, and synthesize comprehensive answers across documents and modalities. Figure 2

    Figure 2: Doc-Researcher Architecture: (a) multimodal deep parsing and indexing, and (b) multimodal deep research.

Evaluation and Benchmarking

To validate the system, the paper introduces M4DocBench, a novel benchmark designed for evaluating multi-modal, multi-hop, multi-document, and multi-turn deep research capabilities. M4DocBench comprises 158 expert-annotated questions across 304 documents, enforcing rigorous testing that existing benchmarks cannot assess.

Experimental Results

The experiments demonstrate that Doc-Researcher achieves a substantial accuracy of 50.6%, which is 3.4 times better than state-of-the-art baselines. This underscores that effective document research requires not just improved retrieval but fundamentally deep parsing to preserve multimodal integrity and support iterative research. Figure 3

Figure 3: Accuracy and retrieval performance of Doc-Researcher using different backbone LLMs and retrieval schemes.

Doc-Researcher's architecture facilitates efficient learning, large-scale document retrieval, and complex multi-modal query processing, as evidenced by the increased performance metrics compared to existing systems (Figure 4). Figure 4

Figure 4: Performance with increasing search depth.

Implications and Future Work

The implications of Doc-Researcher are substantial, establishing a new paradigm for conducting deep research on multimodal document collections. It integrates sophisticated parsing and retrieval systems, proving effective for complex reasoning across multimodal and multi-document scenarios. Future directions include optimizing the system for better cross-modal reasoning mechanisms and developing more compact multimodal embeddings to reduce computational overhead.

Conclusion

In conclusion, Doc-Researcher successfully addresses significant gaps in multimodal document processing systems, providing a robust framework for deep research in complex, multimodal environments. The development of Doc-Researcher marks a pivotal shift towards leveraging LLMs for enhanced document understanding, promising advancements in multimodal research capabilities and setting a benchmark for future systems.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 12 likes about this paper.