π€ AI Summary
This work addresses the limitations of existing academic reasoning approaches, which predominantly rely on goal-directed retrieval and struggle to achieve coherent understanding and verification across entire scholarly papers. We introduce ScholScan, a novel benchmark that pioneers a scan-based academic reasoning paradigm, requiring models to read full papers and cross-verify content to identify inconsistencies. ScholScan encompasses 715 papers across 13 natural science domains, offering 1,800 fine-grained annotated questions with reasoning trajectories targeting nine types of consistency errors. Leveraging multimodal large language models (MLLMs), we evaluate 15 models under 24 input configurations using evidence localization, chain-of-thought annotations, and a unified evaluation protocol. Results reveal that current retrieval-augmented generation (RAG) methods yield no significant improvement, exposing systematic deficiencies of MLLMs in scan-based reasoning and underscoring ScholScanβs challenge and potential to guide future research.
π Abstract
With the rapid progress of multimodal large language models (MLLMs), AI already performs well at literature retrieval and certain reasoning tasks, serving as a capable assistant to human researchers, yet it remains far from autonomous research. The fundamental reason is that current work on academic paper reasoning is largely confined to a search-oriented paradigm centered on pre-specified targets, with reasoning grounded in relevance retrieval, which struggles to support researcher-style full-document understanding, reasoning, and verification. To bridge this gap, we propose \textbf{ScholScan}, a new benchmark for academic paper reasoning. ScholScan introduces a scan-oriented task setting that asks models to read and cross-check entire papers like human researchers, scanning the document to identify consistency issues. The benchmark comprises 1,800 carefully annotated questions drawn from nine error categories across 13 natural-science domains and 715 papers, and provides detailed annotations for evidence localization and reasoning traces, together with a unified evaluation protocol. We assessed 15 models across 24 input configurations and conducted a fine-grained analysis of MLLM capabilities for all error categories. Across the board, retrieval-augmented generation (RAG) methods yield no significant improvements, revealing systematic deficiencies of current MLLMs on scan-oriented tasks and underscoring the challenge posed by ScholScan. We expect ScholScan to be the leading and representative work of the scan-oriented task paradigm.