🤖 AI Summary
The exponential growth of scientific literature impedes domain-specific knowledge acquisition due to complex reasoning requirements, limited access to full texts, and extreme sparsity of target references within vast candidate pools. To address this, we propose a deep reinforcement learning (DRL)-based approach for sparse citation selection that emulates human-like, incremental knowledge construction under information constraints—relying solely on local textual cues such as titles and abstracts to efficiently identify the most informative papers. This work represents the first systematic application of DRL to sparse citation selection in scientific literature. Evaluated on drug–gene relationship discovery, our method significantly improves retrieval precision and accelerates knowledge discovery, demonstrating that both the model and humans can effectively build structured knowledge from restricted textual inputs. The approach establishes a novel paradigm for low-resource scientific intelligence mining.
📝 Abstract
The rapid expansion of scientific literature makes it increasingly difficult to acquire new knowledge, particularly in specialized domains where reasoning is complex, full-text access is restricted, and target references are sparse among a large set of candidates. We present a Deep Reinforcement Learning framework for sparse reference selection that emulates human knowledge construction, prioritizing which papers to read under limited time and cost. Evaluated on drug--gene relation discovery with access restricted to titles and abstracts, our approach demonstrates that both humans and machines can construct knowledge effectively from partial information.