Retriever-and-Memory: Towards Adaptive Note-Enhanced Retrieval-Augmented Generation

๐Ÿ“… 2024-10-11
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing RAG methods for complex open-domain question answering rely on a single LLM invocation to determine retrieval timing and directly generate answers, failing to accurately model dynamic information needs and resulting in inefficient knowledge utilization. Method: We propose Note-RAG, an adaptive note-augmented RAG framework featuring a novel โ€œretrieverโ€“memoryโ€ dual-module paradigm. It enables progressive knowledge construction through iterative retrieval and LLM-driven generation of structured notes; further, it introduces a note-state-aware adaptive memory review and exploration termination mechanism to improve retrieval timing decisions and knowledge interaction quality. Contribution/Results: Note-RAG achieves significant improvements over state-of-the-art RAG baselines across five challenging QA benchmarks. Ablation studies validate the effectiveness of each component. The code and datasets are publicly released.

Technology Category

Application Category

๐Ÿ“ Abstract
Retrieval-Augmented Generation (RAG) mitigates issues of the factual errors and hallucinated outputs generated by Large Language Models (LLMs) in open-domain question-answering tasks (OpenQA) via introducing external knowledge. For complex QA, however, existing RAG methods use LLMs to actively predict retrieval timing and directly use the retrieved information for generation, regardless of whether the retrieval timing accurately reflects the actual information needs, or sufficiently considers prior retrieved knowledge, which may result in insufficient information gathering and interaction, yielding low-quality answers. To address these, we propose a generic RAG approach called Adaptive Note-Enhanced RAG (Adaptive-Note) for complex QA tasks, which includes the iterative information collector, adaptive memory reviewer, and task-oriented generator, while following a new Retriever-and-Memory paradigm. Specifically, Adaptive-Note introduces an overarching view of knowledge growth, iteratively gathering new information in the form of notes and updating them into the existing optimal knowledge structure, enhancing high-quality knowledge interactions. In addition, we employ an adaptive, note-based stop-exploration strategy to decide"what to retrieve and when to stop"to encourage sufficient knowledge exploration. We conduct extensive experiments on five complex QA datasets, and the results demonstrate the superiority and effectiveness of our method and its components. The code and data are at https://github.com/thunlp/Adaptive-Note.
Problem

Research questions and friction points this paper is trying to address.

Mitigates factual errors in LLMs for QA using external knowledge
Improves adaptive retrieval by refining knowledge through note-centric approach
Enhances answer generation by leveraging high-quality accumulated notes
Innovation

Methods, ideas, or system contributions that make the work stand out.

Note-centric adaptive retrieval framework
Iterative knowledge growth assessment
High-density quality knowledge gathering
๐Ÿ”Ž Similar Papers
No similar papers found.
Ruobing Wang
Ruobing Wang
University of Chinese Academy of Sciences
LLMRAG
D
Daren Zha
Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China; School of Cyber Security, University of Chinese Academy of Sciences, Beijing, China
Shi Yu
Shi Yu
Tsinghua University
LLMRAGInformation RetrievalNatural Language Processing
Qingfei Zhao
Qingfei Zhao
University of the Chinese Academy of Sciences
Natural Language ProcessingArtificial Intelligence
Y
Yuxuan Chen
South China University of Technology, China
Y
Yixuan Wang
Department of Computer Science and Technology, Institute for AI, Tsinghua University, China
S
Shuo Wang
Department of Computer Science and Technology, Institute for AI, Tsinghua University, China
Yukun Yan
Yukun Yan
Tsinghua University
Large Language Model
Zhenghao Liu
Zhenghao Liu
Northeastern University
NLPInformation Retrieval
X
Xu Han
Department of Computer Science and Technology, Institute for AI, Tsinghua University, China
Z
Zhiyuan Liu
Department of Computer Science and Technology, Institute for AI, Tsinghua University, China
Maosong Sun
Maosong Sun
Professor of Computer Science and Technology, Tsinghua University
Natural Language ProcessingArtificial IntelligenceSocial Computing