π€ AI Summary
This work addresses the limitation of existing retrievers that rely on surface-level similarity when handling complex logical queries, which constrains the performance of large language models in knowledge-intensive tasks. To overcome this, the authors propose LORE, a method that leverages fine-grained contrastive learning to activate the modelβs inherent logical reasoning capabilities, steering embedding representations to align with deep logical structures rather than shallow semantic cues. LORE requires no external supervision, additional resources, or pre-retrieval analysis, while preserving index compatibility and computational efficiency. Experimental results demonstrate that LORE significantly improves both retrieval accuracy and downstream generation quality across multiple knowledge-intensive benchmarks. The code and datasets are publicly released.
π Abstract
Large language models (LLMs) struggle in knowledge-intensive tasks, as retrievers often overfit to surface similarity and fail on queries involving complex logical relations. The capacity for logical analysis is inherent in model representations but remains underutilized in standard training. LORE (Logic ORiented Retriever Enhancement) introduces fine-grained contrastive learning to activate this latent capacity, guiding embeddings toward evidence aligned with logical structure rather than shallow similarity. LORE requires no external upervision, resources, or pre-retrieval analysis, remains index-compatible, and consistently improves retrieval utility and downstream generation while maintaining efficiency. The datasets and code are publicly available at https://github.com/mazehart/Lore-RAG.