🤖 AI Summary
To address the dual challenges of inefficient large language model (LLM) inference and poor generalization of lightweight dense retrievers, this paper proposes an LLM-driven framework for training compact dense retrievers. Methodologically: (i) a pruned LLM serves as the backbone encoder, balancing representational capacity with computational efficiency; (ii) we introduce the first LLM-based diverse generative data augmentation strategy, substantially improving generalization across multilingual and long-context scenarios; and (iii) we design a single-stage contrastive learning paradigm for end-to-end retrieval optimization. Experiments demonstrate that our approach consistently outperforms conventional dense retrievers across multiple tasks and multilingual benchmarks. Crucially, it achieves strong generalization and high retrieval accuracy while maintaining low latency and a small parameter count—effectively unifying efficiency, scalability, and performance.
📝 Abstract
Large language models (LLMs) have demonstrated strong effectiveness and robustness while fine-tuned as dense retrievers. However, their large parameter size brings significant inference time computational challenges, including high encoding costs for large-scale corpora and increased query latency, limiting their practical deployment. While smaller retrievers offer better efficiency, they often fail to generalize effectively with limited supervised fine-tuning data. In this work, we introduce DRAMA, a training framework that leverages LLMs to train smaller generalizable dense retrievers. In particular, we adopt pruned LLMs as the backbone and train on diverse LLM-augmented data in a single-stage contrastive learning setup. Experiments show that DRAMA offers better multilingual and long-context capabilities than traditional encoder-based retrievers, and achieves strong performance across multiple tasks and languages. These highlight the potential of connecting the training of smaller retrievers with the growing advancements in LLMs, bridging the gap between efficiency and generalization.