π€ AI Summary
Open-domain semantic parsing suffers from poor out-of-distribution generalization, particularly for unseen semantic units; existing neural models rely on heuristic rules and fail to effectively model novel lexical or compositional concepts. To address this, we propose RASPβa retrieval-augmented semantic parsing framework that enables deep synergy between large language models (LLMs) and external lexical knowledge bases. RASP dynamically retrieves relevant lexical knowledge during inference and injects it directly into the LLMβs decoding process, enabling real-time grounding of unfamiliar concepts. Evaluated across multiple open-domain semantic parsing benchmarks, RASP achieves near-doubling in zero-shot accuracy on unseen concepts, substantially outperforming conventional encoder-decoder baselines. These results demonstrate that tight integration of retrieval augmentation with LLMs significantly enhances both effectiveness and robustness in semantic generalization.
π Abstract
Open-domain semantic parsing remains a challenging task, as models often rely on heuristics and struggle to handle unseen concepts. In this paper, we investigate the potential of large language models (LLMs) for this task and introduce Retrieval-Augmented Semantic Parsing (RASP), a simple yet effective approach that integrates external lexical knowledge into the parsing process. Our experiments not only show that LLMs outperform previous encoder-decoder baselines for semantic parsing, but that RASP further enhances their ability to predict unseen concepts, nearly doubling the performance of previous models on out-of-distribution concepts. These findings highlight the promise of leveraging large language models and retrieval mechanisms for robust and open-domain semantic parsing.