🤖 AI Summary
This work addresses three key challenges in zero-shot clinical named entity recognition (NER): fine-grained entity omission, class imbalance, and low recall for rare/long-tail entities. To this end, we propose the Entity Decomposition and Filtering (EDF) framework—the first of its kind—decoupling open-domain NER into two stages: subtype-aware retrieval and collaborative result filtering. EDF leverages open-source, NER-specialized large language models and integrates task decomposition with type-aware mechanisms. Extensive experiments across multiple clinical benchmarks demonstrate that EDF consistently outperforms all baseline methods across all evaluation metrics, model configurations, and entity types. Notably, it achieves substantial improvements in the recognition accuracy of rare and long-tail clinical entities, significantly enhancing zero-shot generalization capability.
📝 Abstract
Clinical named entity recognition (NER) aims to retrieve important entities within clinical narratives. Recent works have demonstrated that large language models (LLMs) can achieve strong performance in this task. While previous works focus on proprietary LLMs, we investigate how open NER LLMs, trained specifically for entity recognition, perform in clinical NER. Our initial experiment reveals significant contrast in performance for some clinical entities and how a simple exploitment on entity types can alleviate this issue. In this paper, we introduce a novel framework, entity decomposition with filtering, or EDF. Our key idea is to decompose the entity recognition task into several retrievals of entity sub-types and then filter them. Our experimental results demonstrate the efficacies of our framework and the improvements across all metrics, models, datasets, and entity types. Our analysis also reveals substantial improvement in recognizing previously missed entities using entity decomposition. We further provide a comprehensive evaluation of our framework and an in-depth error analysis to pave future works.