🤖 AI Summary
Large multimodal models (LMMs) exhibit insufficient generalization in autonomous driving due to misinterpretation of traffic knowledge, complexity of real-world road conditions, and diversity of vehicle states. Method: We propose a novel knowledge editing paradigm that avoids full retraining, and introduce ADS-Edit—the first multimodal knowledge editing benchmark tailored for autonomous driving. ADS-Edit comprises error patterns mined from real-world driving logs, diverse road scenarios, multi-vehicle dynamic states, and multimodal inputs. We further design a cross-modal consistency evaluation mechanism and knowledge-editing prompting strategies to enable precise, efficient correction of LMMs’ traffic cognition. Contribution/Results: Experiments demonstrate substantial improvements in critical capabilities—including traffic rule comprehension and anomaly response—while significantly enhancing robustness and generalization under dynamic driving conditions.
📝 Abstract
Recent advancements in Large Multimodal Models (LMMs) have shown promise in Autonomous Driving Systems (ADS). However, their direct application to ADS is hindered by challenges such as misunderstanding of traffic knowledge, complex road conditions, and diverse states of vehicle. To address these challenges, we propose the use of Knowledge Editing, which enables targeted modifications to a model's behavior without the need for full retraining. Meanwhile, we introduce ADS-Edit, a multimodal knowledge editing dataset specifically designed for ADS, which includes various real-world scenarios, multiple data types, and comprehensive evaluation metrics. We conduct comprehensive experiments and derive several interesting conclusions. We hope that our work will contribute to the further advancement of knowledge editing applications in the field of autonomous driving. Code and data are available in https://github.com/zjunlp/EasyEdit.