DTCCL: Disengagement-Triggered Contrastive Continual Learning for Autonomous Bus Planners

📅 2025-12-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Autonomous buses operating in open urban environments suffer from geographically concentrated, strategy-level failures in highly interactive regions; sparse human接管 data leads to severe overfitting in conventional imitation learning. Method: We propose a接管-driven contrastive continual learning paradigm: human接管 events trigger a cloud-edge collaborative closed loop; unsupervised, geography-aware policy representation optimization is achieved via causal-aware adversarial sample generation and route-context-preserving sampling; multi-agent scenario perturbation modeling further enhances generalization. Contribution/Results: Evaluated on real-world urban bus routes, our approach improves planning success rate by 48.6% over baseline retraining methods—demonstrating the first fully autonomous, scalable, strategy-level closed-loop evolution without manual intervention.

Technology Category

Application Category

📝 Abstract
Autonomous buses run on fixed routes but must operate in open, dynamic urban environments. Disengagement events on these routes are often geographically concentrated and typically arise from planner failures in highly interactive regions. Such policy-level failures are difficult to correct using conventional imitation learning, which easily overfits to sparse disengagement data. To address this issue, this paper presents a Disengagement-Triggered Contrastive Continual Learning (DTCCL) framework that enables autonomous buses to improve planning policies through real-world operation. Each disengagement triggers cloud-based data augmentation that generates positive and negative samples by perturbing surrounding agents while preserving route context. Contrastive learning refines policy representations to better distinguish safe and unsafe behaviors, and continual updates are applied in a cloud-edge loop without human supervision. Experiments on urban bus routes demonstrate that DTCCL improves overall planning performance by 48.6 percent compared with direct retraining, validating its effectiveness for scalable, closed-loop policy improvement in autonomous public transport.
Problem

Research questions and friction points this paper is trying to address.

Addresses planner failures in autonomous buses during disengagement events.
Improves policy learning using contrastive continual learning from sparse data.
Enables scalable, closed-loop policy updates without human supervision.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Disengagement-triggered cloud-based data augmentation for policy learning
Contrastive learning refines representations to distinguish safe behaviors
Cloud-edge continual updates enable unsupervised closed-loop improvement
🔎 Similar Papers
No similar papers found.
Y
Yanding Yang
School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China
Weitao Zhou
Weitao Zhou
Tsinghua University
Autonomous DrivingReinforcement Learning
J
Jinhai Wang
Dongfeng Motor Corporation Research and Development General Institute, Wuhan 430056, China
X
Xiaomin Guo
Dongfeng Yuexiang Technology Co., Ltd., Wuhan 430000, China
J
Junze Wen
School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China
X
Xiaolong Liu
Dongfeng Yuexiang Technology Co., Ltd., Wuhan 430000, China
L
Lang Ding
Dongfeng Yuexiang Technology Co., Ltd., Wuhan 430000, China
Zheng Fu
Zheng Fu
Tsinghua university
J
Jinyu Miao
School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China
Kun Jiang
Kun Jiang
Tsinghua University
autonomous driving
D
Diange Yang
School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China