🤖 AI Summary
This paper addresses collaborative adaptive control of multiple heterogeneous systems under model uncertainty and adversarial data corruption. To counter malicious node interference and system heterogeneity, we propose a unified framework integrating clustering-driven multi-task learning, robust system identification, and attack-resilient aggregation, coupled with a confidence-equivalent control strategy for distributed optimization. Theoretically, under a bounded fraction of malicious nodes, the expected regret decays inversely with the number of honest systems, yielding a non-asymptotic regret bound. Moreover, the clustering mechanism explicitly quantifies how intra-cluster heterogeneity and adversarial behavior degrade LQR performance. Our approach significantly enhances robustness, learning efficiency, and security of distributed control systems, establishing a novel paradigm for trustworthy multi-agent control.
📝 Abstract
We study adversarially robust multitask adaptive linear quadratic control; a setting where multiple systems collaboratively learn control policies under model uncertainty and adversarial corruption. We propose a clustered multitask approach that integrates clustering and system identification with resilient aggregation to mitigate corrupted model updates. Our analysis characterizes how clustering accuracy, intra-cluster heterogeneity, and adversarial behavior affect the expected regret of certainty-equivalent (CE) control across LQR tasks. We establish non-asymptotic bounds demonstrating that the regret decreases inversely with the number of honest systems per cluster and that this reduction is preserved under a bounded fraction of adversarial systems within each cluster.