🤖 AI Summary
Privacy preservation and statistical heterogeneity pose significant challenges for dynamic Bayesian network (DBN) structure learning from distributed time-series data in federated settings.
Method: We propose the first continuous-optimization-based federated DBN structure learning framework. To address inter-client statistical heterogeneity, we design PFDBNL—a personalized regularization method that integrates proximal operator constraints with parameter-level model exchange, enabling collaborative modeling across both homogeneous and heterogeneous horizontally partitioned clients while ensuring data locality and privacy.
Contribution/Results: PFDBNL achieves efficient and robust estimation of the global DBN structure without centralized data access. Extensive experiments on synthetic and real-world datasets demonstrate that our method significantly outperforms existing baselines, particularly in typical federated scenarios characterized by a large number of clients and limited local samples—yielding substantial improvements in structural recovery accuracy.
📝 Abstract
Traditionally, learning the structure of a Dynamic Bayesian Network has been centralized, requiring all data to be pooled in one location. However, in real-world scenarios, data are often distributed across multiple entities (e.g., companies, devices) that seek to collaboratively learn a Dynamic Bayesian Network while preserving data privacy and security. More importantly, due to the presence of diverse clients, the data may follow different distributions, resulting in data heterogeneity. This heterogeneity poses additional challenges for centralized approaches. In this study, we first introduce a federated learning approach for estimating the structure of a Dynamic Bayesian Network from homogeneous time series data that are horizontally distributed across different parties. We then extend this approach to heterogeneous time series data by incorporating a proximal operator as a regularization term in a personalized federated learning framework. To this end, we propose exttt{FDBNL} and exttt{PFDBNL}, which leverage continuous optimization, ensuring that only model parameters are exchanged during the optimization process. Experimental results on synthetic and real-world datasets demonstrate that our method outperforms state-of-the-art techniques, particularly in scenarios with many clients and limited individual sample sizes.