Digital Twin Calibration with Model-Based Reinforcement Learning

📅 2025-01-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Industrial process control in data-scarce, dynamically complex, and highly uncertain domains—such as biopharmaceutical manufacturing—remains challenging due to limited observational data and intrinsic system nonlinearity. Method: This paper proposes an Actor-Simulator dual-component model-based reinforcement learning framework that jointly optimizes digital twin calibration and optimal control policy design—a first-of-its-kind integration. The approach synergistically combines Bayesian system identification, stochastic nonlinear dynamical modeling, and an adaptive exploration-exploitation mechanism, where control policy performance actively guides efficient data acquisition. Convergence is theoretically guaranteed under mild assumptions. Results: Evaluated on a biopharmaceutical simulation benchmark, the method significantly reduces model prediction error, enhances closed-loop robustness, and improves sample efficiency. It achieves comprehensive performance superiority over state-of-the-art baselines across all key metrics.

Technology Category

Application Category

📝 Abstract
This paper presents a novel methodological framework, called the Actor-Simulator, that incorporates the calibration of digital twins into model-based reinforcement learning for more effective control of stochastic systems with complex nonlinear dynamics. Traditional model-based control often relies on restrictive structural assumptions (such as linear state transitions) and fails to account for parameter uncertainty in the model. These issues become particularly critical in industries such as biopharmaceutical manufacturing, where process dynamics are complex and not fully known, and only a limited amount of data is available. Our approach jointly calibrates the digital twin and searches for an optimal control policy, thus accounting for and reducing model error. We balance exploration and exploitation by using policy performance as a guide for data collection. This dual-component approach provably converges to the optimal policy, and outperforms existing methods in extensive numerical experiments based on the biopharmaceutical manufacturing domain.
Problem

Research questions and friction points this paper is trying to address.

Biopharmaceutical Process Control
Data Scarcity
Optimization Methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Digital Twin Calibration
Actor-Critic Method
Biopharmaceutical Processes
🔎 Similar Papers
No similar papers found.
H
Hua Zheng
Northeastern University
W
Wei Xie
Northeastern University
Ilya O. Ryzhov
Ilya O. Ryzhov
Robert H. Smith School of Business, University of Maryland
stochastic optimizationoptimal learningstatistical learning
K
Keilung Choy
Northeastern University