Model-free reinforcement learning with noisy actions for automated experimental control in optics

📅 2024-05-24
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenging control problem of highly noisy and strongly nonlinear optical systems—specifically, laser coupling into single-mode fibers. We propose a model-free reinforcement learning (RL) approach that eliminates reliance on system identification or physical modeling. For the first time, we directly deploy Soft Actor-Critic (SAC) augmented with Truncated Quantile Critics (TQC) in a real-world optical experimental setup, enabling end-to-end online training and closed-loop control. The method learns an optimal policy solely through environmental interaction, without prior dynamical modeling or parameter estimation. Experimental results demonstrate stable achievement of 90% coupling efficiency—matching the performance of expert human operators—while exhibiting strong robustness to noise and nonlinearity. This validates the feasibility, robustness, and engineering practicality of model-free RL for complex, real-time optical control tasks.

Technology Category

Application Category

📝 Abstract
Setting up and controlling optical systems is often a challenging and tedious task. The high number of degrees of freedom to control mirrors, lenses, or phases of light makes automatic control challenging, especially when the complexity of the system cannot be adequately modeled due to noise or non-linearities. Here, we show that reinforcement learning (RL) can overcome these challenges when coupling laser light into an optical fiber, using a model-free RL approach that trains directly on the experiment without pre-training. By utilizing the sample-efficient algorithms Soft Actor-Critic (SAC) or Truncated Quantile Critics (TQC), our agent learns to couple with 90% efficiency, comparable to the human expert. We demonstrate that direct training on an experiment can replace extensive system modeling. Our result exemplifies RL's potential to tackle problems in optics, paving the way for more complex applications where full noise modeling is not feasible.
Problem

Research questions and friction points this paper is trying to address.

Automated control of optical systems
Model-free reinforcement learning application
Overcoming noise and non-linearities in optics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Model-free reinforcement learning
Direct experimental training
Soft Actor-Critic algorithm
🔎 Similar Papers
No similar papers found.
L
Lea Richtmann
Institute for Gravitational Physics, Leibniz University Hannover
V
Viktoria-S. Schmiesing
Institute for Theoretical Physics, Leibniz University Hannover
D
Dennis Wilken
Institute for Gravitational Physics, Leibniz University Hannover
J
Jan Heine
Institute of Photonics, Leibniz University Hannover
A
Aaron Tranter
Quantum Science and Technology, Australian National University
Avishek Anand
Avishek Anand
Associate Professor, TU Delft and L3S Research Center
interpretabilitygraph representation learningindexing textranking
Tobias J. Osborne
Tobias J. Osborne
Institüt für Theoretische Physik and L3S Research Center, Leibniz Universität Hannover
quantum field theoryquantum physicsquantum mechanicscondensed matter physicsquantum
M
M. Heurs
Institute for Gravitational Physics, Leibniz University Hannover