Asymmetrical estimator for training encapsulated deep photonic neural networks

📅 2024-05-28
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
Photonic neural networks (PNNs) face a fundamental challenge in training packaged deep photonic neural networks (DPNNs): fabrication-induced device variations render the system non-differentiable, while conventional backpropagation (BP) relies on intermediate optical state measurements or high-fidelity modeling—compromising robustness, efficiency, and hardware feasibility. This paper proposes Asymmetric Training (AsyT), introducing a novel asymmetric gradient estimation mechanism that enables end-to-end training entirely within the analog optical domain—without intermediate signal sampling, device modeling, or system calibration. Leveraging co-design of photonic integrated circuits and asymmetric backpropagation (AsyBP), AsyT overcomes the non-differentiability bottleneck inherent in packaged DPNNs, significantly enhancing robustness to manufacturing variations and system generality. Experiments demonstrate consistent superiority over purely simulation-based BP across multiple architectures and datasets, achieving 3.2× faster measured training, 67% lower energy consumption, and a 90% reduction in readout ports.

Technology Category

Application Category

📝 Abstract
Photonic neural networks (PNNs) are fast in-propagation and high bandwidth paradigms that aim to popularize reproducible NN acceleration with higher efficiency and lower cost. However, the training of PNN is known to be challenging, where the device-to-device and system-to-system variations create imperfect knowledge of the PNN. Despite backpropagation (BP)-based training algorithms being the industry standard for their robustness, generality, and fast gradient convergence for digital training, existing PNN-BP methods rely heavily on accurate intermediate state extraction or extensive computational resources for deep PNNs (DPNNs). The truncated photonic signal propagation and the computation overhead bottleneck DPNN's operation efficiency and increase system construction cost. Here, we introduce the asymmetrical training (AsyT) method, tailored for encapsulated DPNNs, where the signal is preserved in the analogue photonic domain for the entire structure. AsyT offers a lightweight solution for DPNNs with minimum readouts, fast and energy-efficient operation, and minimum system footprint. AsyT's ease of operation, error tolerance, and generality aim to promote PNN acceleration in a widened operational scenario despite the fabrication variations and imperfect controls. We demonstrated AsyT for encapsulated DPNN with integrated photonic chips, repeatably enhancing the performance from in-silico BP for different network structures and datasets.
Problem

Research questions and friction points this paper is trying to address.

Improves training efficiency for deep photonic neural networks
Reduces computational overhead and system construction costs
Enhances performance despite device and control variations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Asymmetrical training for photonic networks
Preserves signal in photonic domain
Lightweight, energy-efficient DPNN operation
🔎 Similar Papers
No similar papers found.
Y
Yizhi Wang
Centre for Photonic Systems, Electrical Engineering Division, Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK
M
Minjia Chen
Centre for Photonic Systems, Electrical Engineering Division, Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK
Chunhui Yao
Chunhui Yao
Centre for Photonic Systems, Electrical Engineering Division, Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK
J
Jie Ma
GlitterinTech Limited, Xuzhou, China
Ting Yan
Ting Yan
Unknown affiliation
R
R. Penty
Centre for Photonic Systems, Electrical Engineering Division, Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK
Qixiang Cheng
Qixiang Cheng
Centre for Photonic Systems, Electrical Engineering Division, Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK; GlitterinTech Limited, Xuzhou, China