🤖 AI Summary
Conventional color-sequential fluorescence microscopy suffers from slow imaging speed and limited spectral channels, hindering simultaneous, dynamic observation of multiple subcellular structures. To address this, we propose AEMS-Net—a deep learning framework that reconstructs mitochondria and microtubules synchronously and in real time from single-channel widefield fluorescence images. Methodologically, AEMS-Net introduces a novel interpretable architecture integrating brightness-adaptive layers with attention mechanisms; leveraging the Kolmogorov–Arnold representation theorem, it explicitly decouples latent features into morphology-interpretable univariate functions, achieving structural-level interpretability. Experiments demonstrate a 50% reduction in imaging time and over 30% improvement in reconstruction fidelity compared to state-of-the-art deep learning methods. This work establishes a new paradigm for long-term, reliable investigation of dynamic interplay among multiple subcellular structures in live cells.
📝 Abstract
Live-cell imaging of multiple subcellular structures is essential for understanding subcellular dynamics. However, the conventional multi-color sequential fluorescence microscopy suffers from significant imaging delays and limited number of subcellular structure separate labeling, resulting in substantial limitations for real-time live-cell research applications. Here, we present the Adaptive Explainable Multi-Structure Network (AEMS-Net), a deep-learning framework that enables simultaneous prediction of two subcellular structures from a single image. The model normalizes staining intensity and prioritizes critical image features by integrating attention mechanisms and brightness adaptation layers. Leveraging the Kolmogorov-Arnold representation theorem, our model decomposes learned features into interpretable univariate functions, enhancing the explainability of complex subcellular morphologies. We demonstrate that AEMS-Net allows real-time recording of interactions between mitochondria and microtubules, requiring only half the conventional sequential-channel imaging procedures. Notably, this approach achieves over 30% improvement in imaging quality compared to traditional deep learning methods, establishing a new paradigm for long-term, interpretable live-cell imaging that advances the ability to explore subcellular dynamics.