DM-QPMNET: Dual-modality fusion network for cell segmentation in quantitative phase microscopy

📅 2025-10-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In single-shot quantitative phase microscopy (ssQPM), cell segmentation is hindered by noise, high cell density, poor robustness of conventional thresholding methods, and the inability of naive multimodal concatenation to capture cross-modal complementarity. To address these challenges, we propose a Dual-Encoder Attention Fusion Network (DAF-Net). DAF-Net employs two independent encoders to extract features from polarization-intensity images and quantitative phase maps, respectively; integrates a multi-head attention mechanism for content-aware feature-level fusion; and incorporates dual-source skip connections and modality normalization to enhance cross-modal complementarity and training stability. Extensive experiments under varying noise levels and cell densities demonstrate that DAF-Net consistently outperforms unimodal and simple concatenation baselines, achieving superior segmentation accuracy and robustness. The framework establishes a generalizable multimodal learning paradigm for ssQPM-based cellular analysis.

Technology Category

Application Category

📝 Abstract
Cell segmentation in single-shot quantitative phase microscopy (ssQPM) faces challenges from traditional thresholding methods that are sensitive to noise and cell density, while deep learning approaches using simple channel concatenation fail to exploit the complementary nature of polarized intensity images and phase maps. We introduce DM-QPMNet, a dual-encoder network that treats these as distinct modalities with separate encoding streams. Our architecture fuses modality-specific features at intermediate depth via multi-head attention, enabling polarized edge and texture representations to selectively integrate complementary phase information. This content-aware fusion preserves training stability while adding principled multi-modal integration through dual-source skip connections and per-modality normalization at minimal overhead. Our approach demonstrates substantial improvements over monolithic concatenation and single-modality baselines, showing that modality-specific encoding with learnable fusion effectively exploits ssQPM's simultaneous capture of complementary illumination and phase cues for robust cell segmentation.
Problem

Research questions and friction points this paper is trying to address.

Develops dual-modality fusion for cell segmentation in quantitative phase microscopy
Addresses noise sensitivity and cell density challenges in traditional thresholding methods
Enables complementary integration of polarized intensity images and phase maps
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-encoder network processes distinct imaging modalities
Multi-head attention fuses features at intermediate depth
Dual-source skip connections enable principled multi-modal integration
🔎 Similar Papers
No similar papers found.
Rajatsubhra Chakraborty
Rajatsubhra Chakraborty
Doctoral student, Department of Computer Science,University of North Carolina Charlotte
Computer Visionvision language modelsbias mitigationfairnessinterpretible AI systems.
A
Ana Espinosa-Momox
Department of Physics and Optical Science, University of North Carolina at Charlotte, NC, USA
R
Riley Haskin
Department of Physics and Optical Science, University of North Carolina at Charlotte, NC, USA
Depeng Xu
Depeng Xu
University of North Carolina at Charlotte
Machine LearningData PrivacyFairness
R
Rosario Porras-Aguilar
Department of Physics and Optical Science, University of North Carolina at Charlotte, NC, USA; Center for TAIMing AI, University of North Carolina at Charlotte, NC, USA