🤖 AI Summary
This work addresses the limitation of fixed residual connections in multi-modal MRI brain tumor segmentation, which struggle to adaptively fuse modality-specific information. To overcome this, we propose Hyper-Connections—a plug-and-play dynamic feature fusion module—and introduce it for the first time into 3D multi-modal brain tumor segmentation. The method is seamlessly integrated into various architectures, including nnU-Net, SwinUNETR, VT-UNet, U-Net, and U-Net++, enabling end-to-end training with negligible parameter overhead. Our approach significantly enhances sensitivity to clinically critical modalities such as T1ce and FLAIR and improves boundary delineation of enhancing tumor subregions. Evaluated on the BraTS 2021 dataset, it achieves up to a 1.03% absolute improvement in average Dice score across all 3D models, with particularly pronounced gains in volume-level assessments.
📝 Abstract
We present the first study of Hyper-Connections (HC) for volumetric multi-modal brain tumor segmentation, integrating them as a drop-in replacement for fixed residual connections across five architectures: nnU-Net, SwinUNETR, VT-UNet, U-Net, and U-Netpp. Dynamic HC consistently improves all 3D models on the BraTS 2021 dataset, yielding up to +1.03 percent mean Dice gain with negligible parameter overhead. Gains are most pronounced in the Enhancing Tumor sub-region, reflecting improved fine-grained boundary delineation. Modality ablation further reveals that HC-equipped models develop sharper sensitivity toward clinically dominant sequences, specifically T1ce for Tumor Core and Enhancing Tumor, and FLAIR for Whole Tumor, a behavior absent in fixed-connection baselines and consistent across all architectures. In 2D settings, improvements are smaller and configuration-sensitive, suggesting that volumetric spatial context amplifies the benefit of adaptive aggregation. These results establish HC as a simple, efficient, and broadly applicable mechanism for multi-modal feature fusion in medical image segmentation.