IAUNet: Instance-Aware U-Net

πŸ“… 2025-08-03
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the challenging instance segmentation of overlapping and multi-scale cells in biomedical images, this paper proposes IAUNetβ€”the first lightweight, end-to-end framework that deeply integrates a U-Net encoder with a query-based mechanism. Methodologically, IAUNet introduces a lightweight convolutional Pixel Decoder and a multi-scale Transformer Query Decoder to enable efficient instance-aware decoding within the U-Net feature space; it further incorporates cross-scale feature fusion and object queries to significantly improve separation accuracy for highly overlapping cells. As a key contribution, we construct and publicly release Revvity 2025, a high-quality, fully annotated cell segmentation dataset. Extensive experiments demonstrate that IAUNet consistently outperforms state-of-the-art CNN-, Transformer-, and query-based methods across multiple public benchmarks and Revvity 2025, establishing a new performance baseline for cell instance segmentation.

Technology Category

Application Category

πŸ“ Abstract
Instance segmentation is critical in biomedical imaging to accurately distinguish individual objects like cells, which often overlap and vary in size. Recent query-based methods, where object queries guide segmentation, have shown strong performance. While U-Net has been a go-to architecture in medical image segmentation, its potential in query-based approaches remains largely unexplored. In this work, we present IAUNet, a novel query-based U-Net architecture. The core design features a full U-Net architecture, enhanced by a novel lightweight convolutional Pixel decoder, making the model more efficient and reducing the number of parameters. Additionally, we propose a Transformer decoder that refines object-specific features across multiple scales. Finally, we introduce the 2025 Revvity Full Cell Segmentation Dataset, a unique resource with detailed annotations of overlapping cell cytoplasm in brightfield images, setting a new benchmark for biomedical instance segmentation. Experiments on multiple public datasets and our own show that IAUNet outperforms most state-of-the-art fully convolutional, transformer-based, and query-based models and cell segmentation-specific models, setting a strong baseline for cell instance segmentation tasks. Code is available at https://github.com/SlavkoPrytula/IAUNet
Problem

Research questions and friction points this paper is trying to address.

Improves instance segmentation of overlapping biomedical objects like cells
Enhances U-Net efficiency with lightweight Pixel decoder and Transformer decoder
Introduces new dataset for benchmarking cell instance segmentation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Lightweight convolutional Pixel decoder enhances efficiency
Transformer decoder refines multi-scale object features
Query-based U-Net architecture for instance segmentation
πŸ”Ž Similar Papers
No similar papers found.