计算机科学
编码器
分割
人工智能
变压器
图像分割
推论
图像分辨率
模式识别(心理学)
计算机视觉
量子力学
操作系统
物理
电压
作者
Nan Wang,Shaohui Lin,Xiaoxiao Li,Ke Li,Yunhang Shen,Yue Gao,Lizhuang Ma
标识
DOI:10.1109/tmi.2023.3264433
摘要
U-Nets have achieved tremendous success in medical image segmentation. Nevertheless, it may have limitations in global (long-range) contextual interactions and edge-detail preservation. In contrast, the Transformer module has an excellent ability to capture long-range dependencies by leveraging the self-attention mechanism into the encoder. Although the Transformer module was born to model the long-range dependency on the extracted feature maps, it still suffers high computational and spatial complexities in processing high-resolution 3D feature maps. This motivates us to design an efficient Transformer-based UNet model and study the feasibility of Transformer-based network architectures for medical image segmentation tasks. To this end, we propose to self-distill a Transformer-based UNet for medical image segmentation, which simultaneously learns global semantic information and local spatial-detailed features. Meanwhile, a local multi-scale fusion block is first proposed to refine fine-grained details from the skipped connections in the encoder by the main CNN stem through self-distillation, only computed during training and removed at inference with minimal overhead. Extensive experiments on BraTS 2019 and CHAOS datasets show that our MISSU achieves the best performance over previous state-of-the-art methods. Code and models are available at: https://github.com/wangn123/MISSU.git.
科研通智能强力驱动
Strongly Powered by AbleSci AI