计算机科学
分割
图像分割
人工智能
块(置换群论)
匹配(统计)
编码(集合论)
建筑
图像(数学)
模式识别(心理学)
计算机视觉
视觉艺术
艺术
统计
集合(抽象数据类型)
程序设计语言
数学
几何学
作者
Jianfeng Qin,Jun Yu,Jinhai Xiang,Xinwei He,Wen Zhang,Lulu Wu
标识
DOI:10.1109/bibm58861.2023.10385828
摘要
In medical image segmentation, U-Net has consistently played a vital role. Recently, the U-Net networks based on the Vision Transformer (ViT) architecture have become more and more popular. ViT exhibits superior capabilities in handling long-range dependencies and capturing global contextual information. However, it requires significant computational cost, and does not explore the optimal matching and the potential dependencies between different patches. To address the aforementioned issues, we propose a novel network framework, called AiA-UNet, for medical image segmentation. The AiA-UNet makes two main contributions. A convolutional self-attention mechanism is proposed to replace the self-attention module in ViT, effectively reducing computational complexity. Moreover, an Attention in Attention module (AiA) is applied within the ViT block. Experimental results on the Synapse multi-organ segmentation dataset demonstrate that AiA-UNet outperforms Trans-UNet by 5.40% and Swin-UNet by 3.75%. Code and models are available at https://github.com/xiaoqin1998/AiA-UNet.
科研通智能强力驱动
Strongly Powered by AbleSci AI