计算机科学
分割
变压器
人工智能
图像分割
数据一致性
雅卡索引
空间分析
上下文模型
模式识别(心理学)
交通标志识别
合成数据
特征提取
数据建模
计算机视觉
机器学习
稳健性(进化)
尺度空间分割
相互信息
背景(考古学)
计算机辅助设计
一致性(知识库)
实体造型
标记数据
数据挖掘
维数(图论)
遮罩(插图)
空间语境意识
上下文图像分类
频道(广播)
目标检测
作者
Xinyu Liu,Zhen Chen,Wuyang Li,Chenxin Li,Yixuan Yuan
标识
DOI:10.1109/tpami.2025.3640233
摘要
Transformers have shown remarkable performance in 3D medical image segmentation, but their high computational requirements and need for large amounts of labeled data limit their applicability. To address these challenges, we consider two crucial aspects: model efficiency and data efficiency. Specifically, we propose Light-UNETR, a lightweight transformer designed to achieve model efficiency. Light-UNETR features a Lightweight Dimension Reductive Attention (LIDR) module, which reduces spatial and channel dimensions while capturing both global and local features via multi-branch attention. Additionally, we introduce a Compact Gated Linear Unit (CGLU) to selectively control channel interaction with minimal parameters. Furthermore, we introduce a Contextual Synergic Enhancement (CSE) learning strategy, which aims to boost the data efficiency of Transformers. It first leverages the extrinsic contextual information to support the learning of unlabeled data with Attention-Guided Replacement, then applies Spatial Masking Consistency that utilizes intrinsic contextual information to enhance the spatial context reasoning for unlabeled data. Extensive experiments on various benchmarks demonstrate the superiority of our approach in both performance and efficiency. For example, with only 10% labeled data on the Left Atrial Segmentation dataset, our method surpasses BCP by 1.43% Jaccard while drastically reducing the FLOPs by 90.8% and parameters by 85.8%.
科研通智能强力驱动
Strongly Powered by AbleSci AI