跳到主要导航 跳到搜索 跳到主要内容

Aggregate twice more efficiently: Dual feature aggregation transformer for medical image segmentation

  • Northwestern Polytechnical University Xian

科研成果: 期刊稿件文章同行评审

摘要

Accurate medical image segmentation provides precise descriptions of anatomical structures and pathological regions, which plays a crucial role in formulating effective treatment plans, guiding surgeries and monitoring disease progression. Recently, hybrid models combining Convolutional Neural Networks (CNNs) and Transformers have been able to compensate for the limitations of traditional CNNs in capturing long-range dependencies. However, these models often exhibit insufficient generalization ability when confronted with unknown medical data. On the other hand, purely Transformer-based models, while possessing strong global modeling capabilities, face challenges of high computational complexity. To address these problems, this paper proposes a novel U-shaped pure Transformer architecture, called Dual Feature Aggregation Transformer (DFAFormer). A Dual Feature Aggregation Transformer Block (DFATB) is designed based on Feature Aggregation Feed-Forward Network (FAFN), which enhances the model’s ability to capture richer contextual information and complex features by integrating spatial aggregation attention and channel aggregation attention mechanisms. The FAFN module introduces a gating mechanism to capture nonlinear spatial information and reduce channel redundancy, achieving efficient feature extraction while reducing the computational complexity of the model. Additionally, the Differential Transformer is innovatively incorporated, which focuses on key information and suppresses unnecessary noise through differential operations, improving the model’s robustness and generalization capabilities. Extensive comparison and ablation experiments are conducted on the Synapse, ISIC 2018 and WORD dataset, achieving average Dice scores of 83.60 %, 92.27 % and 87.78 % respectively. Experiments have shown that the proposed method outperforms state-of-the-art methods, reducing computational complexity while exhibiting strong generalization ability and promising application prospects. The code will be released via https://github.com/Sunflower-li369/DFAFormer .

源语言英语
文章编号103996
期刊Information Fusion
129
DOI
出版状态已出版 - 5月 2026

指纹

探究 'Aggregate twice more efficiently: Dual feature aggregation transformer for medical image segmentation' 的科研主题。它们共同构成独一无二的指纹。

引用此