首页|基于多模态特征重组和尺度交叉注意力机制的全自动脑肿瘤分割算法

基于多模态特征重组和尺度交叉注意力机制的全自动脑肿瘤分割算法

扫码查看
脑肿瘤对人体危害极大,并且在医学影像中的占比较小,边界模糊,可能以任何形状出现在脑区的任意位置,给脑肿瘤分割任务带来了极大挑战。本文综合脑肿瘤形态学和解剖学特点,提出了一种基于多模态特征重组和尺度交叉注意力机制的全自动脑肿瘤分割U形网络模型——MR-SC-UNet。它采用多任务分割框架,针对完整肿瘤(WT)、核心肿瘤(TC)、增强肿瘤(ET)等不同子区域的分割任务设计多模态特征重组模块,同时使用学习到的不同权重有效融合不同模态的信息输入,能够获得更具针对性的病灶特征。这与不同模态MRI影像凸显不同脑肿瘤病灶子区域特征的思想一致。同时,MR-SC-UNet以UNet为基准网络,在深层跳跃连接中添加尺度交叉注意力模块,可以获取多种尺度的全局信息。利用公开脑肿瘤数据集进行的相关实验结果表明,MR-SC-UNet架构对WT、TC和ET分割的Dice系数均值分别可以达到91。13%、87。46%和87。98%。这证明了所示设计的网络可以有效利用多模态肿瘤数据信息,提取和融合不同尺度的肿瘤特征,提高了脑肿瘤分割的准确性。
Full-Automatic Brain Tumor Segmentation Based on Multimodal Feature Recombination and Scale Cross Attention Mechanism
Objective Brain tumors pose a significant threat to human health,and fully automatic magnetic resonance imaging(MRI)segmentation of brain tumors and their subregions is fundamental to their computer-aided clinical diagnosis.During brain MRI segmentation using deep learning networks,tumors occupy a small volume of medical images,have blurred boundaries,and may appear in any shape and location in the brain,presenting significant challenges to brain tumor segmentation tasks.In this study,the morphological and anatomical characteristics of brain tumors are integrated,and a UNet with a multimodal recombination module and scale cross attention(MR-SC-UNet)is proposed.In the MR-SC-UNet,a multitask segmentation framework is employed,and a multimodal feature recombination module is designed for segmenting different subregions,such as the whole tumor(WT),tumor core(TC),and enhancing tumor(ET).In addition,the learned weights are used to effectively integrate information from different modalities,thereby obtaining more targeted lesion features.This approach aligns with the idea that different MRI modalities highlight different subregions of brain tumor lesions.Methods To address the feature differences required for segmenting the different subregions of brain tumors,a segmentation framework was proposed in this study,which takes the segmentation task of three lesion regions as independent sub-tasks.In this framework,complementary and shared information among various modalities is fully considered,and a multimodal feature recombination module was designed to automatically learn the attention weights of each modality.The recombined features derived by integrating these learned attention weights with the traditionally extracted features are then input into the segmentation network.In the segmentation network,the module automatically learns the attention weights of each modality and recombines these weights with traditionally extracted features.By treating the segmentation tasks of the three lesion regions as independent subtasks,accurate segmentation of the gliomas is achieved,thereby addressing the problem of differing multimodal information requirements for different regions.To address the inability of a 3DUNet to fully extract global features and fuse multiscale information,a U-shaped network based on scale cross attention(SC-U-Net)was proposed.Specifically,a scale cross attention(SC)module was designed and incorporated into the deep skip connections of a 3DUNet.By leveraging the global modeling capability of the transformer model,SC extracts the global features of the image and fully integrates multiscale information.Results Figure 7 shows the results of the ablation experiments with different configurations of the SC module.When the SC module is added to the 3rd to 5th skip connections,the network achieves the best integration of deep multiscale features,thereby enhancing the feature extraction capability of the model.The average Dice coefficient of the three regions reaches 87.98%,and the mean 95%Hausdorff distance is 5.82 mm,thereby achieving optimal performance.Table 1 lists the ablation experimental results.The best results are obtained when the proposed MR and SC modules are used together,with the Dice coefficients for the three subregions increased by 1.34,2.33,and 7.08 percentage points.Table 2 presents the comparison results of the six state-of-the-art methods,indicating superior performance in most metrics.Figures 8 and 9 show the segmentation visualization results,revealing that the improved model can more accurately identify the tumor tissue,resulting in smoother segmentation boundaries.Additionally,by integrating multiscale features,the model gains a larger receptive field,reducing the unreasonable segmentation caused by a single-scale and limited receptive field.Therefore,the segmentation results are closer to the annotated images with minimal false-positive regions.Conclusion In this study,a deep learning network framework,MR-SC-UNet,is proposed and applied to glioma segmentation tasks.The test results on the BraTS2019 dataset show that the proposed method achieves average Dice scores of 91.13%,87.46%,and 87.98%for the WT,TC,and ET regions,respectively,demonstrating its feasibility and effectiveness.In clinical applications,accurate tumor segmentation can significantly improve the capabilities of radiologists and neurosurgeons for disease assessment and provide a scientific basis for precise treatment planning and risk assessment of patients.

machine visionmultimodal feature recombinationscale cross attention mechanismbrain tumor segmentation

田恒屹、王瑜、肖洪兵

展开 >

北京工商大学计算机与人工智能学院,北京 100048

机器视觉 多模态特征重组 尺度交叉注意力机制 脑肿瘤分割

2024

中国激光
中国光学学会 中科院上海光机所

中国激光

CSTPCD北大核心
影响因子:2.204
ISSN:0258-7025
年,卷(期):2024.51(21)