This study presents a cascaded neural network that learns both global semantic information and local spatial details for medical image segmentation of brain tumors,solving the problem that convolutional neural networks(CNN)are greatly restricted in learning global contextual information and edge details.First,the input voxels are fed into the CNN and Transformer branches separately.After the encoding phase,a two-branch fusion module is used to effectively combine the features learned in both branches to achieve the fusion of global and local information.The two-branch fu-sion module uses Hadamard products to model the fine-grained interactions between the two-branch features,while us-ing multiple attention mechanisms to fully extract the feature map channels and spatial information and suppress the in-valid noise information.The method of this paper has been evaluated on the BraTS competition website,with Dice scores of 77.92%,89.20%and 81.20%for the enhanced tumor region,full tumor region and tumor core region on the BraTS2019 validation set,respectively.Compared with other advanced 3D medical image segmentation methods,this method shows better segmentation performance,which provides a reliable basis for clinicians to make accurate brain tu-mor cell assessment and treatment plans.