Affiliation:
1. College of Mechanical Engineering, Quzhou University Quzhou China
2. The First Affiliated Hospital, Gannan Medical University Ganzhou China
3. School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University Hangzhou China
4. Department of Magnetic Resonance Imaging The First Affiliated Hospital of Zhengzhou University Zhengzhou China
5. Bioengineering Department and Imperial‐X Imperial College London London UK
6. National Heart and Lung Institute, Imperial College London London UK
7. Cardiovascular Research Centre, Royal Brompton Hospital London UK
8. School of Biomedical Engineering & Imaging Sciences, King's College London London UK
Abstract
ABSTRACTBrain gliomas, common in adults, pose significant diagnostic challenges. Accurate segmentation from multimodal magnetic resonance imaging (MRI) scans is critical for effective treatment planning. Traditional manual segmentation methods, labor‐intensive and error‐prone, often lead to inconsistent diagnoses. To overcome these limitations, our study presents a sophisticated framework for the automated segmentation of brain gliomas from multimodal MRI images. This framework consists of three integral components: a 3D UNet, a classifier, and a Classifier Weight Transformer (CWT). The 3D UNet, acting as both an encoder and decoder, is instrumental in extracting comprehensive features from MRI scans. The classifier, employing a streamlined 1 × 1 convolutional architecture, performs detailed pixel‐wise classification. The CWT integrates self‐attention mechanisms through three linear layers, a multihead attention module, and layer normalization, dynamically refining the classifier's parameters based on the features extracted by the 3D UNet, thereby improving segmentation accuracy. Our model underwent a two‐stage training process for maximum efficiency: in the first stage, supervised learning was used to pre‐train the encoder and decoder, focusing on robust feature representation. In the second stage, meta‐training was applied to the classifier, with the encoder and decoder remaining unchanged, ensuring precise fine‐tuning based on the initially developed features. Extensive evaluation of datasets such as BraTS2019, BraTS2020, BraTS2021, and a specialized private dataset (ZZU) underscored the robustness and clinical potential of our framework, highlighting its superiority and competitive advantage over several state‐of‐the‐art approaches across various segmentation metrics in training and validation sets.
Funder
Horizon 2020 Framework Programme
Royal Society