DBGAN: Dual Branch Generative Adversarial Network for Multi-Modal MRI Translation
-
Published:2024-06-13
Issue:8
Volume:20
Page:1-22
-
ISSN:1551-6857
-
Container-title:ACM Transactions on Multimedia Computing, Communications, and Applications
-
language:en
-
Short-container-title:ACM Trans. Multimedia Comput. Commun. Appl.
Author:
Lyu Jun1ORCID,
Yan Shouang2ORCID,
Hossain M. Shamim3ORCID
Affiliation:
1. Brigham and Women's Hospital, Boston, United States
2. Yantai University, Yantai, China
3. Department of Software Engineering, King Saud University College of Computer and Information Sciences, Riyadh 12372, Saudi Arabia
Abstract
Existing magnetic resonance imaging translation models rely on generative adversarial networks, primarily employing simple convolutional neural networks. Unfortunately, these networks struggle to capture global representations and contextual relationships within magnetic resonance images. While the advent of Transformers enables capturing long-range feature dependencies, they often compromise the preservation of local feature details. To address these limitations and enhance both local and global representations, we introduce
DBGAN
, a novel dual-branch generative adversarial network. In this framework, the Transformer branch comprises sparse attention blocks and dense self-attention blocks, allowing for a wider receptive field while simultaneously capturing local and global information. The convolutional neural network branch, built with integrated residual convolutional layers, enhances local modeling capabilities. Additionally, we propose a fusion module that cleverly integrates features extracted from both branches. Extensive experimentation on two public datasets and one clinical dataset validates significant performance improvements with DBGAN. On Brats2018, it achieves a 10% improvement in MAE, 3.2% in PSNR, and 4.8% in SSIM for image generation tasks compared to RegGAN. Notably, the generated MRIs receive positive feedback from radiologists, underscoring the potential of our proposed method as a valuable tool in clinical settings.
Funder
Researchers Supporting Project
King Saud University, Riyadh, Saudi Arabia
National Natural Science Foundation of China
Yantai Basic Research Key Project
Youth Innovation Science and Technology Support Program of Shandong Province
Publisher
Association for Computing Machinery (ACM)
Reference53 articles.
1. Andy Adam, Adrian K. Dixon, Jonathan H. Gillard, Cornelia Schaefer-Prokop, Ronald G. Grainger, and David J. Allison. 2014. Grainger & Allison’s Diagnostic Radiology E-Book. Elsevier Health Sciences.
2. DUNIT: Detection-Based Unsupervised Image-to-Image Translation
3. Large scale GAN training for high fidelity natural image synthesis;Brock Andrew;arXiv preprint arXiv:1809.11096,2018
4. A deep learning approach to generate contrast-enhanced computerised tomography angiography without the use of intravenous contrast agents;Chandrashekar Anirudh;arXiv preprint arXiv:2003.01223,2020
5. TransUNet: Transformers make strong encoders for medical image segmentation;Chen Jieneng;arXiv preprint arXiv:2102.04306,2021