Respiratory motion management using a single rapid MRI scan for a 0.35 T MRI‐Linac system

Author:

Chen Sihao1,Eldeniz Cihat2,Fraum Tyler J.2,Ludwig Daniel R.2,Gan Weijie3,Liu Jiaming4,Kamilov Ulugbek S.34,Yang Deshan5,Gach H. Michael126,An Hongyu1247

Affiliation:

1. Department of Biomedical Engineering Washington University in St. Louis St. Louis Missouri USA

2. Mallinckrodt Institute of Radiology Washington University in St. Louis St. Louis Missouri USA

3. Department of Computer Science & Engineering Washington University in St. Louis St. Louis Missouri USA

4. Department of Electrical & Systems Engineering Washington University in St. Louis St. Louis Missouri USA

5. Department of Radiation Oncology Duke University Durham North Carolina USA

6. Department of Radiation Oncology Washington University in St. Louis St. Louis Missouri USA

7. Department of Neurology Washington University in St. Louis St. Louis Missouri USA

Abstract

AbstractBackgroundMRI has a rapidly growing role in radiation therapy (RT) for treatment planning, real‐time image guidance, and beam gating (e.g., MRI‐Linac). Free‐breathing 4D‐MRI is desirable in respiratory motion management for therapy. Moreover, high‐quality 3D‐MRIs without motion artifacts are needed to delineate lesions. Existing MRI methods require multiple scans with lengthy acquisition times or are limited by low spatial resolution, contrast, and signal‐to‐noise ratio.PurposeWe developed a novel method to obtain motion‐resolved 4D‐MRIs and motion‐integrated 3D‐MRI reconstruction using a single rapid (35‐45 s scan on a 0.35 T MRI‐Linac.MethodsGolden‐angle radial stack‐of‐stars MRI scans were acquired from a respiratory motion phantom and 12 healthy volunteers (n = 12) on a 0.35 T MRI‐Linac. A self‐navigated method was employed to detect respiratory motion using 2000 (acquisition time = 5–7 min) and the first 200 spokes (acquisition time = 35–45 s). Multi‐coil non‐uniform fast Fourier transform (MCNUFFT), compressed sensing (CS), and deep‐learning Phase2Phase (P2P) methods were employed to reconstruct motion‐resolved 4D‐MRI using 2000 spokes (MCNUFFT2000) and 200 spokes (CS200 and P2P200). Deformable motion vector fields (MVFs) were computed from the 4D‐MRIs and used to reconstruct motion‐corrected 3D‐MRIs with the MOtion Transformation Integrated forward‐Fourier (MOTIF) method. Image quality was evaluated quantitatively using the structural similarity index measure (SSIM) and the root mean square error (RMSE), and qualitatively in a blinded radiological review.ResultsEvaluation using the respiratory motion phantom experiment showed that the proposed method reversed the effects of motion blurring and restored edge sharpness. In the human study, P2P200 had smaller inaccuracy in MVFs estimation than CS200. P2P200 had significantly greater SSIMs (p < 0.0001) and smaller RMSEs (p < 0.001) than CS200 in motion‐resolved 4D‐MRI and motion‐corrected 3D‐MRI. The radiological review found that MOTIF 3D‐MRIs using MCNUFFT2000 exhibited the highest image quality (scoring > 8 out of 10), followed by P2P200 (scoring > 5 out of 10), and then motion‐uncorrected (scoring < 3 out of 10) in sharpness, contrast, and artifact‐freeness.ConclusionsWe have successfully demonstrated a method for respiratory motion management for MRI‐guided RT. The method integrated self‐navigated respiratory motion detection, deep‐learning P2P 4D‐MRI reconstruction, and a motion integrated reconstruction (MOTIF) for 3D‐MRI using a single rapid MRI scan (35‐45 s) on a 0.35 T MRI‐Linac system.

Funder

National Institutes of Health

Publisher

Wiley

Subject

General Medicine

Cited by 3 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3