Affiliation:
1. Center for Advanced Medical Computing and Analysis Massachusetts General Hospital and Harvard Medical School Boston Massachusetts USA
Abstract
AbstractBackgroundPatient head motion is a common source of image artifacts in computed tomography (CT) of the head, leading to degraded image quality and potentially incorrect diagnoses. The partial angle reconstruction (PAR) means dividing the CT projection into several consecutive angular segments and reconstructing each segment individually. Although motion estimation and compensation using PAR has been developed and investigated in cardiac CT scans, its potential for reducing motion artifacts in head CT scans remains unexplored.PurposeTo develop a deep learning (DL) model capable of directly estimating head motion from PAR images of head CT scans and to integrate the estimated motion into an iterative reconstruction process to compensate for the motion.MethodsHead motion is considered as a rigid transformation described by six time‐variant variables, including the three variables for translation and three variables for rotation. Each motion variable is modeled using a B‐spline defined by five control points (CP) along time. We split the full projections from 360° into 25 consecutive PARs and subsequently input them into a convolutional neural network (CNN) that outputs the estimated CPs for each motion variable. The estimated CPs are used to calculate the object motion in each projection, which are incorporated into the forward and backprojection of an iterative reconstruction algorithm to reconstruct the motion‐compensated image. The performance of our DL model is evaluated through both simulation and phantom studies.ResultsThe DL model achieved high accuracy in estimating head motion, as demonstrated in both the simulation study (mean absolute error (MAE) ranging from 0.28 to 0.45 mm or degree across different motion variables) and the phantom study (MAE ranging from 0.40 to 0.48 mm or degree). The resulting motion‐corrected image, , exhibited a significant reduction in motion artifacts when compared to the traditional filtered back‐projection reconstructions, which is evidenced both in the simulation study (image MAE drops from 178 33HU to 37 9HU, structural similarity index (SSIM) increases from 0.60 0.06 to 0.98 0.01) and the phantom study (image MAE drops from 117 17HU to 42 19HU, SSIM increases from 0.83 0.04 to 0.98 0.02).ConclusionsWe demonstrate that using PAR and our proposed deep learning model enables accurate estimation of patient head motion and effectively reduces motion artifacts in the resulting head CT images.
Funder
National Institute of Biomedical Imaging and Bioengineering