Transfer learning for auto‐segmentation of 17 organs‐at‐risk in the head and neck: Bridging the gap between institutional and public datasets

Author:

Clark Brett12,Hardcastle Nicholas234,Johnston Leigh A.156,Korte James12

Affiliation:

1. Department of Biomedical Engineering University of Melbourne Melbourne Australia

2. Department of Physical Sciences Peter MacCallum Cancer Centre Melbourne Australia

3. Centre for Medical Radiation Physics University of Wollongong Wollongong Australia

4. Sir Peter MacCallum Department of Oncology University of Melbourne Melbourne Australia

5. Melbourne Brain Centre Imaging Unit University of Melbourne Melbourne Australia

6. Graeme Clark Institute University of Melbourne Melbourne Australia

Abstract

AbstractBackgroundAuto‐segmentation of organs‐at‐risk (OARs) in the head and neck (HN) on computed tomography (CT) images is a time‐consuming component of the radiation therapy pipeline that suffers from inter‐observer variability. Deep learning (DL) has shown state‐of‐the‐art results in CT auto‐segmentation, with larger and more diverse datasets showing better segmentation performance. Institutional CT auto‐segmentation datasets have been small historically (n < 50) due to the time required for manual curation of images and anatomical labels. Recently, large public CT auto‐segmentation datasets (n > 1000 aggregated) have become available through online repositories such as The Cancer Imaging Archive. Transfer learning is a technique applied when training samples are scarce, but a large dataset from a closely related domain is available.PurposeThe purpose of this study was to investigate whether a large public dataset could be used in place of an institutional dataset (n > 500), or to augment performance via transfer learning, when building HN OAR auto‐segmentation models for institutional use.MethodsAuto‐segmentation models were trained on a large public dataset (public models) and a smaller institutional dataset (institutional models). The public models were fine‐tuned on the institutional dataset using transfer learning (transfer models). We assessed both public model generalizability and transfer model performance by comparison with institutional models. Additionally, the effect of institutional dataset size on both transfer and institutional models was investigated. All DL models used a high‐resolution, two‐stage architecture based on the popular 3D U‐Net. Model performance was evaluated using five geometric measures: the dice similarity coefficient (DSC), surface DSC, 95th percentile Hausdorff distance, mean surface distance (MSD), and added path length.ResultsFor a small subset of OARs (left/right optic nerve, spinal cord, left submandibular), the public models performed significantly better (p < 0.05) than, or showed no significant difference to, the institutional models under most of the metrics examined. For the remaining OARs, the public models were inferior to the institutional models, although performance differences were small (DSC ≤ 0.03, MSD < 0.5 mm) for seven OARs (brainstem, left/right lens, left/right parotid, mandible, right submandibular). The transfer models performed significantly better than the institutional models for seven OARs (brainstem, right lens, left/right optic nerve, left/right parotid, spinal cord) with a small margin of improvement (DSC ≤ 0.02, MSD < 0.4 mm). When numbers of institutional training samples were limited, public and transfer models outperformed the institutional models for most OARs (brainstem, left/right lens, left/right optic nerve, left/right parotid, spinal cord, and left/right submandibular).ConclusionTraining auto‐segmentation models with public data alone was suitable for a small number of OARs. Using only public data incurred a small performance deficit for most other OARs, when compared with institutional data alone, but may be preferable over time‐consuming curation of a large institutional dataset. When a large institutional dataset was available, transfer learning with models pretrained on a large public dataset provided a modest performance improvement for several OARs. When numbers of institutional samples were limited, using the public dataset alone, or as a pretrained model, was beneficial for most OARs.

Publisher

Wiley

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3