Author:
Tariq Amara,Patel Bhavik N.,Banerjee Imon
Abstract
AbstractSelf-supervised pretraining can reduce the amount of labeled training data needed by pre-learning fundamental visual characteristics of the medical imaging data. In this study, we investigate several self-supervised training strategies for chest computed tomography exams and their effects of downstream applications. we bench-mark five well-known self-supervision strategies (masked image region prediction, next slice prediction, rotation prediction, flip prediction and denoising) on 15M chest CT slices collected from four sites of Mayo Clinic enterprise. These models were evaluated for two downstream tasks on public datasets; pulmonary embolism (PE) detection (classification) and lung nodule segmentation. Image embeddings generated by these models were also evaluated for prediction of patient age, race, and gender to study inherent biases in models’ understanding of chest CT exams. Use of pretraining weights, especially masked regions prediction based weights, improved performance and reduced computational effort needed for downstream tasks compared to task-specific state-of-the-art (SOTA) models. Performance improvement for PE detection was observed for training dataset sizes as large aswith maximum gain of 5% over SOTA. Segmentation model initialized with pretraining weights learned twice as fast as randomly initialized model. While gender and age predictors built using self-supervised training weights showed no performance improvement over randomly initialized predictors, the race predictor experienced a 10% performance boost when using self-supervised training weights. We released models and weights under open-source academic license. These models can then be finetuned with limited task-specific annotated data for a variety of downstream imaging tasks thus accelerating research in biomedical imaging informatics.
Publisher
Cold Spring Harbor Laboratory
Reference34 articles.
1. I. Banerjee , K. Bhattacharjee , J. L. Burns , H. Trivedi , S. Purkayastha , L. Seyyed-Kalantari , B. N. Patel , R. Shiradkar , and J. Gichoya . “short-cuts” causing bias in radiology artificial intelligence: causes, evaluation and mitigation. Journal of the American College of Radiology, 2023.
2. On the opportunities and risks of foundation models;arXiv preprint,2021
3. B. Bozorgtabar , D. Mahapatra , and J.-P. Thiran . Amae: Adaptation of pre-trained masked autoencoder for dual-distribution anomaly detection in chest x-rays. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 195–205. Springer, 2023.
4. Detecting shortcut learning for fair medical ai using shortcut testing;Nature Communications,2023
5. T. Chen , S. Kornblith , M. Norouzi , and G. Hinton . A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597–1607. PMLR, 2020.