Affiliation:
1. Department of Computer Science, Iowa State University
2. Department of Medicine, University of Minnesota
Abstract
The performance of supervised deep learning image classifiers has significantly improved with large, labeled datasets and increased computing power. However, obtaining large, labeled image datasets in areas like medicine is expensive. This study seeks to improve model performance on limited labeled datasets by reducing confusion. We observed that
misclassification (or confusion) between classes is usually more prevalent between specific classes.
Thus, we developed synthesized image training techniques (SIT2), a novel
confusion-based training framework
that identifies pairs of classes with high confusion and synthesizes
not-sure
images from these pairs. The
not-sure
images are utilized in three new training strategies as follows. (1) The
not-sure
training strategy pretrains a model using
not-sure
images and the original training images. (2) The
sure-or-not
strategy pretrains with synthesized
sure
or
not-sure
images. (3) The
multi-label
strategy pretrains with synthesized images but predicts the original class(es) of the synthesized images. Finally, the pretrained model is finetuned on the original dataset. An extensive evaluation was conducted on five medical and non-medical datasets. Several improvements are statistically significant, which shows the promising future of our confusion-based training framework.
Publisher
Association for Computing Machinery (ACM)
Reference50 articles.
1. ImageNet: A large-scale hierarchical image database
2. Jeremy Irvin et al, “CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison,” presented at the AAAI, 2019. [Online]. Available: https://arxiv.org/pdf/1901.07031.pdf
3. “The 2nd Learning from Limited Labeled Data (LLD) Workshop,” Learning with Limited Labeled Data. https://lld-workshop.github.io/(accessed Aug. 18, 2022).
4. Hippocampal atrophy based Alzheimer’s disease diagnosis via machine learning methods