Enhancing Medical Imaging Segmentation with GB-SAM: A Novel Approach to Tissue Segmentation Using Granular Box Prompts
Author:
Villanueva-Miranda Ismael1, Rong Ruichen1ORCID, Quan Peiran1, Wen Zhuoyu1ORCID, Zhan Xiaowei1ORCID, Yang Donghan M.1ORCID, Chi Zhikai2ORCID, Xie Yang134, Xiao Guanghua134
Affiliation:
1. Quantitative Biomedical Research Center, Peter O’Donnell Jr. School of Public Health, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA 2. Department of Pathology, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA 3. Simmons Comprehensive Cancer Center, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA 4. Department of Bioinformatics, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA
Abstract
Recent advances in foundation models have revolutionized model development in digital pathology, reducing dependence on extensive manual annotations required by traditional methods. The ability of foundation models to generalize well with few-shot learning addresses critical barriers in adapting models to diverse medical imaging tasks. This work presents the Granular Box Prompt Segment Anything Model (GB-SAM), an improved version of the Segment Anything Model (SAM) fine-tuned using granular box prompts with limited training data. The GB-SAM aims to reduce the dependency on expert pathologist annotators by enhancing the efficiency of the automated annotation process. Granular box prompts are small box regions derived from ground truth masks, conceived to replace the conventional approach of using a single large box covering the entire H&E-stained image patch. This method allows a localized and detailed analysis of gland morphology, enhancing the segmentation accuracy of individual glands and reducing the ambiguity that larger boxes might introduce in morphologically complex regions. We compared the performance of our GB-SAM model against U-Net trained on different sizes of the CRAG dataset. We evaluated the models across histopathological datasets, including CRAG, GlaS, and Camelyon16. GB-SAM consistently outperformed U-Net, with reduced training data, showing less segmentation performance degradation. Specifically, on the CRAG dataset, GB-SAM achieved a Dice coefficient of 0.885 compared to U-Net’s 0.857 when trained on 25% of the data. Additionally, GB-SAM demonstrated segmentation stability on the CRAG testing dataset and superior generalization across unseen datasets, including challenging lymph node segmentation in Camelyon16, which achieved a Dice coefficient of 0.740 versus U-Net’s 0.491. Furthermore, compared to SAM-Path and Med-SAM, GB-SAM showed competitive performance. GB-SAM achieved a Dice score of 0.900 on the CRAG dataset, while SAM-Path achieved 0.884. On the GlaS dataset, Med-SAM reported a Dice score of 0.956, whereas GB-SAM achieved 0.885 with significantly less training data. These results highlight GB-SAM’s advanced segmentation capabilities and reduced dependency on large datasets, indicating its potential for practical deployment in digital pathology, particularly in settings with limited annotated datasets.
Funder
National Institutes of Health
Reference30 articles.
1. Artificial intelligence and digital pathology: Challenges and opportunities;Tizhoosh;J. Pathol. Inform.,2018 2. Whole slide imaging in pathology: Advantages, limitations, and emerging perspectives;Farahani;Pathol. Lab. Med. Int.,2015 3. Yang, Y., Sun, K., Gao, Y., Wang, K., and Yu, G. (2023). Preparing data for artificial intelligence in pathology with clinical-grade performance. Diagnostics, 13. 4. Huang, Z., Yang, E., Shen, J., Gratzinger, D., Eyerer, F., Liang, B., Nirschl, J., Bingham, D., Dussaq, A.M., and Kunder, C. (2024). A pathologist–AI collaboration framework for enhancing diagnostic accuracies and efficiencies. Nat. Biomed. Eng., 1–16. Online ahead of print. 5. Digital pathology and artificial intelligence applications in pathology;Go;Brain Tumor Res. Treat.,2022
|
|