One model to use them all: training a segmentation model with complementary datasets
-
Published:2024-04-27
Issue:6
Volume:19
Page:1233-1241
-
ISSN:1861-6429
-
Container-title:International Journal of Computer Assisted Radiology and Surgery
-
language:en
-
Short-container-title:Int J CARS
Author:
Jenke Alexander C.ORCID, Bodenstedt Sebastian, Kolbinger Fiona R., Distler Marius, Weitz Jürgen, Speidel Stefanie
Abstract
Abstract
Purpose
Understanding surgical scenes is crucial for computer-assisted surgery systems to provide intelligent assistance functionality. One way of achieving this is via scene segmentation using machine learning (ML). However, such ML models require large amounts of annotated training data, containing examples of all relevant object classes, which are rarely available. In this work, we propose a method to combine multiple partially annotated datasets, providing complementary annotations, into one model, enabling better scene segmentation and the use of multiple readily available datasets.
Methods
Our method aims to combine available data with complementary labels by leveraging mutual exclusive properties to maximize information. Specifically, we propose to use positive annotations of other classes as negative samples and to exclude background pixels of these binary annotations, as we cannot tell if a positive prediction by the model is correct.
Results
We evaluate our method by training a DeepLabV3 model on the publicly available Dresden Surgical Anatomy Dataset, which provides multiple subsets of binary segmented anatomical structures. Our approach successfully combines 6 classes into one model, significantly increasing the overall Dice Score by 4.4% compared to an ensemble of models trained on the classes individually. By including information on multiple classes, we were able to reduce the confusion between classes, e.g. a 24% drop for stomach and colon.
Conclusion
By leveraging multiple datasets and applying mutual exclusion constraints, we developed a method that improves surgical scene segmentation performance without the need for fully annotated datasets. Our results demonstrate the feasibility of training a model on multiple complementary datasets. This paves the way for future work further alleviating the need for one specialized large, fully segmented dataset but instead the use of already existing datasets.
Funder
Bundesministerium für Gesundheit Deutsches Krebsforschungszentrum Deutsche Forschungsgemeinschaft Horizon 2020 Framework Programme Joachim Herz Stiftung
Publisher
Springer Science and Business Media LLC
Reference17 articles.
1. Jin Y, Yu Y, Chen C, Zhao Z, Heng P-A, Stoyanov D (2022) Exploring intra- and inter-video relation for surgical semantic scene segmentation. IEEE Trans Med Imaging 41(11):2991–3002. https://doi.org/10.1109/TMI.2022.3177077 2. Mohammed A, Yildirim S, Farup I, Pedersen M, Hovde Ø (2019) StreoScenNet: surgical stereo robotic scene segmentation. In: Medical imaging 2019: image-guided procedures, robotic interventions, and modeling, vol 10951, p 109510. SPIE. https://doi.org/10.1117/12.2512518. International Society for Optics and Photonics 3. Yoon J, Hong S, Hong S, Lee J, Shin S, Park B, Sung N, Yu H, Kim S, Park S, Hyung WJ, Choi M-K (2022) Surgical scene segmentation using semantic image synthesis with a virtual surgery environment. In: Medical image computing and computer assisted intervention—MICCAI 2022. Springer, Cham, pp 551–561 4. Fuentes-Hurtado F, Kadkhodamohammadi A, Flouty E, Barbarisi S, Luengo I, Stoyanov D (2019) Easylabels: weak labels for scene segmentation in laparoscopic videos. Int J Comput Assist Radiol Surg 14(7):1247–1257 5. Allan M, Kondo S, Bodenstedt S, Leger S, Kadkhodamohammadi R, Luengo I, Fuentes F, Flouty E, Mohammed A, Pedersen M, Kori A, Alex V, Krishnamurthi G, Rauber D, Mendel R, Palm C, Bano S, Saibro G, Shih C-S, Chiang H-A, Zhuang J, Yang J, Iglovikov V, Dobrenkii A, Reddiboina M, Reddy A, Liu X, Gao C, Unberath M, Kim M, Kim C, Kim C, Kim H, Lee G, Ullah I, Luna M, Park SH, Azizian M, Stoyanov D, Maier-Hein L, Speidel S (2020) 2018 robotic scene segmentation challenge. https://doi.org/10.48550/ARXIV.2001.11190
|
|