loading page

Automatic lung segmentation in chest X-ray images using SAM with prompts from YOLO
  • +1
  • Ebrahim Khalili,
  • Blanca Priego-Torres,
  • Antonio Leon-Jimenez,
  • Daniel Sanchez-Morillo
Ebrahim Khalili
INiBICA, Puerta del Mar University Hospital, 11009 Cádiz, Spain, Department of Automation Engineering, Electronics and Computer Architecture and Networks, School of Engineering, University of Cadiz, Puerto Real, 11519, Cádiz, Spain
Blanca Priego-Torres
INiBICA, Puerta del Mar University Hospital, 11009 Cádiz, Spain, Department of Automation Engineering, Electronics and Computer Architecture and Networks, School of Engineering, University of Cadiz, Puerto Real, 11519, Cádiz, Spain
Antonio Leon-Jimenez
INiBICA, Puerta del Mar University Hospital, 11009 Cádiz, Spain, Pneumology Department, Puerta del Mar University Hospital, Cádiz, 11009, Cádiz, Spain
Daniel Sanchez-Morillo
INiBICA, Puerta del Mar University Hospital, 11009 Cádiz, Spain, Department of Automation Engineering, Electronics and Computer Architecture and Networks, School of Engineering, University of Cadiz, Puerto Real, 11519, Cádiz, Spain

Corresponding Author:[email protected]

Author Profile

Abstract

Despite the impressive performance of current deep learning models in the field of medical imaging, the transfer of the lung segmentation task in X-ray images to clinical practice is still a pending task. In this study, we explored the performance of a fully automatic framework for lung fields segmentation in chest X-ray images, based on the combination of the Segment Anything Model (SAM) with prompt capabilities, and the You Only Look Once (YOLO) model to provide effective prompts. Transfer learning, loss functions and several validation strategies were intensively assessed. This provides a complete benchmark that enables future research studies to fairly compare new segmentation strategies. The results achieved demonstrate significant robustness and generalization capability against the variability in sensors, populations, disease manifestations, device processing and imaging conditions. The proposed framework is computationally efficient, can address bias in training over multiple datasets, and has the potential to be applied across other domains and modalities.
06 Mar 2024Submitted to TechRxiv
11 Mar 2024Published in TechRxiv