Program now available
Registration open
>
16 - 20 February 2025
San Diego, California, US
Conference 13407 > Paper 13407-107
Paper 13407-107

COVSeg-VLM: Vision-language model for reliable segmenting COVID-19 infections in chest x-ray images

19 February 2025 • 5:30 PM - 7:00 PM PST | Golden State Ballroom

Abstract

Automated lung infected region delineation in X-ray images is essential for accurately diagnosing COVID-19 and assessing disease progression. In this paper, we introduced the COVID-19 Infection Segmentation Vision Language Model (COVSeg-VLM) network for infected region segmentation in 2D X-ray images. Although the advancements in vision deep-learning neural networks have significantly enhanced the auto-contouring of normal organs and anatomical structures, accurate delineation of infected region of COVID-19 remains a challenge. Our model adeptly incorporates large language models to extract text-rich features from clinical reports, captures both vision- and text-contextual information for accurate segmentation. We assessed our model using the QaTa-COV19 Dataset, showcasing its superior performance compared to conventional models, especially in difficult cases. The results underscore our network's robustness and precision, validating its efficiency in both typical and complex segmentation scenarios. These improvements herald a new era of reliable and accurate medical image analysis, with promising implications for enhancing clinical outcomes.

Presenter

Vanessa Su
Emory Univ. (United States)
This is a master student from the Department of Computer Science at Emory University.
Application tracks: AI/ML
Presenter/Author
Vanessa Su
Emory Univ. (United States)
Author
Xiaohan Yuan
Emory Univ. (United States)
Author
Emory Univ. (United States)
Author
Emory Univ. (United States)