Detail View
Context-Guided Medical Visual Question Answering
WEB OF SCIENCE
SCOPUS
- Title
- Context-Guided Medical Visual Question Answering
- Issued Date
- 2024-10-06
- Citation
- Arsalane, Wafa. (2024-10-06). Context-Guided Medical Visual Question Answering. 1st MICCAI Student Board Workshop on Empowering Medical Information Computing and Research through Early-Career Expertise, EMERGE 2024, Held in Conjunction with MICCAI 2024, 245–255. doi: 10.1007/978-3-031-79103-1_25
- Type
- Conference Paper
- ISBN
- 9783031791031
- ISSN
- 1865-0929
- Abstract
-
Given a medical image and a question in natural language, medical VQA systems are required to predict clinically relevant answers. Integrating information from visual and textual modalities requires complex fusion techniques due to the semantic gap between images and text, as well as the diversity of medical question types. To address this challenge, we propose aligning image and text features in VQA models by using text from medical reports to provide additional context during training. Specifically, we introduce a transformer-based alignment module that learns to align the image with the textual context, thereby incorporating supplementary medical features that can enhance the VQA model’s predictive capabilities. During the inference stage, VQA operates robustly without requiring any medical report. Our experiments on the Rad-Restruct dataset demonstrate a significant impact of the proposed strategy and show promising improvements, positioning our approach as competitive with state-of-the-art methods in this task. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
더보기
- Publisher
- Medical Image Computing and Computer Assisted Intervention Society
File Downloads
- There are no files associated with this item.
공유
Total Views & Downloads
???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???:
