NLM IRP Seminar Schedule
UPCOMING SEMINARS
-
Oct. 8, 2024 Jing Wang
Enhancing Heart Failure Prediction through LLM-backed Doctor Simulation -
Oct. 15, 2024 Tanvi Patel
Generative and Diagnostic Medical Imaging through AI -
Oct. 22, 2024 Lakshminarayan Iyer
TBD -
Oct. 29, 2024 Rezarta Islamaj
TBD -
Nov. 5, 2024 Max Burroughs
TBD
RECENT SEMINARS
-
Oct. 1, 2024 Timothy Doerr
Electrostatics for biomolecular systems: ionic screening and more -
Sept. 24, 2024 Natalya Yutin
Virus discovery in the era of massive metagenomic sequencing -
Sept. 10, 2024 Diego Salazar Barreto
A phenome-wide association study to identify adverse events related with glucagon-like protein-1 agonists in Type 2 Diabetes cohort. -
July 23, 2024 Yu group
Yu Group Research Update -
July 18, 2024 Xiaofang Jiang
Jiang Lab research updates
Scheduled Seminars on May 30, 2024
Contact NLM_IRP_Seminar_Scheduling@mail.nih.gov with questions about this seminar.
Abstract:
The increase in the availability of online videos has transformed the way we access information and knowledge. A growing number of individuals now prefer instructional videos as they offer a series of step-by-step procedures to accomplish particular tasks. Instructional videos from the medical domain may provide the best possible visual answers to first aid, medical emergency, and medical education questions. This talk focuses on answering health-related questions asked by health consumers by providing visual answers from medical videos. The scarcity of large-scale datasets in the medical domain is a key challenge that hinders the development of applications that can help the public with their health-related questions. To address this issue, we first proposed a pipelined approach to create two large-scale datasets: HealthVidQA-CRF and HealthVidQA-Prompt. Leveraging the datasets, we developed monomodal and multimodal approaches that can effectively provide visual answers from medical videos to natural language questions. We conducted a comprehensive analysis of the results and outlined the findings, focusing on the impact of the created datasets on model training and the significance of visual features in enhancing the performance of the monomodal and multi-modal approaches for medical visual answer localization task.