NLM IRP Seminar Schedule
UPCOMING SEMINARS
-
May 7, 2024 OPEN
TBD -
May 9, 2024 Pascal Mutz
The Riboviria protein structurome expands virus protein annotation and highlights protein relations -
May 14, 2024 Stanley Liang
TBD -
May 16, 2024 Diego Salazar
TBD -
May 21, 2024 Ziynet Kesimoglu
TBD
RECENT SEMINARS
-
May 2, 2024 OPEN
TBD -
April 30, 2024 Wenya Rowe
The conformal central charge of the spin-1/2 XX model derived from long-chain asymptotics -
April 25, 2024 Ermin Hodzic
Condition-Aware Cell Type Deconvolution of Bulk Tissues -
April 16, 2024 Jaya Srivastava
Regulatory plasticity of the human genome -
April 11, 2024 Sergey Shmakov
Comprehensive survey of the TnpB RNA-guided nucleases
Scheduled Seminars on May 12, 2022
Contact NLM_IRP_Seminar_Scheduling@mail.nih.gov with questions about this seminar.
Abstract:
Videos that semantically correspond to a text query provide highly condensed information that can give a complete answer to the query. Videos relevant to medical instructional questions (e.g., how to use a tourniquet) are especially useful for first aid, medical emergency, and education questions. However, the number of publicly available, benchmark datasets with medical instructional videos is nonexistent. Thus we introduce two new datasets to push research toward designing and comparing algorithms that can recognize medical instructional videos and locate visual answers from them to natural language queries. We propose the datasets, MedVidCL and MedVidQA, for the tasks of Medical Video Classification (MVC) and Medical Visual Answer Localization (MVAL), two tasks that emphasize multi-modal (language and video) understanding. The MedVidCL dataset includes 6117 annotated videos for the MVC task, while the MedVidQA dataset contains 3010 annotated questions with corresponding answer segments from 899 videos for the MVAL task. We have benchmarked both tasks with both datasets via deep learning models that set competitive and comparative baselines for future research.