NLM DIR Seminar Schedule
UPCOMING SEMINARS
-
April 8, 2025 Jaya Srivastava
Leveraging a deep learning model to assess the impact of regulatory variants on traits and diseases -
April 15, 2025 Pascal Mutz
TBD -
April 18, 2025 Valentina Boeva, Department of Computer Science, ETH Zurich
Decoding tumor heterogeneity: computational methods for scRNA-seq and spatial omics -
April 22, 2025 Stanley Liang
TBD -
April 29, 2025 MG Hirsch
TBD
RECENT SEMINARS
-
April 1, 2025 Roman Kogay
Horizontal transfer of bacterial operons into eukaryote genomes -
March 25, 2025 Yifan Yang
Adversarial Manipulation and Data Memorization in Large Language Models for Medicine -
March 11, 2025 Sofya Garushyants
Tmn – bacterial anti-phage defense system -
March 4, 2025 Sanasar Babajanyan
Evolution of antivirus defense in prokaryotes depending on the environmental virus load -
Feb. 25, 2025 Zhizheng Wang
GeneAgent: Self-verification Language Agent for Gene Set Analysis using Domain Databases
Scheduled Seminars on May 12, 2022
Contact NLMDIRSeminarScheduling@mail.nih.gov with questions about this seminar.
Abstract:
Videos that semantically correspond to a text query provide highly condensed information that can give a complete answer to the query. Videos relevant to medical instructional questions (e.g., how to use a tourniquet) are especially useful for first aid, medical emergency, and education questions. However, the number of publicly available, benchmark datasets with medical instructional videos is nonexistent. Thus we introduce two new datasets to push research toward designing and comparing algorithms that can recognize medical instructional videos and locate visual answers from them to natural language queries. We propose the datasets, MedVidCL and MedVidQA, for the tasks of Medical Video Classification (MVC) and Medical Visual Answer Localization (MVAL), two tasks that emphasize multi-modal (language and video) understanding. The MedVidCL dataset includes 6117 annotated videos for the MVC task, while the MedVidQA dataset contains 3010 annotated questions with corresponding answer segments from 899 videos for the MVAL task. We have benchmarked both tasks with both datasets via deep learning models that set competitive and comparative baselines for future research.