CIS seminar series - Arsha Nagrani (Google AI Research)
Date and time
Location
Online event
Multimodal learning for videos
About this event
(The Zoom link will be provided upon completion of the Eventbrite registration)
Title: Multimodal learning for videos
Presenter: Arsha Nagrani, Google AI Research
Abstract
How can a video recognition system benefit from other modalities such as audio and text? This talk will focus on what multimodal learning is, why we need it (from both a human and machine perspective) and how to use it. We will cover some recent papers accepted to CVPR and NeurIPS, and finally brainstorm on some of the biggest challenges facing multimodal learning.
Bio
Arsha Nagrani is a Research Scientist at Google AI Research. She obtained her PhD from the University of Oxford, where her thesis received the ELLIS PhD Award. During her PhD she also spent time at Google AI in California and the Wadhwani AI non-profit organisation in Mumbai. Her work has been recognised by a Best Student Paper Award at Interspeech, an Outstanding Paper Award at ICASSP, a Google PhD Fellowship and a Townsend Scholarship, and has been covered by news outlets such as The New Scientist, MIT Tech review and Verdict. Her research is focused on cross-modal and multi-modal machine learning techniques for video understanding.
Programme
13:30 - 13:35: Opening and introduction
13:35 - 14:20: Talk by Arsha Nagrani
14:20 - 14:30: Q&A and closing remarks
Host: Changjae Oh, Centre for Intelligent Sensing, Queen Mary University of London