Skip to main content

Multimodal Machine Learning for Social Robots

Time: Wed 2024-02-21 16.00

Location: Room 1440 "Henrik Eriksson", Osquars backe 2, floor 4

Video link: Zoom link

Language: English

Participating: Angelica Lim, Ph.D. Assistant Professor, School of Computing Science, Simon Fraser University

Export to calendar

Multimodal interaction between humans and robots

Abstract

In this talk, I will discuss the challenges and opportunities in creating multimodal machine learning systems that can analyze, detect and generate non-verbal communication, including gestures, gaze, auditory signals, and facial expressions. Specifically, I will provide examples in how we might allow robots to analyze human social signals (including emotions, mental states, and attitudes) across cultures, as well as to recognize and generate expressions with diversity in mind.

Bio

Dr. Angelica Lim is the Director of the Rosie Lab , and an Assistant Professor in the School of Computing Science at Simon Fraser University. Previously, she led the Emotion and Expressivity teams for the Pepper humanoid robot at SoftBank Robotics. She received her B.Sc. in Computing Science (Artificial Intelligence Specialization) from SFU and a Ph.D. and Masters in Computer Science (Intelligence Science) from Kyoto University, Japan. She has been featured on the BBC, TEDx, hosted a TV documentary on robotics, and was recently featured in Forbes 20 Leading Women in AI. 

Passionate about diversity, she is the Director of Invent the Future (AI4ALL)  program for Canada, teaches CMPT 120: Introduction to Computing Science and Programming and has developed the SFU Computer Science Teaching Toolkit , with the goal of inspiring people of all backgrounds to pursue Computing Science. She also teaches Introduction to Software Engineering  and Affective Computing

Page responsible:Roberto Bresin
Belongs to: NAVET
Last changed: Feb 19, 2024