AI Seminar: Inverse constraint learning and risk averse reinforcement learning for safe AI

Image
Pascal Poupart
Event Speaker
Pascal Poupart
Event Speaker Description
Professor
David R. Cheriton School of Computer Science
University of Waterloo
Event Type
Artificial Intelligence
Date
Event Location
BEXL 320 and Zoom
Event Description

Zoom: https://oregonstate.zoom.us/j/91611213801?pwd=Wm9JSkN1eW84RUpiS2JEd0E5T…

In many applications of reinforcement learning (RL) and control, policies need to satisfy constraints to ensure feasibility, safety or thresholds about key performance indicators. However, some constraints may be difficult to specify. For instance, in autonomous driving, it is relatively easy to specify a reward function to reach a destination, but implicit constraints followed by expert human drivers to ensure a safe, smooth and comfortable ride are much more difficult to specify. I will present some techniques to learn soft constraints from expert trajectories in autonomous driving and robotics. I will also present an alternative to variance based on Gini deviation for risk-averse reinforcement learning.

Speaker Biography

Pascal Poupart is a Professor in the David R. Cheriton School of Computer Science at the University of Waterloo (Canada). He is also a Canada CIFAR AI Chair at the Vector Institute and a member of the Waterloo AI Institute. He serves on the advisory board of the NSF AI Institute for Advances in Optimization (2022-present) at Georgia Tech. He served as Research Director and Principal Research Scientist at the Waterloo Borealis AI Research Lab at the Royal Bank of Canada (2018-2020). He also served as scientific advisor for ProNavigator (2017-2019), ElementAI (2017-2018) and DialPad (2017-2018). His research focuses on the development of algorithms for Machine Learning with application to Natural Language Processing and Material Discovery. He is most well-known for his contributions to the development of Reinforcement Learning algorithms. Notable projects that his research team are currently working on include inverse constraint learning, mean field RL, RL foundation models, Bayesian federated learning, uncertainty quantification, probabilistic deep learning, conversational agents, transcription error correction, sport analytics, adaptive satisfiability and material discovery for CO2 recycling.