Scaling inverse reinforcement learning for human-compatible AI
- π€ Speaker: Adam Gleave, UC Berkeley
- π Date & Time: Tuesday 23 October 2018, 17:00 - 18:30
- π Venue: Cambridge University Engineering Department, CBL Seminar room BE4-38. See https://www.openstreetmap.org/#map=18/52.19804/0.11969
Abstract
Inverse reinforcement learning (IRL) is a technique for inferring human preferences from demonstrations of a target behaviour. Classical approaches make strong assumptions on human rationality, are designed for only a single agent and do not scale to high-dimensional environments. In this talk, Adam Gleave will discuss recent work by himself and collaborators scaling inverse reinforcement learning to video games, demonstrations from multiple users with differing preferences, and the very hard problem of learning from users with cognitive biases. The talk will be based on Inverse reinforcement learning for video games, Multi-task Maximum Causal Entropy Inverse Reinforcement Learning and Inferring Reward Functions from Demonstrators with Unknown Biases.
Adam is a PhD student at UC Berkeley working with Stuart Russell in the Center for Human-Compatible AI. After his talk, we will have time to discuss with him how he started working in alignment, what are the most promising approaches and ways of getting involved, and more.
There will definitely be snacks and pizza.
Series This talk is part of the Engineering Safe AI series.
Included in Lists
- Cambridge talks
- Cambridge University Engineering Department, CBL Seminar room BE4-38. See https://www.openstreetmap.org/#map=18/52.19804/0.11969
- Chris Davis' list
- Engineering Safe AI
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)


Tuesday 23 October 2018, 17:00-18:30