AI Safety Gridworlds: Is my agent 'safe'?
- π€ Speaker: Jessica Yung (University of Cambridge)
- π Date & Time: Wednesday 28 February 2018, 17:00 - 18:30
- π Venue: Cambridge University Engineering Department, CBL Seminar room BE4-38. For directions see http://learning.eng.cam.ac.uk/Public/Directions
Abstract
AI Safety Gridworlds are a suite of 2D reinforcement learning environments that test for desirable safety properties of an agent, such as correct objective specification and robustness.
We will first discuss the paper’s approach to formalising safety properties in environments. Next, we will demo some of the environments and discuss whether they are reasonable tests of desirable properties. Finally, we will discuss why certain algorithms (among variations of RAINBOW and A2C ) seem to have better safety performance than others.
This week’s talk is a good opportunity to get a big-picture view of AI safety from a practical perspective. No prior knowledge of AI safety is needed.
You can try the environments for yourself by cloning this git repo: https://github.com/deepmind/ai-safety-gridworlds/tree/master/ai_safety_gridworlds
Paper: AI Safety Gridworlds (Leike et. al., 2017) https://arxiv.org/abs/1711.09883
Series This talk is part of the Engineering Safe AI series.
Included in Lists
- Cambridge talks
- Cambridge University Engineering Department, CBL Seminar room BE4-38. For directions see http://learning.eng.cam.ac.uk/Public/Directions
- Chris Davis' list
- Engineering Safe AI
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)


Wednesday 28 February 2018, 17:00-18:30