Deep Reinforcement Learning from Human Preferences
- 👤 Speaker: Jessica Yung (University of Cambridge)
- 📅 Date & Time: Wednesday 15 November 2017, 17:00 - 18:30
- 📍 Venue: Cambridge University Engineering Department, CBL Seminar room BE4-38. For directions see http://learning.eng.cam.ac.uk/Public/Directions
Abstract
How do you teach an algorithm to do a backflip or play a game where rewards are sparse? In this seminar we will discuss how algorithms can learn from human preferences as opposed to from pre-specified goal functions.
Removing the need for humans to write goal functions is important because getting them slightly wrong could lead to dangerous behaviour. Here this is only used to learn physical behaviours, but one can imagine that it could apply to learning moral values as well.
We will be looking at the paper ‘Deep Reinforcement Learning from Human Preferences’ (Christiano et. al., 2017). We will discuss the model used and experiments in three domains: simulated robotics, Atari arcade games and novel behaviours.
Link to paper: https://arxiv.org/abs/1706.03741
Slides: https://valuealignment.ml/talks/2017-11-15-deeprl-human-prefs.pdf
Series This talk is part of the Engineering Safe AI series.
Included in Lists
- Cambridge talks
- Cambridge University Engineering Department, CBL Seminar room BE4-38. For directions see http://learning.eng.cam.ac.uk/Public/Directions
- Chris Davis' list
- Engineering Safe AI
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)


Wednesday 15 November 2017, 17:00-18:30