Proximal Policy Optimization in the Fisher-Rao geometry
- đ¤ Speaker: Razvan-Andrei Lascu (RIKEN)
- đ Date & Time: Tuesday 11 November 2025, 16:30 - 17:10
- đ Venue: Seminar Room 1, Newton Institute
Abstract
PPO is one of the most widely used algorithms in reinforcement learning, offering a practical policy gradient method with strong empirical performance. However, despite its popularity, PPO lacks rigorous theoretical guarantees for policy improvement and convergence. The method employs a clipped surrogate objective, derived from linearising the value function in a flat geometric setting. In this talk, we introduce a refined surrogate objective based on the Fisher–Rao geometry, leading to a new variant, Fisher–Rao PPO (FR-PPO). Our approach provides robust theoretical guarantees, including monotonic policy improvement and sub-linear convergence rates, representing a substantial advance toward formal convergence results for the wider class of PPO algorithms. This talk is based on joint work with David Siska and Lukasz Szpruch.
Series This talk is part of the Isaac Newton Institute Seminar Series series.
Included in Lists
- All CMS events
- bld31
- dh539
- Featured lists
- INI info aggregator
- Isaac Newton Institute Seminar Series
- School of Physical Sciences
- Seminar Room 1, Newton Institute
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Razvan-Andrei Lascu (RIKEN)
Tuesday 11 November 2025, 16:30-17:10