University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Match, Train, Improve: From Matched Data to Local Policy Improvement in Molecular Design

Match, Train, Improve: From Matched Data to Local Policy Improvement in Molecular Design

Download to your calendar using vCal

If you have a question about this talk, please contact nobody.

SCLW02 - Reinforcement Learning for Science: Discovery and Automation

Scientific discovery often operates in a regime where evaluations are expensive, data are limited, and useful interventions must remain close to known good candidates. In this talk, I present a matched-data approach to molecular property enhancement that is naturally connected to reinforcement learning. The key idea is to construct local pairs in which one molecule is both nearby and better than another, and to train a model to learn these local improving moves. Iterating this operator yields a practical strategy for lead optimization.I will argue that this method is best viewed as a critic-free, offline, local policy-improvement procedure rather than full RL. This framing clarifies its relationship to supervised fine-tuning and direct preference optimization: all three methods learn from paired data, but matched training uses locality as an additional inductive bias, allowing each pair to convey directional information about how to improve. I will also discuss extensions based on generative modeling over matched datasets, self-training, and robust out-of-distribution generalization, and conclude with opportunities for active data collection and uncertainty-aware planning in scientific design.

This talk is part of the Isaac Newton Institute Seminar Series series.

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

Š 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity