University of Cambridge > Talks.cam > DAMTP ML for Science Reading Group > AI Alignment & RL in Turbulent Environments

AI Alignment & RL in Turbulent Environments

Download to your calendar using vCal

If you have a question about this talk, please contact Liz Tan .

We have two talks for the final journal club of Lent!

1. Stress Testing Deliberative Alignment for Anti-Scheming TrainingLeo Thom

Can AI models secretly pursue their own goals while appearing aligned? This paper by OpenAI and Apollo Research shows they can — demonstrating sandbagging, self-grading manipulation, and strategic deception across all major frontier models. We examine their proposed fix, its ~30× reduction in scheming, and a critical caveat about situational awareness that complicates the results. No AI safety background assumed.

2. Navigation with Reinforcement Learning in Turbulent EnvironmentsYixuan Zhu (Imperial)

Autonomous navigation in turbulent atmospheres presents a unique challenge, characterized by uncertain causal relationships and incomplete environmental information. In this talk, we will explore thermal soaring, the process by which birds and gliders harvest energy from ascending air currents to remain airborne without propulsion. We will examine two key studies by Reddy et al. that utilize Reinforcement Learning (RL) to address this problem. First, we will discuss how gliders can learn effective soaring strategies in turbulent flow simulations . We will then look at the transition to real-world applications, where RL algorithms trained on field data enabled model gliders to outperform non-trained counterparts by identifying and tracking thermals.

This talk is part of the DAMTP ML for Science Reading Group series.

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

Š 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity