Learning Rate Schedules, Scaling Laws, and Techniques for Pretraining LLMs
- 👤 Speaker: Alex Hägele, EPFL
- 📅 Date & Time: Tuesday 08 April 2025, 14:00 - 15:00
- 📍 Venue: Computer Lab, FW26
Abstract
Large Language Model (LLM) pretraining relies on complex strategies for large-scale optimization, with the learning rate schedule being particularly important yet often following conventional rules. In this talk, I will discuss our recent NeurIPS Spotlight that investigates a simple but effective strategy: a constant learning rate followed by strategic cooldowns. Our analysis demonstrates that this approach does not only perform reliably, but it offers practical advantages as it does not require predetermined training lengths and easily allows continual training. Importantly, these findings enable more efficient scaling law experiments, as they allow for reuse of training runs and thereby substantially reduce compute and GPU hours. In a followup work, we investigate theoretical explanations of the unique behavior of such learning rate schedules, leveraging last-iterate convergence bounds which closely match real experiments. At the end of the talk, I will conclude by introducing the Swiss AI initiative (https://www.swiss-ai.org/) which deploys the world’s first national research infrastructure with 10,000 NVIDIA Grace Hopper GPUs. This initiative leverages our research innovations, such as the above, to develop state-of-the-art open and multilingual LLMs, with the goal of advancing fully transparent scientific research on foundation models.
Bio: Alex Hägele is a PhD Student at EPFL in the Machine Learning and Optimization group (MLO) supervised by Martin Jaggi. Currently, he is part of the inaugural Anthropic Fellowship for AI Safety research, based in London. Previously, he completed his BSc+MSc in Computer Science at ETH Z ürich and was a visiting Student Researcher at Apple MLR in Paris. His research explores scaling behavior and training of language models, spanning optimization, data, and architectures.
Series This talk is part of the Cambridge ML Systems Seminar Series series.
Included in Lists
- All Talks (aka the CURE list)
- bld31
- Cambridge talks
- Computer Lab, FW26
- Department of Computer Science and Technology talks and seminars
- Interested Talks
- School of Technology
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Alex Hägele, EPFL
Tuesday 08 April 2025, 14:00-15:00