Mechanistic interpretability (cont.) + reasoning in LLMs
- ๐ค Speaker: Chris Chiu (University of Cambridge)
- ๐ Date & Time: Tuesday 04 November 2025, 15:30 - 16:30
- ๐ Venue: B1.19 Potters Room, Centre for Mathematical Sciences, Cambridge CB3 0WA
Abstract
In the second journal club, we will first continue with the discussion on mechanistic interpretability by diving into a specific example of mechanistic interpretability work from the following paper: โWhen Models Manipulate Manifolds: The Geometry of a Counting Taskโ Anthropic 2025:ย https://transformer-circuits.pub/2025/linebreaks/index.html
Then, we will discuss some highly cited ML work on reasoning capabilities in LLMs.
Description: From CoT to R1 - An overview of reasoning in LLMs
Papers: Chain-of-Thought Prompting Elicits Reasoning in Large Language Models DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
Overview:ย Reasoning in LLMs have shown to provide significant improvement in model performance. We provide an overview of how reasoning capabilities develop in LLMs and its wider implications
It is not necessary to read the above literature before the session!
Series This talk is part of the DAMTP ML for Science Reading Group series.
Included in Lists
- B1.19 Potters Room, Centre for Mathematical Sciences, Cambridge CB3 0WA
- DAMTP ML for Science Reading Group
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)


Tuesday 04 November 2025, 15:30-16:30