University of Cambridge > Talks.cam > DAMTP ML for Science Reading Group > Mechanistic interpretability (cont.) + reasoning in LLMs

Mechanistic interpretability (cont.) + reasoning in LLMs

Download to your calendar using vCal

If you have a question about this talk, please contact Rachel Zhang .

In the second journal club, we will first continue with the discussion on mechanistic interpretability by diving into a specific example of mechanistic interpretability work from the following paper: โ€˜When Models Manipulate Manifolds: The Geometry of a Counting Taskโ€™ Anthropic 2025:ย https://transformer-circuits.pub/2025/linebreaks/index.html

Then, we will discuss some highly cited ML work on reasoning capabilities in LLMs.

Description: From CoT to R1 - An overview of reasoning in LLMs

Papers: Chain-of-Thought Prompting Elicits Reasoning in Large Language Models DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning

Overview:ย  Reasoning in LLMs have shown to provide significant improvement in model performance. We provide an overview of how reasoning capabilities develop in LLMs and its wider implications

It is not necessary to read the above literature before the session!

This talk is part of the DAMTP ML for Science Reading Group series.

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

ยฉ 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity