Neural Sequence Models for Mathematical Reasoning
- đ¤ Speaker: Yuhuai(Tony) Wu, Stanford University & Google
- đ Date & Time: Tuesday 14 June 2022, 16:00 - 17:00
- đ Venue: Zoom
Abstract
Advanced mathematical reasoning is unique in human intelligence, and it is a fundamental building block for many intellectual pursuits and scientific developments. In this talk, I will describe generic approaches to mathematical reasoning using neural sequence models. Unlike previous approaches with highly constrained action spaces, sequence models allow the prover to freely synthesize arbitrary proof steps, including the synthesis of conjectures, lemmas, and definitions. We created a tool that can help formal mathematicians prove theorems in Lean, and our model made contributions accepted by professional mathematicians, adding new proofs to the Lean Mathlib library. The aforementioned work highlights the power of modern sequence models for automated theorem proving. However, there are still major reasoning capabilities that are missing. First, for reasoning tasks such as theorem proving or program synthesis, one must be able to work with large and continuously changing theorem databases and code repositories. I introduced a new transformer architecture that can memorize the internal representations of past inputs, allowing the utilization of newly added facts or codes without the need for retraining. Second, reasoning is often the result of extended chains of thought, while current models mostly perform fast, intuitive pattern matching. I introduced a learning algorithm that enables models to generate explicit rationales when tackling challenging reasoning problems, by bootstrapping from a pre-trained large language model. I will conclude my talk with future research directions towards the goal of building a strong mathematical reasoning model.
Bio:
Yuhuai Tony Wu is a Postdoctoral Scholar at Stanford University working with Percy Liang and Jay McClelland, and a Research Scientist at Google. His long-term research interest is to build a machine that can automate mathematics. Towards this goal, his current research focuses on building fundamental reasoning architectures as well as improving the reasoning capabilities of large language models. During his Ph.D at University of Toronto, he was awarded a Google PhD Fellowship.
Series This talk is part of the Artificial Intelligence Research Group Talks (Computer Laboratory) series.
Included in Lists
- All Talks (aka the CURE list)
- Artificial Intelligence Research Group Talks (Computer Laboratory)
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Chris Davis' list
- Department of Computer Science and Technology talks and seminars
- Guy Emerson's list
- Hanchen DaDaDash
- Interested Talks
- Martin's interesting talks
- ndk22's list
- ob366-ai4er
- PhD related
- rp587
- School of Technology
- Speech Seminars
- Trust & Technology Initiative - interesting events
- yk373's list
- yk449
- Zoom
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Yuhuai(Tony) Wu, Stanford University & Google
Tuesday 14 June 2022, 16:00-17:00