Building and Understanding Human-scale Language Models
- đ¤ Speaker: Aaron Mueller (Boston University)
- đ Date & Time: Friday 05 December 2025, 15:00 - 16:00
- đ Venue: ONLINE ONLY. Google Meet Link: https://meet.google.com/sbr-mvua-fcc
Abstract
Abstract: Humans learn language from less than 100 million words. Todayâs state-of-the-art language models are exposed to trillions of words. What do todayâs human-scale language models learnâand what donât they? How can we close this gap in data efficiency? In this talk, I will start by presenting insights from 3 years of the BabyLM Challenge. The purpose of BabyLM is to encourage researchers to train language models using only as much data as a human would need when first learning language, and to democratize access to language modeling research. Participants have submitted a wide variety of systems; the most highly performing systems tend to come from innovations to the architecture of training objective. Then, I will present recent work on the training dynamics of both human-scale and large-scale language models. I will present a method for understanding what concepts a model is learning at specific points in training. Using subject-verb agreement as a case study, I will show that simpler word-matching features are learned early in training, while more abstract grammatical number detectorsâincluding more abstract cross-linguistic number featuresâare learned far later in training. I will conclude by discussing the future of BabyLM, and the future of interpretability as a tool for understandingâand improvingâlanguage model training.
Bio: Aaron Mueller is an Assistant Professor (Lecturer) of Computer Science (Informatics) and, by courtesy, of Data Science at Boston University. His research centers on developing language modeling methods and evaluations inspired by causal and linguistic principles, and applying these to precisely control and improve the generalization of computational models of language. He completed his Ph.D. at Johns Hopkins University. His work has been published in ML and NLP venues (such as ICML , ACL, and EMNLP ) and has won awards at TMLR and ACL . He is a recurring organizer of the BlackboxNLP and BabyLM workshops.
Series This talk is part of the NLIP Seminar Series series.
Included in Lists
- All Talks (aka the CURE list)
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Chris Davis' list
- Computer Education Research
- Computing Education Research
- Department of Computer Science and Technology talks and seminars
- Graduate-Seminars
- Guy Emerson's list
- Interested Talks
- Language Sciences for Graduate Students
- ndk22's list
- NLIP Seminar Series
- ob366-ai4er
- ONLINE ONLY. Google Meet Link: https://meet.google.com/sbr-mvua-fcc
- PMRFPS's
- rp587
- School of Technology
- Simon Baker's List
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Aaron Mueller (Boston University)
Friday 05 December 2025, 15:00-16:00