Sentence-level Topic Models
- ๐ค Speaker: Kris Cao (University of Cambridge) ๐ Website
- ๐ Date & Time: Friday 02 June 2017, 12:00 - 13:00
- ๐ Venue: FW26, Computer Laboratory
Abstract
We present two generative models of documents which generate whole sentences from underlying topics. This relaxes the word exchangeability assumption of traditional generative models of documents to sentence exchangeability, and can hence capture inter-word dependencies that LDA misses. Despite the additional model complexity, model training and inference is still feasible using state-of-the-art approximate inference techniques. We show that both our proposed models achieve lower perplexities than a standard LDA topic model and a strong LSTM language model on held-out documents. We also manually inspect samples from the topics learnt, and show that the topics both models learn are coherent. Finally, we show that on a shallow document classification task, LDA outperforms our models, and analyse the reasons behind this.
Series This talk is part of the NLIP Seminar Series series.
Included in Lists
- All Talks (aka the CURE list)
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Chris Davis' list
- Computer Education Research
- Computing Education Research
- Department of Computer Science and Technology talks and seminars
- FW26, Computer Laboratory
- Graduate-Seminars
- Guy Emerson's list
- Interested Talks
- Language Sciences for Graduate Students
- ndk22's list
- NLIP Seminar Series
- ob366-ai4er
- PMRFPS's
- rp587
- School of Technology
- Simon Baker's List
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)



Friday 02 June 2017, 12:00-13:00