This talk is part of the NLP Seminar Series.

Grammar, reasoning, learning: Three short stories on comparative & rational analysis of language model capabilities

Andrew Lampinen, Google Deepmind
Date: 11:00am - 12:00 noon PT, Dec 5 2024
Venue: Room 287, Gates Computer Science Building

Abstract

There has been substantial debate about the capabilities of language models—which aspects of language they can acquire, whether they can be said to ‘reason’, and whether they can truly ‘learn’ in context. In this talk, I will suggest that approaches from cognitive science can provide useful tools for approaching these questions. Specifically, I will focus on comparative methods (comparing capabilities across different systems) and rational analysis (analyzing behavior as a rational adaptation to an environment). I will illustrate different aspects of these ideas through three examples from our recent work: 1) comparing processing of recursive syntactic structures in language models and humans, 2) evaluating the way that both language models and humans entangle content in their responses to logical reasoning problems, and 3) understanding how in-context learning emerges from properties of training distributions. I will outline how these disparate phenomena can be understood using these cognitive methods, and the implications for evaluating and understanding language model behaviors.

Bio

Andrew Lampinen’s research bridges cognitive science and artificial intelligence, often with a focus on how the complex behaviors and representations of models, agents, or humans emerge from their learning experiences or data. His work covers topics ranging from interpretability, to explanations as a learning signal, to embodied intelligence. He is currently a Staff Research Scientist at Google DeepMind. Before that, he completed his PhD in cognitive psychology at Stanford University, and his BA in mathematics and physics at UC Berkeley. Fun fact: Andrew enjoys rock climbing and used to be a routesetter at the Stanford climbing wall.