Read about our projects on
Prospective lab members are encouraged to review the Future directions within each project section, and then report back about the directions that pique their interest.
Behavioral scientists must often choose between two computational paradigms with complementary strengths and weaknesses: Bayesian models or neural networks. An ideal paradigm would facilitate testing different priors; Bayesian models make this easy, while neural networks do not. Similarly, an ideal paradigm would avoid over-simplifications; neural networks make this easy, while Bayesian models do not. We aim to combine the best of both with an approach we call Bayesian distillation with Behavioral Tuning (BBT). BBT allows for flexible modeling of behavior using powerful Transformers, even from minimal behavioral data, by distilling Bayesian priors into the networks and then by fine-tuning the networks further on raw behavioral examples.
Recent articles
Future directions
Simulating human reasoning in an open-ended task space has been a long-standing dream of cognitive science. Recent advances in LLMs brings this closer to reality, yet these models are not developed with human-like reasoning as an objective. In this project, we fine-tune reasoning-capable LLMs on large datasets of people reasoning “out loud” to align their reasoning processes with human patterns, akin to the Centaur LLM from Binz et al. but for reasoning. We will evaluate whether fine-tuned models better predict human reasoning and whether they can simulate human-like reasoning on novel problems.
Recent articles
Future directions
Young children have wide-ranging and sophisticated knowledge of the world. Where does this early knowledge come from? How much can be explained through generic learning mechanisms applied to sensory data, and how much requires stronger innate inductive biases? We examine these classic nature vs. nurture questions by training large-scale neural networks through the eyes and ears of a single developing child, using longitudinal baby headcam videos. We call these Child’s View AI (CVAI) models and build on the amazing SAYCam dataset. Remarkably, we find that broadly useful visual features and word-referent mappings can emerge from CVAI models trained in a self-supervised way on just slices of one child’s everyday experiences.
Recent articles
Future directions
People make rule-like, compositional generalizations in language, thought, and action — a hallmark of human intelligence often called systematic generalization. For example, once a person learns how to “photobomb” she immediately understands how to “photobomb twice” or “photobomb vigorously.” Despite recent advances, AI systems (including LLMs) still struggle with this kind of generalization. We are studying how humans learn to generalize compositionally and developing Meta-Learning for Compositionality (MLC) models to better capture and understand these human abilities. We also explore how systematic generalization relates to safety challenges in LLM, where failures of generalization can lead to unexpected or unsafe behavior.
Recent articles
Future directions
Human conceptual representations are rich in structural and statistical knowledge. Symbolic models excel at capturing compositional and causal structure, but they struggle with complex correlations. In contrast, neural network models excel at processing raw stimuli and capturing complex statistics, but they struggle with compositional and causal knowledge. The human mind seems to transcend this dichotomy: learning both structural and statistical knowledge from raw inputs. We are developing neuro-symbolic models that can capture the dual structural and statistical natures of human conceptual representation.
Recent articles
Future directions