• Event Date: 2025-10-14
  • Event Start Time: 2:00 PM
  • Event End Time: 3:20 PM
  • Event Location: 152 Frelinghuysen Rd, Psych Bldg., Busch Campus, Room 105
  • Event Type: Talks: RuCCS Colloquia
  • Event Semester: Fall 2025

 Abstract:

Young children have sophisticated representations of their visual and linguistic environment. Where do these representations come from? How much knowledge arises through generic learning mechanisms applied to sensory data, and how much requires more substantive (possibly innate) inductive biases? We examine these questions by training neural networks solely on longitudinal data collected from a single child (Sullivan et al., 2020), consisting of egocentric video and audio streams. Our principal findings are as follows: 1) Based on visual only training, neural networks can acquire high-level visual features that are broadly useful across categorization and segmentation tasks. 2) Based on language only training, networks can acquire meaningful clusters of words and sentence-level syntactic sensitivity. 3) Based on paired visual and language training, networks can acquire word-referent mappings from tens of noisy examples and align their multi-modal conceptual systems. Taken together, our results show how sophisticated visual and linguistic representations can arise through data-driven learning applied to one child’s first-person experience.

Bio:  Dr. Brenden Lake

Dr. Brenden Lake is an Associate Professor with the Department of Computer Science at Princeton University. He received his Ph.D. in Cognitive Science from Massachusetts Institute of Technology. His research uses advances in machine intelligence to better understand human intelligence and applies insights from human intelligence to develop more fruitful kinds of machine intelligence.