David Crandall
Indiana University, USA
While early work in computer vision was inspired by studies of human perception, most recent work has focused on techniques that work well in practice but probably have little biological basis. But low-cost, lightweight wearable cameras and gaze trackers can now record people's actual fields of view as they go about their everyday lives. Such first-person, "egocentric" video contains rich information about how people see and interact with the world around them, potentially helping us better understand human perception and behavior while also yielding insights that could improve computer vision. I'll describe a recent interdisciplinary project (with Chen Yu and Linda Smith) in which we used computer vision to try to characterize the properties of childrens' egocentric views as they interact with objects -- the "training data" of the child's learning system -- and then showed that injecting similar properties into the training data of computer vision algorithms could improve the algorithms' accuracies as well.
Alejandrina Cristia
PSL University, France
Rhodri Cusack
Trinity College Dublin, Ireland
Hana D’Souza
Cardiff University, UK
Emmanuel Dupoux
EHESS, France
Abdellah Fourtassi
Aix-Marseille University, France
Michael Frank
Stanford Universty, USA
Kristen Grauman
University of Texas at Austin, USA
Uri Hasson
Princeton Universty, USA
Felix Hill
DeepMind, UK
Judy Hoffman
Georgia Institute of Technology, USA
Celeste Kidd
UC Berkeley, USA
I will talk about our lab's current work-in-progress exploring interventions designed to give children a greater ability to discern truth from falsity. I will discuss some of the foundational empirical studies in progress on two types of interventions designed to facilitate children’s ability to discern fact from fiction. The first set of interventions target factors external to the child relating to the information ecosystems in which they are making judgements. The second set of interventions involve investigating internal mechanisms children may have available for helping them detect misinformed opinions. Both sets of work build off the lab's previous behavioral experiments and computational models about how children sample subsets of information from the world based on their uncertainty in order to form their beliefs and guide their subsequent sampling decisions. I will briefly provide some background on how our new work is building off of our prior papers.
Eon-Suk Ko
Chosun University, Korea
Maithilee Kunda
Vanderbilt University, USA
Casey Lew-Williams
Princeton University, USA
Jitendra Malik
UC Berkeley, USA
Atsushi Nakazawa
Kyoto University, Japan
Pierre-Yves Oudeyer
Inria, France
Marc’Aurelio Ranzato
DeepMind, UK
Jim Rehg
Georgia Institute of Technology, USA
A classical topic in computer vision and psychology is the link between knowledge of 3D object shape and the ability to categorize objects. In this talk we revisit this link in two machine learning contexts that are connected to development: few-shot learning and continual learning. We show that learning a representation of 3D shape in the form of dense local descriptors provides a surprisingly powerful cue for rapid object categorization. Our shape-based approach to low-shot learning outperforms state-of-the-art models trained on category labels. We also present the first investigation of continual learning of 3D shape and demonstrate significant differences relative to continual category learning, finding that 3D shape learning does not suffer from catastrophic forgetting.
Rebecca Saxe
MIT, USA
Olivier Sigaud
Sorbonne University, France
Linda Smith
Indiana University, USA
Daniel Swingley
University of Pennsylvania, USA
Sho Tsuji
University of Tokyo, Japan
Different views on language acquisition suggest a range of cues are used, from structure found in the linguistic signal, to information gleaned from the environmental context or through social interaction. Technological advances make it now possible to collect large quantities of ecologically valid data from young children's environment, but we still lack frameworks to extract and integrate such different kinds of cues from the input. SCALa (Socio-Computational Architecture of Language Acquisition) proposes a blueprint for computational models that makes explicit the connection between the kinds of information available to the social early language learner and the computational mechanisms required to extract language-relevant information and learn from it. SCALa further allows us to make precise recommendations for future large-scale empirical research.
Ingmar Visser
University of Amsterdam, Netherlands
Anne Warlaumont
UCLA, USA
Gert Westermann
Lancaster University, UK
Curiosity in infants and computational models
Much of what we know about infants' cognitive development comes from studies in which infants are passive recipients of information presented to them on a computer screen in an order and duration determined by the experimenter. While this body of work has provided us with many insights about infants' learning and their cognitive abilities, these methods ignore a fundamental aspect of real-life learning: outside the lab, infants are actively involved in their learning through exploring their environment and engaging with information in the order and duration they choose. In our lab we investigate infants' information seeking using behavioural, eye tracking, EEG and computational modelling methods. I will give a very brief overview of the methods and studies currently going on in my lab, and then describe a simple auto-encoder neural network model used to simulate intrinsically-motivated exploration that is based on maximizing in-the-moment learning progress. This model learns a stimulus set used in seminal studies of infant category learning as well as a non-curious model embedded in an optimally structured external environment.
Chen Yu
University of Texas at Austin, USA
Andrew Zisserman
University of Oxford, UK
Pre/Post Doctoral Flash Talks
EAGER: Asking and Answering Questions for Automatic Reward Shaping in Language-guided RL
How social interaction facilitates semantic formation and differentiation of early words
Unsupervised language learning from child-centered long-form recordings
Modelling bilingual language acquisition
Balancing generalisation and specificity in learning (and what's language got to do with it)
Cultural priors for artificial agents: Language Models as culture models
Multi-View Self-Supervised Learning for Low-Shot Object Category Recognition
Multi-View Object Discovery and Representation Learning Facilitates Fast Mapping