Stochastic Gradient Descent Captures How Children Learn About Physics
- URL: http://arxiv.org/abs/2209.12344v1
- Date: Sun, 25 Sep 2022 22:56:14 GMT
- Title: Stochastic Gradient Descent Captures How Children Learn About Physics
- Authors: Luca M. Schulze Buschoff, Eric Schulz, Marcel Binz
- Abstract summary: We investigate how children's developmental trajectories compare to the learning trajectories of artificial systems.
We find that the model's learning trajectory captures the developmental trajectories of children, thereby providing support to the idea of development as optimization.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As children grow older, they develop an intuitive understanding of the
physical processes around them. They move along developmental trajectories,
which have been mapped out extensively in previous empirical research. We
investigate how children's developmental trajectories compare to the learning
trajectories of artificial systems. Specifically, we examine the idea that
cognitive development results from some form of stochastic optimization
procedure. For this purpose, we train a modern generative neural network model
using stochastic gradient descent. We then use methods from the developmental
psychology literature to probe the physical understanding of this model at
different degrees of optimization. We find that the model's learning trajectory
captures the developmental trajectories of children, thereby providing support
to the idea of development as stochastic optimization.
Related papers
- Development of Cognitive Intelligence in Pre-trained Language Models [3.1815791977708834]
Recent studies show evidence for emergent cognitive abilities in Large Pre-trained Language Models.
The developmental trajectories of PLMs consistently exhibit a window of maximal alignment to human cognitive development.
After that window, training appears to serve the engineering goal of reducing loss but not the scientific goal of increasing alignment with human cognition.
arXiv Detail & Related papers (2024-07-01T07:56:36Z) - Neural Lineage [56.34149480207817]
We introduce a novel task known as neural lineage detection, aiming at discovering lineage relationships between parent and child models.
For practical convenience, we introduce a learning-free approach, which integrates an approximation of the finetuning process into the neural network representation similarity metrics.
For the pursuit of accuracy, we introduce a learning-based lineage detector comprising encoders and a transformer detector.
arXiv Detail & Related papers (2024-06-17T01:11:53Z) - The Acquisition of Physical Knowledge in Generative Neural Networks [9.799950649945597]
We investigate how the learning trajectories of deep generative neural networks compare to children's developmental trajectories using physical understanding as a testbed.
We find that while our models are able to accurately predict a number of physical processes, their learning trajectories under both hypotheses do not follow the developmental trajectories of children.
arXiv Detail & Related papers (2023-10-30T18:58:03Z) - Learning Theory of Mind via Dynamic Traits Attribution [59.9781556714202]
We propose a new neural ToM architecture that learns to generate a latent trait vector of an actor from the past trajectories.
This trait vector then multiplicatively modulates the prediction mechanism via a fast weights' scheme in the prediction neural network.
We empirically show that the fast weights provide a good inductive bias to model the character traits of agents and hence improves mindreading ability.
arXiv Detail & Related papers (2022-04-17T11:21:18Z) - Gradient-Based Trajectory Optimization With Learned Dynamics [80.41791191022139]
We use machine learning techniques to learn a differentiable dynamics model of the system from data.
We show that a neural network can model highly nonlinear behaviors accurately for large time horizons.
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car.
arXiv Detail & Related papers (2022-04-09T22:07:34Z) - A Few-shot Learning Graph Multi-Trajectory Evolution Network for
Forecasting Multimodal Baby Connectivity Development from a Baseline
Timepoint [53.73316520733503]
We propose a Graph Multi-Trajectory Evolution Network (GmTE-Net), which adopts a teacher-student paradigm.
This is the first teacher-student architecture tailored for brain graph multi-trajectory growth prediction.
arXiv Detail & Related papers (2021-10-06T08:26:57Z) - Continual Developmental Neurosimulation Using Embodied Computational Agents [0.0]
Continual developmental neurosimulation allows us to consider the role of developmental trajectories in bridging the related phenomena of nervous system morphogenesis, developmental learning, and plasticity.
Our approach is tightly integrated with developmental embodiment, and can be implemented using a type of agent called developmental Braitenberg Vehicles (dBVs)
dBVs begin their lives as a set of undefined structures that transform into agent-based systems including a body, sensors, effectors, and nervous system.
arXiv Detail & Related papers (2021-03-07T07:22:49Z) - An Open-World Simulated Environment for Developmental Robotics [1.2955718209635252]
SEDRo allows a learning agent to have similar experiences that a human infant goes through from the fetus stage up to 12 months.
A series of simulated tests based on developmental psychology will be used to evaluate the progress of a learning model.
arXiv Detail & Related papers (2020-07-18T01:16:13Z) - A Developmental Neuro-Robotics Approach for Boosting the Recognition of
Handwritten Digits [91.3755431537592]
Recent evidence shows that a simulation of the children's embodied strategies can improve the machine intelligence too.
This article explores the application of embodied strategies to convolutional neural network models in the context of developmental neuro-robotics.
arXiv Detail & Related papers (2020-03-23T14:55:00Z) - The large learning rate phase of deep learning: the catapult mechanism [50.23041928811575]
We present a class of neural networks with solvable training dynamics.
We find good agreement between our model's predictions and training dynamics in realistic deep learning settings.
We believe our results shed light on characteristics of models trained at different learning rates.
arXiv Detail & Related papers (2020-03-04T17:52:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.