Baby Sophia: A Developmental Approach to Self-Exploration through Self-Touch and Hand Regard
- URL: http://arxiv.org/abs/2511.09727v1
- Date: Fri, 14 Nov 2025 01:06:25 GMT
- Title: Baby Sophia: A Developmental Approach to Self-Exploration through Self-Touch and Hand Regard
- Authors: Stelios Zarifis, Ioannis Chalkiadakis, Artemis Chardouveli, Vasiliki Moutzouri, Aggelos Sotirchos, Katerina Papadimitriou, Panagiotis Filntisis, Niki Efthymiou, Petros Maragos, Katerina Pastra,
- Abstract summary: We propose a Reinforcement Learning framework for autonomous self-exploration in a robotic agent, Baby Sophia.<n>The agent learns self-touch and hand regard behaviors through intrinsic rewards that mimic an infant's curiosity-driven exploration of its own body.
- Score: 16.432856040952327
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inspired by infant development, we propose a Reinforcement Learning (RL) framework for autonomous self-exploration in a robotic agent, Baby Sophia, using the BabyBench simulation environment. The agent learns self-touch and hand regard behaviors through intrinsic rewards that mimic an infant's curiosity-driven exploration of its own body. For self-touch, high-dimensional tactile inputs are transformed into compact, meaningful representations, enabling efficient learning. The agent then discovers new tactile contacts through intrinsic rewards and curriculum learning that encourage broad body coverage, balance, and generalization. For hand regard, visual features of the hands, such as skin-color and shape, are learned through motor babbling. Then, intrinsic rewards encourage the agent to perform novel hand motions, and follow its hands with its gaze. A curriculum learning setup from single-hand to dual-hand training allows the agent to reach complex visual-motor coordination. The results of this work demonstrate that purely curiosity-based signals, with no external supervision, can drive coordinated multimodal learning, imitating an infant's progression from random motor babbling to purposeful behaviors.
Related papers
- From Curiosity to Competence: How World Models Interact with the Dynamics of Exploration [0.0]
We show how evolving internal representations mediate the trade-off between curiosity and competence.<n>Our findings formalize adaptive exploration as a balance between pursuing the unknown and the controllable.
arXiv Detail & Related papers (2025-07-10T22:45:28Z) - Emergent Active Perception and Dexterity of Simulated Humanoids from Visual Reinforcement Learning [69.71072181304066]
We introduce Perceptive Dexterous Control (PDC), a framework for vision-driven whole-body control with simulated humanoids.<n>PDC operates solely on egocentric vision for task specification, enabling object search, target placement, and skill selection through visual cues.<n>We show that training from scratch with reinforcement learning can produce emergent behaviors such as active search.
arXiv Detail & Related papers (2025-05-18T07:33:31Z) - A computational model of infant sensorimotor exploration in the mobile paradigm [13.666777211441286]
We present a computational model of the mechanisms that may determine infants' behavior in the "mobile paradigm"<n>In this paradigm, a mobile is connected to one of the infant's limbs, prompting the infant to preferentially move that "connected" limb.<n>Our model incorporates a neural network, action-outcome prediction, exploration, motor noise, preferred activity level, and biologically-inspired motor control.
arXiv Detail & Related papers (2025-04-24T21:02:06Z) - Toddlers' Active Gaze Behavior Supports Self-Supervised Object Learning [4.612042044544857]
We show that toddlers' gaze strategy supports the learning of invariant object representations.<n>Our work reveals how toddlers' gaze behavior may support their development of view-invariant object recognition.
arXiv Detail & Related papers (2024-11-04T10:44:46Z) - MIMo: A Multi-Modal Infant Model for Studying Cognitive Development [3.5009119465343033]
We present MIMo, an open-source infant model for studying early cognitive development through computer simulations.
MIMo perceives its surroundings via binocular vision, a vestibular system, proprioception, and touch perception through a full-body virtual skin.
arXiv Detail & Related papers (2023-12-07T14:21:31Z) - Active Vision Reinforcement Learning under Limited Visual Observability [46.99501921691587]
We investigate Active Vision Reinforcement Learning (ActiveVision-RL) where an embodied agent simultaneously learns action policy for the task while also controlling its visual observations in partially observable environments.
We propose SUGARL, Sensorimotor Understanding Guided Active Reinforcement Learning, a framework that models motor and sensory policies separately, but jointly learns them using with an intrinsic sensorimotor reward.
arXiv Detail & Related papers (2023-06-01T17:59:05Z) - Developmental Curiosity and Social Interaction in Virtual Agents [2.8894038270224858]
We create a virtual infant agent and place it in a developmentally-inspired 3D environment with no external rewards.
We test intrinsic reward functions that are similar to motivations that have been proposed to drive exploration in humans.
We find that learning a world model in the presence of an attentive caregiver helps the infant agent learn how to predict scenarios.
arXiv Detail & Related papers (2023-05-22T18:17:07Z) - Incremental procedural and sensorimotor learning in cognitive humanoid
robots [52.77024349608834]
This work presents a cognitive agent that can learn procedures incrementally.
We show the cognitive functions required in each substage and how adding new functions helps address tasks previously unsolved by the agent.
Results show that this approach is capable of solving complex tasks incrementally.
arXiv Detail & Related papers (2023-04-30T22:51:31Z) - Dexterous Manipulation from Images: Autonomous Real-World RL via Substep
Guidance [71.36749876465618]
We describe a system for vision-based dexterous manipulation that provides a "programming-free" approach for users to define new tasks.
Our system includes a framework for users to define a final task and intermediate sub-tasks with image examples.
experimental results with a four-finger robotic hand learning multi-stage object manipulation tasks directly in the real world.
arXiv Detail & Related papers (2022-12-19T22:50:40Z) - Development of collective behavior in newborn artificial agents [0.0]
We use deep reinforcement learning and curiosity-driven learning to build newborn artificial agents that develop collective behavior.
Our agents learn collective behavior without external rewards, using only intrinsic motivation (curiosity) to drive learning.
This work bridges the divide between high-dimensional sensory inputs and collective action, resulting in a pixels-to-actions model of collective animal behavior.
arXiv Detail & Related papers (2021-11-06T03:46:31Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - AGENT: A Benchmark for Core Psychological Reasoning [60.35621718321559]
Intuitive psychology is the ability to reason about hidden mental variables that drive observable actions.
Despite recent interest in machine agents that reason about other agents, it is not clear if such agents learn or hold the core psychology principles that drive human reasoning.
We present a benchmark consisting of procedurally generated 3D animations, AGENT, structured around four scenarios.
arXiv Detail & Related papers (2021-02-24T14:58:23Z) - Mutual Information-based State-Control for Intrinsically Motivated
Reinforcement Learning [102.05692309417047]
In reinforcement learning, an agent learns to reach a set of goals by means of an external reward signal.
In the natural world, intelligent organisms learn from internal drives, bypassing the need for external signals.
We propose to formulate an intrinsic objective as the mutual information between the goal states and the controllable states.
arXiv Detail & Related papers (2020-02-05T19:21:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.