Enhancing a Neurocognitive Shared Visuomotor Model for Object
Identification, Localization, and Grasping With Learning From Auxiliary Tasks
- URL: http://arxiv.org/abs/2009.12674v1
- Date: Sat, 26 Sep 2020 19:45:15 GMT
- Title: Enhancing a Neurocognitive Shared Visuomotor Model for Object
Identification, Localization, and Grasping With Learning From Auxiliary Tasks
- Authors: Matthias Kerzel (1), Fares Abawi (1), Manfred Eppe (1), Stefan Wermter
(1) ((1) University of Hamburg)
- Abstract summary: We present a follow-up study on our unified visuomotor neural model for the robotic tasks of identifying, localizing, and grasping a target object in a scene with multiple objects.
Our Retinanet-based model enables end-to-end training of visuomotor abilities in a biologically inspired developmental approach.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a follow-up study on our unified visuomotor neural model for the
robotic tasks of identifying, localizing, and grasping a target object in a
scene with multiple objects. Our Retinanet-based model enables end-to-end
training of visuomotor abilities in a biologically inspired developmental
approach. In our initial implementation, a neural model was able to grasp
selected objects from a planar surface. We embodied the model on the NICO
humanoid robot. In this follow-up study, we expand the task and the model to
reaching for objects in a three-dimensional space with a novel dataset based on
augmented reality and a simulation environment. We evaluate the influence of
training with auxiliary tasks, i.e., if learning of the primary visuomotor task
is supported by learning to classify and locate different objects. We show that
the proposed visuomotor model can learn to reach for objects in a
three-dimensional space. We analyze the results for biologically-plausible
biases based on object locations or properties. We show that the primary
visuomotor task can be successfully trained simultaneously with one of the two
auxiliary tasks. This is enabled by a complex neurocognitive model with shared
and task-specific components, similar to models found in biological systems.
Related papers
- Towards a "universal translator" for neural dynamics at single-cell, single-spike resolution [10.49121904052395]
We build towards a first foundation model for neural spiking data that can solve a diverse set of tasks across multiple brain areas.
Prediction tasks include single-neuron and region-level activity prediction, forward prediction, and behavior decoding.
arXiv Detail & Related papers (2024-07-19T21:05:28Z) - Planning for Learning Object Properties [117.27898922118946]
We formalize the problem of automatically training a neural network to recognize object properties as a symbolic planning problem.
We use planning techniques to produce a strategy for automating the training dataset creation and the learning process.
We provide an experimental evaluation in both a simulated and a real environment.
arXiv Detail & Related papers (2023-01-15T09:37:55Z) - Learn to Predict How Humans Manipulate Large-sized Objects from
Interactive Motions [82.90906153293585]
We propose a graph neural network, HO-GCN, to fuse motion data and dynamic descriptors for the prediction task.
We show the proposed network that consumes dynamic descriptors can achieve state-of-the-art prediction results and help the network better generalize to unseen objects.
arXiv Detail & Related papers (2022-06-25T09:55:39Z) - Lifelong 3D Object Recognition and Grasp Synthesis Using Dual Memory
Recurrent Self-Organization Networks [0.0]
Humans learn to recognize and manipulate new objects in lifelong settings without forgetting the previously gained knowledge.
In most conventional deep neural networks, this is not possible due to the problem of catastrophic forgetting.
We propose a hybrid model architecture consisting of a dual-memory recurrent neural network and an autoencoder to tackle object recognition and grasping simultaneously.
arXiv Detail & Related papers (2021-09-23T11:14:13Z) - Capturing the objects of vision with neural networks [0.0]
Human visual perception carves a scene at its physical joints, decomposing the world into objects.
Deep neural network (DNN) models of visual object recognition, by contrast, remain largely tethered to the sensory input.
We review related work in both fields and examine how these fields can help each other.
arXiv Detail & Related papers (2021-09-07T21:49:53Z) - 3D Neural Scene Representations for Visuomotor Control [78.79583457239836]
We learn models for dynamic 3D scenes purely from 2D visual observations.
A dynamics model, constructed over the learned representation space, enables visuomotor control for challenging manipulation tasks.
arXiv Detail & Related papers (2021-07-08T17:49:37Z) - Model-Based Visual Planning with Self-Supervised Functional Distances [104.83979811803466]
We present a self-supervised method for model-based visual goal reaching.
Our approach learns entirely using offline, unlabeled data.
We find that this approach substantially outperforms both model-free and model-based prior methods.
arXiv Detail & Related papers (2020-12-30T23:59:09Z) - Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in a
First-person Simulated 3D Environment [73.9469267445146]
First-person object-interaction tasks in high-fidelity, 3D, simulated environments such as the AI2Thor pose significant sample-efficiency challenges for reinforcement learning agents.
We show that one can learn object-interaction tasks from scratch without supervision by learning an attentive object-model as an auxiliary task.
arXiv Detail & Related papers (2020-10-28T19:27:26Z) - Understanding Information Processing in Human Brain by Interpreting
Machine Learning Models [1.14219428942199]
The thesis explores the role machine learning methods play in creating intuitive computational models of neural processing.
This perspective makes the case in favor of the larger role that exploratory and data-driven approach to computational neuroscience could play.
arXiv Detail & Related papers (2020-10-17T04:37:26Z) - Goal-Aware Prediction: Learning to Model What Matters [105.43098326577434]
One of the fundamental challenges in using a learned forward dynamics model is the mismatch between the objective of the learned model and that of the downstream planner or policy.
We propose to direct prediction towards task relevant information, enabling the model to be aware of the current task and encouraging it to only model relevant quantities of the state space.
We find that our method more effectively models the relevant parts of the scene conditioned on the goal, and as a result outperforms standard task-agnostic dynamics models and model-free reinforcement learning.
arXiv Detail & Related papers (2020-07-14T16:42:59Z) - Learning intuitive physics and one-shot imitation using
state-action-prediction self-organizing maps [0.0]
Humans learn by exploration and imitation, build causal models of the world, and use both to flexibly solve new tasks.
We suggest a simple but effective unsupervised model which develops such characteristics.
We demonstrate its performance on a set of several related, but different one-shot imitation tasks, which the agent flexibly solves in an active inference style.
arXiv Detail & Related papers (2020-07-03T12:29:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.