Self-Calibrating Active Binocular Vision via Active Efficient Coding
with Deep Autoencoders
- URL: http://arxiv.org/abs/2101.11391v1
- Date: Wed, 27 Jan 2021 13:40:16 GMT
- Title: Self-Calibrating Active Binocular Vision via Active Efficient Coding
with Deep Autoencoders
- Authors: Charles Wilmot, Bertram E. Shi, Jochen Triesch
- Abstract summary: We present a model of the self-calibration of active binocular vision comprising the simultaneous learning of visual representations, vergence, and pursuit eye movements.
The model follows the principle of Active Efficient Coding (AEC), a recent extension of the classic Efficient Coding Hypothesis to active perception.
- Score: 5.653716495767271
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We present a model of the self-calibration of active binocular vision
comprising the simultaneous learning of visual representations, vergence, and
pursuit eye movements. The model follows the principle of Active Efficient
Coding (AEC), a recent extension of the classic Efficient Coding Hypothesis to
active perception. In contrast to previous AEC models, the present model uses
deep autoencoders to learn sensory representations. We also propose a new
formulation of the intrinsic motivation signal that guides the learning of
behavior. We demonstrate the performance of the model in simulations.
Related papers
- DSReLU: A Novel Dynamic Slope Function for Superior Model Training [2.2057562301812674]
The rationale behind this approach is to overcome limitations associated with traditional activation functions, such as ReLU.
Evaluated on the Mini-ImageNet, CIFAR-100, and MIT-BIH datasets, our method demonstrated improvements in classification metrics and generalization capabilities.
arXiv Detail & Related papers (2024-08-17T10:01:30Z) - A spatiotemporal style transfer algorithm for dynamic visual stimulus
generation [0.0]
We introduce the Spatiotemporal Style Transfer (STST) algorithm, a dynamic visual stimulus generation framework.
It is based on a two-stream deep neural network model that factorizes spatial and temporal features to generate dynamic visual stimuli.
We show that our algorithm enables the generation of model metamers, dynamic stimuli whose layer activations are matched to those of natural videos.
arXiv Detail & Related papers (2024-03-07T23:07:46Z) - Generative Model-based Feature Knowledge Distillation for Action
Recognition [11.31068233536815]
Our paper introduces an innovative knowledge distillation framework, with the generative model for training a lightweight student model.
The efficacy of our approach is demonstrated through comprehensive experiments on diverse popular datasets.
arXiv Detail & Related papers (2023-12-14T03:55:29Z) - Exploring Model Transferability through the Lens of Potential Energy [78.60851825944212]
Transfer learning has become crucial in computer vision tasks due to the vast availability of pre-trained deep learning models.
Existing methods for measuring the transferability of pre-trained models rely on statistical correlations between encoded static features and task labels.
We present an insightful physics-inspired approach named PED to address these challenges.
arXiv Detail & Related papers (2023-08-29T07:15:57Z) - Predictive Experience Replay for Continual Visual Control and
Forecasting [62.06183102362871]
We present a new continual learning approach for visual dynamics modeling and explore its efficacy in visual control and forecasting.
We first propose the mixture world model that learns task-specific dynamics priors with a mixture of Gaussians, and then introduce a new training strategy to overcome catastrophic forgetting.
Our model remarkably outperforms the naive combinations of existing continual learning and visual RL algorithms on DeepMind Control and Meta-World benchmarks with continual visual control tasks.
arXiv Detail & Related papers (2023-03-12T05:08:03Z) - Masked World Models for Visual Control [90.13638482124567]
We introduce a visual model-based RL framework that decouples visual representation learning and dynamics learning.
We demonstrate that our approach achieves state-of-the-art performance on a variety of visual robotic tasks.
arXiv Detail & Related papers (2022-06-28T18:42:27Z) - Self-Regulated Learning for Egocentric Video Activity Anticipation [147.9783215348252]
Self-Regulated Learning (SRL) aims to regulate the intermediate representation consecutively to produce representation that emphasizes the novel information in the frame of the current time-stamp.
SRL sharply outperforms existing state-of-the-art in most cases on two egocentric video datasets and two third-person video datasets.
arXiv Detail & Related papers (2021-11-23T03:29:18Z) - Learning Oculomotor Behaviors from Scanpath [4.611116211281628]
We develop a novel method that creates rich representations of oculomotor scanpaths to facilitate the learning of downstream tasks.
The proposed stimulus-agnostic Oculomotor Behavior Framework (OBF) model learns human oculomotor behaviors from unsupervised and semi-supervised tasks.
arXiv Detail & Related papers (2021-08-11T04:38:17Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Model-Based Inverse Reinforcement Learning from Visual Demonstrations [20.23223474119314]
We present a gradient-based inverse reinforcement learning framework that learns cost functions when given only visual human demonstrations.
The learned cost functions are then used to reproduce the demonstrated behavior via visual model predictive control.
We evaluate our framework on hardware on two basic object manipulation tasks.
arXiv Detail & Related papers (2020-10-18T17:07:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.