Unsupervised Online Learning for Robotic Interestingness with Visual
Memory
- URL: http://arxiv.org/abs/2111.09793v2
- Date: Fri, 19 Nov 2021 05:02:35 GMT
- Title: Unsupervised Online Learning for Robotic Interestingness with Visual
Memory
- Authors: Chen Wang, Yuheng Qiu, Wenshan Wang, Yafei Hu, Seungchan Kim,
Sebastian Scherer
- Abstract summary: We develop a method that automatically adapts online to the environment to report interesting scenes quickly.
We achieve an average of 20% higher accuracy than the state-of-the-art unsupervised methods in a subterranean tunnel environment.
- Score: 9.189959184116962
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Autonomous robots frequently need to detect "interesting" scenes to decide on
further exploration, or to decide which data to share for cooperation. These
scenarios often require fast deployment with little or no training data. Prior
work considers "interestingness" based on data from the same distribution.
Instead, we propose to develop a method that automatically adapts online to the
environment to report interesting scenes quickly. To address this problem, we
develop a novel translation-invariant visual memory and design a three-stage
architecture for long-term, short-term, and online learning, which enables the
system to learn human-like experience, environmental knowledge, and online
adaption, respectively. With this system, we achieve an average of 20% higher
accuracy than the state-of-the-art unsupervised methods in a subterranean
tunnel environment. We show comparable performance to supervised methods for
robot exploration scenarios showing the efficacy of our approach. We expect
that the presented method will play an important role in the robotic
interestingness recognition exploration tasks.
Related papers
- KOI: Accelerating Online Imitation Learning via Hybrid Key-state Guidance [51.09834120088799]
We introduce the hybrid Key-state guided Online Imitation (KOI) learning method.
We use visual-language models to extract semantic key states from expert trajectory, indicating the objectives of "what to do"
Within the intervals between semantic key states, optical flow is employed to capture motion key states to understand the mechanisms of "how to do"
arXiv Detail & Related papers (2024-08-06T02:53:55Z) - Enhancing Robot Learning through Learned Human-Attention Feature Maps [6.724036710994883]
We think that embedding auxiliary information about focus point into robot learning would enhance efficiency and robustness of the learning process.
In this paper, we propose a novel approach to model and emulate the human attention with an approximate prediction model.
We test our approach on two learning tasks - object detection and imitation learning.
arXiv Detail & Related papers (2023-08-29T14:23:44Z) - Incremental 3D Scene Completion for Safe and Efficient Exploration
Mapping and Planning [60.599223456298915]
We propose a novel way to integrate deep learning into exploration by leveraging 3D scene completion for informed, safe, and interpretable mapping and planning.
We show that our method can speed up coverage of an environment by 73% compared to the baselines with only minimal reduction in map accuracy.
Even if scene completions are not included in the final map, we show that they can be used to guide the robot to choose more informative paths, speeding up the measurement of the scene with the robot's sensors by 35%.
arXiv Detail & Related papers (2022-08-17T14:19:33Z) - Robot in a China Shop: Using Reinforcement Learning for
Location-Specific Navigation Behaviour [24.447207633447363]
We propose a new approach to navigation, where it is treated as a multi-task learning problem.
This enables the robot to learn to behave differently in visual navigation tasks for different environments.
arXiv Detail & Related papers (2021-06-02T19:31:27Z) - Actionable Models: Unsupervised Offline Reinforcement Learning of
Robotic Skills [93.12417203541948]
We propose the objective of learning a functional understanding of the environment by learning to reach any goal state in a given dataset.
We find that our method can operate on high-dimensional camera images and learn a variety of skills on real robots that generalize to previously unseen scenes and objects.
arXiv Detail & Related papers (2021-04-15T20:10:11Z) - Rapid Exploration for Open-World Navigation with Latent Goal Models [78.45339342966196]
We describe a robotic learning system for autonomous exploration and navigation in diverse, open-world environments.
At the core of our method is a learned latent variable model of distances and actions, along with a non-parametric topological memory of images.
We use an information bottleneck to regularize the learned policy, giving us (i) a compact visual representation of goals, (ii) improved generalization capabilities, and (iii) a mechanism for sampling feasible goals for exploration.
arXiv Detail & Related papers (2021-04-12T23:14:41Z) - Cognitive architecture aided by working-memory for self-supervised
multi-modal humans recognition [54.749127627191655]
The ability to recognize human partners is an important social skill to build personalized and long-term human-robot interactions.
Deep learning networks have achieved state-of-the-art results and demonstrated to be suitable tools to address such a task.
One solution is to make robots learn from their first-hand sensory data with self-supervision.
arXiv Detail & Related papers (2021-03-16T13:50:24Z) - ViNG: Learning Open-World Navigation with Visual Goals [82.84193221280216]
We propose a learning-based navigation system for reaching visually indicated goals.
We show that our system, which we call ViNG, outperforms previously-proposed methods for goal-conditioned reinforcement learning.
We demonstrate ViNG on a number of real-world applications, such as last-mile delivery and warehouse inspection.
arXiv Detail & Related papers (2020-12-17T18:22:32Z) - NavRep: Unsupervised Representations for Reinforcement Learning of Robot
Navigation in Dynamic Human Environments [28.530962677406627]
We train two end-to-end, and 18 unsupervised-learning-based architectures, and compare them, along with existing approaches, in unseen test cases.
Our results show that unsupervised learning methods are competitive with end-to-end methods.
This release also includes OpenAI-gym-compatible environments designed to emulate the training conditions described by other papers.
arXiv Detail & Related papers (2020-12-08T12:51:14Z) - Visual Memorability for Robotic Interestingness via Unsupervised Online
Learning [8.747798544090314]
We propose a novel translation-invariant visual memory for recalling and identifying interesting scenes.
This enables our system to learn human-like experience, environmental knowledge, and online adaption.
Our approach achieves much higher accuracy than the state-of-the-art algorithms on challenging robotic interestingness datasets.
arXiv Detail & Related papers (2020-05-18T16:00:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.