A Virtual Reality Tool for Representing, Visualizing and Updating Deep
Learning Models
- URL: http://arxiv.org/abs/2305.15353v1
- Date: Wed, 24 May 2023 17:06:59 GMT
- Title: A Virtual Reality Tool for Representing, Visualizing and Updating Deep
Learning Models
- Authors: Hannes Kath, Bengt L\"uers, Thiago S. Gouv\^ea, Daniel Sonntag
- Abstract summary: We demonstrate a virtual reality tool for automating the process of assigning data inputs to different categories.
A dataset is represented as a cloud of points in virtual space.
The user explores the cloud through movement and uses hand gestures to categorise portions of the cloud.
- Score: 1.9785872350085878
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep learning is ubiquitous, but its lack of transparency limits its impact
on several potential application areas. We demonstrate a virtual reality tool
for automating the process of assigning data inputs to different categories. A
dataset is represented as a cloud of points in virtual space. The user explores
the cloud through movement and uses hand gestures to categorise portions of the
cloud. This triggers gradual movements in the cloud: points of the same
category are attracted to each other, different groups are pushed apart, while
points are globally distributed in a way that utilises the entire space. The
space, time, and forces observed in virtual reality can be mapped to
well-defined machine learning concepts, namely the latent space, the training
epochs and the backpropagation. Our tool illustrates how the inner workings of
deep neural networks can be made tangible and transparent. We expect this
approach to accelerate the autonomous development of deep learning applications
by end users in novel areas.
Related papers
- Point Cloud Matters: Rethinking the Impact of Different Observation Spaces on Robot Learning [58.69297999175239]
In robot learning, the observation space is crucial due to the distinct characteristics of different modalities.
In this study, we explore the influence of various observation spaces on robot learning, focusing on three predominant modalities: RGB, RGB-D, and point cloud.
arXiv Detail & Related papers (2024-02-04T14:18:45Z) - Generalized Label-Efficient 3D Scene Parsing via Hierarchical Feature
Aligned Pre-Training and Region-Aware Fine-tuning [55.517000360348725]
This work presents a framework for dealing with 3D scene understanding when the labeled scenes are quite limited.
To extract knowledge for novel categories from the pre-trained vision-language models, we propose a hierarchical feature-aligned pre-training and knowledge distillation strategy.
Experiments with both indoor and outdoor scenes demonstrated the effectiveness of our approach in both data-efficient learning and open-world few-shot learning.
arXiv Detail & Related papers (2023-12-01T15:47:04Z) - Self-supervised Learning for Pre-Training 3D Point Clouds: A Survey [25.51613543480276]
Self-supervised point cloud representation learning has attracted increasing attention in recent years.
This paper presents a comprehensive survey of self-supervised point cloud representation learning using DNNs.
arXiv Detail & Related papers (2023-05-08T13:20:55Z) - Sim2real Transfer Learning for Point Cloud Segmentation: An Industrial
Application Case on Autonomous Disassembly [55.41644538483948]
We present an industrial application case that uses sim2real transfer learning for point cloud data.
We provide insights on how to generate and process synthetic point cloud data.
A novel patch-based attention network is proposed additionally to tackle this problem.
arXiv Detail & Related papers (2023-01-12T14:00:37Z) - Stochastic Coherence Over Attention Trajectory For Continuous Learning
In Video Streams [64.82800502603138]
This paper proposes a novel neural-network-based approach to progressively and autonomously develop pixel-wise representations in a video stream.
The proposed method is based on a human-like attention mechanism that allows the agent to learn by observing what is moving in the attended locations.
Our experiments leverage 3D virtual environments and they show that the proposed agents can learn to distinguish objects just by observing the video stream.
arXiv Detail & Related papers (2022-04-26T09:52:31Z) - Towards Optimal Strategies for Training Self-Driving Perception Models
in Simulation [98.51313127382937]
We focus on the use of labels in the synthetic domain alone.
Our approach introduces both a way to learn neural-invariant representations and a theoretically inspired view on how to sample the data from the simulator.
We showcase our approach on the bird's-eye-view vehicle segmentation task with multi-sensor data.
arXiv Detail & Related papers (2021-11-15T18:37:43Z) - Evaluating Continual Learning Algorithms by Generating 3D Virtual
Environments [66.83839051693695]
Continual learning refers to the ability of humans and animals to incrementally learn over time in a given environment.
We propose to leverage recent advances in 3D virtual environments in order to approach the automatic generation of potentially life-long dynamic scenes with photo-realistic appearance.
A novel element of this paper is that scenes are described in a parametric way, thus allowing the user to fully control the visual complexity of the input stream the agent perceives.
arXiv Detail & Related papers (2021-09-16T10:37:21Z) - SuperCaustics: Real-time, open-source simulation of transparent objects
for deep learning applications [0.0]
SuperCaustics is a real-time, open-source simulation of transparent objects designed for deep learning applications.
We trained a deep neural network from scratch to segment transparent objects in difficult lighting scenarios.
Our neural network achieved performance comparable to the state-of-the-art on a real-world dataset using only 10% of the training data.
arXiv Detail & Related papers (2021-07-23T03:11:47Z) - Real or Virtual? Using Brain Activity Patterns to differentiate Attended
Targets during Augmented Reality Scenarios [10.739605873338592]
We use machine learning techniques to classify electroencephalographic (EEG) data collected in Augmented Reality scenarios.
A shallow convolutional neural net classified 3 second data windows from 20 participants in a person-dependent manner.
arXiv Detail & Related papers (2021-01-12T19:08:39Z) - Boosting Deep Open World Recognition by Clustering [37.5993398894786]
We show how we can boost the performance of deep open world recognition algorithms by means of a new loss formulation.
We propose a strategy to learn class-specific rejection thresholds, instead of estimating a single global threshold.
Experiments on RGB-D Object and Core50 show the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-20T12:07:39Z) - MNEW: Multi-domain Neighborhood Embedding and Weighting for Sparse Point
Clouds Segmentation [1.2380933178502298]
We propose MNEW, including multi-domain neighborhood embedding, and attention weighting based on their geometry distance, feature similarity, and neighborhood sparsity.
MNEW achieves the top performance for sparse point clouds, which is important to the application of LiDAR-based automated driving perception.
arXiv Detail & Related papers (2020-04-05T18:02:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.