Empowering Knowledge Distillation via Open Set Recognition for Robust 3D
Point Cloud Classification
- URL: http://arxiv.org/abs/2010.13114v1
- Date: Sun, 25 Oct 2020 13:26:48 GMT
- Title: Empowering Knowledge Distillation via Open Set Recognition for Robust 3D
Point Cloud Classification
- Authors: Ayush Bhardwaj, Sakshee Pimpale, Saurabh Kumar, Biplab Banerjee
- Abstract summary: We propose a joint Knowledge Distillation and Open Set recognition training methodology for three-dimensional object recognition.
We demonstrate the effectiveness of the proposed method via various experiments on how it allows us to obtain a much smaller model.
- Score: 20.591508284285368
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world scenarios pose several challenges to deep learning based computer
vision techniques despite their tremendous success in research. Deeper models
provide better performance, but are challenging to deploy and knowledge
distillation allows us to train smaller models with minimal loss in
performance. The model also has to deal with open set samples from classes
outside the ones it was trained on and should be able to identify them as
unknown samples while classifying the known ones correctly. Finally, most
existing image recognition research focuses only on using two-dimensional
snapshots of the real world three-dimensional objects. In this work, we aim to
bridge these three research fields, which have been developed independently
until now, despite being deeply interrelated. We propose a joint Knowledge
Distillation and Open Set recognition training methodology for
three-dimensional object recognition. We demonstrate the effectiveness of the
proposed method via various experiments on how it allows us to obtain a much
smaller model, which takes a minimal hit in performance while being capable of
open set recognition for 3D point cloud data.
Related papers
- ZeroKey: Point-Level Reasoning and Zero-Shot 3D Keypoint Detection from Large Language Models [57.57832348655715]
We propose a novel zero-shot approach for keypoint detection on 3D shapes.
Our method utilizes the rich knowledge embedded within Multi-Modal Large Language Models.
arXiv Detail & Related papers (2024-12-09T08:31:57Z) - Generalized Label-Efficient 3D Scene Parsing via Hierarchical Feature
Aligned Pre-Training and Region-Aware Fine-tuning [55.517000360348725]
This work presents a framework for dealing with 3D scene understanding when the labeled scenes are quite limited.
To extract knowledge for novel categories from the pre-trained vision-language models, we propose a hierarchical feature-aligned pre-training and knowledge distillation strategy.
Experiments with both indoor and outdoor scenes demonstrated the effectiveness of our approach in both data-efficient learning and open-world few-shot learning.
arXiv Detail & Related papers (2023-12-01T15:47:04Z) - Leveraging Large-Scale Pretrained Vision Foundation Models for
Label-Efficient 3D Point Cloud Segmentation [67.07112533415116]
We present a novel framework that adapts various foundational models for the 3D point cloud segmentation task.
Our approach involves making initial predictions of 2D semantic masks using different large vision models.
To generate robust 3D semantic pseudo labels, we introduce a semantic label fusion strategy that effectively combines all the results via voting.
arXiv Detail & Related papers (2023-11-03T15:41:15Z) - Open-CRB: Towards Open World Active Learning for 3D Object Detection [40.80953254074535]
LiDAR-based 3D object detection has recently seen significant advancements through active learning (AL)
In real-world deployments where streaming point clouds may include unknown or novel objects, the ability of current AL methods to capture such objects remains unexplored.
This paper investigates a more practical and challenging research task: Open World Active Learning for 3D Object Detection (OWAL-3D)
arXiv Detail & Related papers (2023-10-16T13:32:53Z) - OpenPatch: a 3D patchwork for Out-Of-Distribution detection [16.262921993755892]
We present an approach for the task of semantic novelty detection on real-world point cloud samples when the reference known data are synthetic.
OpenPatch builds on a large pre-trained model and simply extracts from its intermediate features a set of patch representations that describe each known class.
We demonstrate that OpenPatch excels in both the full and few-shot known sample scenarios.
arXiv Detail & Related papers (2023-10-05T08:49:51Z) - On-Device Constrained Self-Supervised Speech Representation Learning for
Keyword Spotting via Knowledge Distillation [13.08005728839078]
We propose a knowledge distillation-based self-supervised speech representation learning architecture for on-device keyword spotting.
Our approach used a teacher-student framework to transfer knowledge from a larger, more complex model to a smaller, light-weight model.
We evaluated our model's performance on an Alexa keyword spotting detection task using a 16.6k-hour in-house dataset.
arXiv Detail & Related papers (2023-07-06T02:03:31Z) - Fine-grained 3D object recognition: an approach and experiments [0.0]
Three-dimensional (3D) object recognition technology is being used as a core technology in advanced technologies such as autonomous driving of automobiles.
There are two sets of approaches for 3D object recognition: (i) hand-crafted approaches like Global Orthographic Object Descriptor (GOOD), and (ii) deep learning-based approaches such as MobileNet and VGG.
In this paper, we first implemented an offline 3D object recognition system that takes an object view as input and generates category labels as output.
In the offline stage, instance-based learning (IBL) is used to form a new
arXiv Detail & Related papers (2023-06-28T04:48:21Z) - Explore In-Context Learning for 3D Point Cloud Understanding [71.20912026561484]
We introduce a novel framework, named Point-In-Context, designed especially for in-context learning in 3D point clouds.
We propose the Joint Sampling module, carefully designed to work in tandem with the general point sampling operator.
We conduct extensive experiments to validate the versatility and adaptability of our proposed methods in handling a wide range of tasks.
arXiv Detail & Related papers (2023-06-14T17:53:21Z) - Generalized Few-Shot 3D Object Detection of LiDAR Point Cloud for
Autonomous Driving [91.39625612027386]
We propose a novel task, called generalized few-shot 3D object detection, where we have a large amount of training data for common (base) objects, but only a few data for rare (novel) classes.
Specifically, we analyze in-depth differences between images and point clouds, and then present a practical principle for the few-shot setting in the 3D LiDAR dataset.
To solve this task, we propose an incremental fine-tuning method to extend existing 3D detection models to recognize both common and rare objects.
arXiv Detail & Related papers (2023-02-08T07:11:36Z) - Advancing 3D finger knuckle recognition via deep feature learning [51.871256510747465]
Contactless 3D finger knuckle patterns have emerged as an effective biometric identifier due to its discriminativeness, visibility from a distance, and convenience.
Recent research has developed a deep feature collaboration network which simultaneously incorporates intermediate features from deep neural networks with multiple scales.
This paper advances this approach by investigating the possibility of learning a discriminative feature vector with the least possible dimension for representing 3D finger knuckle images.
arXiv Detail & Related papers (2023-01-07T20:55:16Z) - Lifelong Ensemble Learning based on Multiple Representations for
Few-Shot Object Recognition [6.282068591820947]
We present a lifelong ensemble learning approach based on multiple representations to address the few-shot object recognition problem.
To facilitate lifelong learning, each approach is equipped with a memory unit for storing and retrieving object information instantly.
We have performed extensive sets of experiments to assess the performance of the proposed approach in offline, and open-ended scenarios.
arXiv Detail & Related papers (2022-05-04T10:29:10Z) - Anomaly Detection in Video via Self-Supervised and Multi-Task Learning [113.81927544121625]
Anomaly detection in video is a challenging computer vision problem.
In this paper, we approach anomalous event detection in video through self-supervised and multi-task learning at the object level.
arXiv Detail & Related papers (2020-11-15T10:21:28Z) - Joint Supervised and Self-Supervised Learning for 3D Real-World
Challenges [16.328866317851187]
Point cloud processing and 3D shape understanding are challenging tasks for which deep learning techniques have demonstrated great potentials.
Here we consider several possible scenarios involving synthetic and real-world point clouds where supervised learning fails due to data scarcity and large domain gaps.
We propose to enrich standard feature representations by leveraging self-supervision through a multi-task model that can solve a 3D puzzle while learning the main task of shape classification or part segmentation.
arXiv Detail & Related papers (2020-04-15T23:34:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.