AirDet: Few-Shot Detection without Fine-tuning for Autonomous
Exploration
- URL: http://arxiv.org/abs/2112.01740v1
- Date: Fri, 3 Dec 2021 06:41:07 GMT
- Title: AirDet: Few-Shot Detection without Fine-tuning for Autonomous
Exploration
- Authors: Bowen Li, Chen Wang, Pranay Reddy, Seungchan Kim, Sebastian Scherer
- Abstract summary: We present AirDet, which is free of fine-tuning by learning class relation with support images.
AirDet achieves comparable or even better results than the exhaustively finetuned methods, reaching up to 40-60% improvements on the baseline.
We present evaluation results on real-world exploration tests from the DARPA Subterranean Challenge.
- Score: 16.032316550612336
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot object detection has rapidly progressed owing to the success of
meta-learning strategies. However, the requirement of a fine-tuning stage in
existing methods is timeconsuming and significantly hinders their usage in
real-time applications such as autonomous exploration of low-power robots. To
solve this problem, we present a brand new architecture, AirDet, which is free
of fine-tuning by learning class agnostic relation with support images.
Specifically, we propose a support-guided cross-scale (SCS) feature fusion
network to generate object proposals, a global-local relation network (GLR) for
shots aggregation, and a relation-based prototype embedding network (R-PEN) for
precise localization. Exhaustive experiments are conducted on COCO and PASCAL
VOC datasets, where surprisingly, AirDet achieves comparable or even better
results than the exhaustively finetuned methods, reaching up to 40-60%
improvements on the baseline. To our excitement, AirDet obtains favorable
performance on multi-scale objects, especially the small ones. Furthermore, we
present evaluation results on real-world exploration tests from the DARPA
Subterranean Challenge, which strongly validate the feasibility of AirDet in
robotics. The source code, pre-trained models, along with the real world data
for exploration, will be made public.
Related papers
- Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Label-Efficient 3D Object Detection For Road-Side Units [10.663986706501188]
Collaborative perception can enhance the perception of autonomous vehicles via deep information fusion with intelligent roadside units (RSU)
The data-hungry nature of these methods creates a major hurdle for their real-world deployment, particularly due to the need for annotated RSU data.
We devise a label-efficient object detection method for RSU based on unsupervised object discovery.
arXiv Detail & Related papers (2024-04-09T12:29:16Z) - AirShot: Efficient Few-Shot Detection for Autonomous Exploration [18.09990606406497]
Few-shot object detection has drawn increasing attention in the field of robotic exploration, where robots are required to find unseen objects with a few online provided examples.
Existing methods still face performance and efficiency challenges, mainly due to unreliable features and exhaustive class loops.
We propose a new paradigm AirShot, and discover that, by fully exploiting the valuable correlation map, AirShot can result in a more robust few-shot object detection system.
arXiv Detail & Related papers (2024-04-07T20:39:31Z) - Innovative Horizons in Aerial Imagery: LSKNet Meets DiffusionDet for
Advanced Object Detection [55.2480439325792]
We present an in-depth evaluation of an object detection model that integrates the LSKNet backbone with the DiffusionDet head.
The proposed model achieves a mean average precision (MAP) of approximately 45.7%, which is a significant improvement.
This advancement underscores the effectiveness of the proposed modifications and sets a new benchmark in aerial image analysis.
arXiv Detail & Related papers (2023-11-21T19:49:13Z) - Comparing Active Learning Performance Driven by Gaussian Processes or
Bayesian Neural Networks for Constrained Trajectory Exploration [0.0]
Currently, humans drive robots to meet scientific objectives, but depending on the robot's location, the exchange of information and driving commands may cause undue delays in mission fulfillment.
An autonomous robot encoded with a scientific objective and an exploration strategy incurs no communication delays and can fulfill missions more quickly.
Active learning algorithms offer this capability of intelligent exploration, but the underlying model structure varies the performance of the active learning algorithm in accurately forming an understanding of the environment.
arXiv Detail & Related papers (2023-09-28T02:45:14Z) - Efficient Real-time Smoke Filtration with 3D LiDAR for Search and Rescue
with Autonomous Heterogeneous Robotic Systems [56.838297900091426]
Smoke and dust affect the performance of any mobile robotic platform due to their reliance on onboard perception systems.
This paper proposes a novel modular computation filtration pipeline based on intensity and spatial information.
arXiv Detail & Related papers (2023-08-14T16:48:57Z) - Aerial View Goal Localization with Reinforcement Learning [6.165163123577484]
We present a framework that emulates a search-and-rescue (SAR)-like setup without requiring access to actual UAVs.
In this framework, an agent operates on top of an aerial image (proxy for a search area) and is tasked with localizing a goal that is described in terms of visual cues.
We propose AiRLoc, a reinforcement learning (RL)-based model that decouples exploration (searching for distant goals) and exploitation (localizing nearby goals)
arXiv Detail & Related papers (2022-09-08T10:27:53Z) - Rethinking Drone-Based Search and Rescue with Aerial Person Detection [79.76669658740902]
The visual inspection of aerial drone footage is an integral part of land search and rescue (SAR) operations today.
We propose a novel deep learning algorithm to automate this aerial person detection (APD) task.
We present the novel Aerial Inspection RetinaNet (AIR) algorithm as the combination of these contributions.
arXiv Detail & Related papers (2021-11-17T21:48:31Z) - Adaptive Informative Path Planning Using Deep Reinforcement Learning for
UAV-based Active Sensing [2.6519061087638014]
We propose a new approach for informative path planning based on deep reinforcement learning (RL)
Our method combines Monte Carlo tree search with an offline-learned neural network predicting informative sensing actions.
By deploying the trained network during a mission, our method enables sample-efficient online replanning on physical platforms with limited computational resources.
arXiv Detail & Related papers (2021-09-28T09:00:55Z) - Batch Exploration with Examples for Scalable Robotic Reinforcement
Learning [63.552788688544254]
Batch Exploration with Examples (BEE) explores relevant regions of the state-space guided by a modest number of human provided images of important states.
BEE is able to tackle challenging vision-based manipulation tasks both in simulation and on a real Franka robot.
arXiv Detail & Related papers (2020-10-22T17:49:25Z) - AutoOD: Automated Outlier Detection via Curiosity-guided Search and
Self-imitation Learning [72.99415402575886]
Outlier detection is an important data mining task with numerous practical applications.
We propose AutoOD, an automated outlier detection framework, which aims to search for an optimal neural network model.
Experimental results on various real-world benchmark datasets demonstrate that the deep model identified by AutoOD achieves the best performance.
arXiv Detail & Related papers (2020-06-19T18:57:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.