MeanAP-Guided Reinforced Active Learning for Object Detection
- URL: http://arxiv.org/abs/2310.08387v1
- Date: Thu, 12 Oct 2023 14:59:22 GMT
- Title: MeanAP-Guided Reinforced Active Learning for Object Detection
- Authors: Zhixuan Liang, Xingyu Zeng, Rui Zhao, Ping Luo
- Abstract summary: This paper introduces MeanAP-Guided Reinforced Active Learning for Object Detection (MAGRAL)
Built upon LSTM architecture, the agent efficiently explores and selects subsequent training instances.
We assess MAGRAL's efficacy across popular benchmarks, PASCAL VOC and MS COCO.
- Score: 34.19741444116433
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Active learning presents a promising avenue for training high-performance
models with minimal labeled data, achieved by judiciously selecting the most
informative instances to label and incorporating them into the task learner.
Despite notable advancements in active learning for image recognition, metrics
devised or learned to gauge the information gain of data, crucial for query
strategy design, do not consistently align with task model performance metrics,
such as Mean Average Precision (MeanAP) in object detection tasks. This paper
introduces MeanAP-Guided Reinforced Active Learning for Object Detection
(MAGRAL), a novel approach that directly utilizes the MeanAP metric of the task
model to devise a sampling strategy employing a reinforcement learning-based
sampling agent. Built upon LSTM architecture, the agent efficiently explores
and selects subsequent training instances, and optimizes the process through
policy gradient with MeanAP serving as reward. Recognizing the time-intensive
nature of MeanAP computation at each step, we propose fast look-up tables to
expedite agent training. We assess MAGRAL's efficacy across popular benchmarks,
PASCAL VOC and MS COCO, utilizing different backbone architectures. Empirical
findings substantiate MAGRAL's superiority over recent state-of-the-art
methods, showcasing substantial performance gains. MAGRAL establishes a robust
baseline for reinforced active object detection, signifying its potential in
advancing the field.
Related papers
- Parameter-Efficient Active Learning for Foundational models [7.799711162530711]
Foundational vision transformer models have shown impressive few shot performance on many vision tasks.
This research presents a novel investigation into the application of parameter efficient fine-tuning methods within an active learning (AL) framework.
arXiv Detail & Related papers (2024-06-13T16:30:32Z) - Monte Carlo Tree Search Boosts Reasoning via Iterative Preference Learning [55.96599486604344]
We introduce an approach aimed at enhancing the reasoning capabilities of Large Language Models (LLMs) through an iterative preference learning process.
We use Monte Carlo Tree Search (MCTS) to iteratively collect preference data, utilizing its look-ahead ability to break down instance-level rewards into more granular step-level signals.
The proposed algorithm employs Direct Preference Optimization (DPO) to update the LLM policy using this newly generated step-level preference data.
arXiv Detail & Related papers (2024-05-01T11:10:24Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Learning to Rank for Active Learning via Multi-Task Bilevel Optimization [29.207101107965563]
We propose a novel approach for active learning, which aims to select batches of unlabeled instances through a learned surrogate model for data acquisition.
A key challenge in this approach is developing an acquisition function that generalizes well, as the history of data, which forms part of the utility function's input, grows over time.
arXiv Detail & Related papers (2023-10-25T22:50:09Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Temporal Output Discrepancy for Loss Estimation-based Active Learning [65.93767110342502]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2022-12-20T19:29:37Z) - Active Learning with Effective Scoring Functions for Semi-Supervised
Temporal Action Localization [15.031156121516211]
This paper focuses on a rarely investigated yet practical task named semi-supervised TAL.
We propose an effective active learning method, named AL-STAL.
Experiment results show that AL-STAL outperforms the existing competitors and achieves satisfying performance compared with fully-supervised learning.
arXiv Detail & Related papers (2022-08-31T13:39:38Z) - Consistency-based Active Learning for Object Detection [10.794744492493262]
Active learning aims to improve the performance of task model by selecting the most informative samples with a limited budget.
We propose an effective Consistency-based Active Learning method for object Detection (CALD), which fully explores the consistency between original and augmented data.
arXiv Detail & Related papers (2021-03-18T17:00:34Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - Meta-Reinforcement Learning Robust to Distributional Shift via Model
Identification and Experience Relabeling [126.69933134648541]
We present a meta-reinforcement learning algorithm that is both efficient and extrapolates well when faced with out-of-distribution tasks at test time.
Our method is based on a simple insight: we recognize that dynamics models can be adapted efficiently and consistently with off-policy data.
arXiv Detail & Related papers (2020-06-12T13:34:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.