Parameterized Prompt for Incremental Object Detection
- URL: http://arxiv.org/abs/2510.27316v2
- Date: Tue, 04 Nov 2025 06:13:12 GMT
- Title: Parameterized Prompt for Incremental Object Detection
- Authors: Zijia An, Boyu Diao, Ruiqi Liu, Libo Huang, Chuanguang Yang, Fei Wang, Zhulin An, Yongjun Xu,
- Abstract summary: Existing prompts pool based approaches assume disjoint class sets across incremental tasks.<n>In co-occurring scenarios, unlabeled objects from previous tasks may appear in current task images, leading to confusion in prompts pool.<n>In this paper, we hold that prompt structures should exhibit adaptive consolidation properties across tasks, with constrained updates to prevent catastrophic forgetting.
- Score: 40.077943384096805
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have demonstrated that incorporating trainable prompts into pretrained models enables effective incremental learning. However, the application of prompts in incremental object detection (IOD) remains underexplored. Existing prompts pool based approaches assume disjoint class sets across incremental tasks, which are unsuitable for IOD as they overlook the inherent co-occurrence phenomenon in detection images. In co-occurring scenarios, unlabeled objects from previous tasks may appear in current task images, leading to confusion in prompts pool. In this paper, we hold that prompt structures should exhibit adaptive consolidation properties across tasks, with constrained updates to prevent catastrophic forgetting. Motivated by this, we introduce Parameterized Prompts for Incremental Object Detection (P$^2$IOD). Leveraging neural networks global evolution properties, P$^2$IOD employs networks as the parameterized prompts to adaptively consolidate knowledge across tasks. To constrain prompts structure updates, P$^2$IOD further engages a parameterized prompts fusion strategy. Extensive experiments on PASCAL VOC2007 and MS COCO datasets demonstrate that P$^2$IOD's effectiveness in IOD and achieves the state-of-the-art performance among existing baselines.
Related papers
- Beyond Prompt Degradation: Prototype-guided Dual-pool Prompting for Incremental Object Detection [18.985709082532992]
We propose a novel prompt-decoupled framework called PDP.<n>It explicitly separates task-general and task-specific prompts, preventing interference between prompts and mitigating prompt coupling.<n>It achieves state-of-the-art performance on MS-COCO and PASCAL VOC benchmarks, highlighting its potential in balancing stability and plasticity.
arXiv Detail & Related papers (2026-03-02T12:09:38Z) - BOTS: A Unified Framework for Bayesian Online Task Selection in LLM Reinforcement Finetuning [82.925106913459]
Reinforcement finetuning (RFT) is a key technique for aligning Large Language Models (LLMs) with human preferences and enhancing reasoning.<n>We introduce textbfS, a unified framework for textbfBayesian textbfOnline textbfTask textbfSelection in LLM reinforcement finetuning.
arXiv Detail & Related papers (2025-10-30T11:15:23Z) - DSS-Prompt: Dynamic-Static Synergistic Prompting for Few-Shot Class-Incremental Learning [4.957021413601961]
We introduce DSS-Prompt, a simple yet effective approach that transforms the pre-trained Vision Transformer with minimal modifications.<n>We conduct extensive experiments on four benchmarks to validate the effectiveness of our DSS-Prompt.<n>We show that it consistently achieves better performance than existing approaches on all datasets.
arXiv Detail & Related papers (2025-08-13T13:10:18Z) - Visual Consensus Prompting for Co-Salient Object Detection [26.820772908765083]
We propose an interaction-effective and parameter-efficient concise architecture for the co-salient object detection task.<n>A parameter-efficient prompt tuning paradigm and seamlessly embeds consensus into the prompts to formulate task-specific Visual Consensus Prompts (VCP)<n>OurVCP outperforms 13 cutting-edge full fine-tuning models, achieving the new state of the art (with 6.8% improvement in F_m metrics on the most challenging CoCA dataset)
arXiv Detail & Related papers (2025-04-19T10:12:39Z) - Seamless Detection: Unifying Salient Object Detection and Camouflaged Object Detection [73.85890512959861]
We propose a task-agnostic framework to unify Salient Object Detection (SOD) and Camouflaged Object Detection (COD)<n>We design a simple yet effective contextual decoder involving the interval-layer and global context, which achieves an inference speed of 67 fps.<n> Experiments on public SOD and COD datasets demonstrate the superiority of our proposed framework in both supervised and unsupervised settings.
arXiv Detail & Related papers (2024-12-22T03:25:43Z) - LW2G: Learning Whether to Grow for Prompt-based Continual Learning [55.552510632228326]
Recent Prompt-based Continual learning has achieved remarkable performance with pre-trained models.<n>These approaches expand a prompt pool by adding a new set of prompts while learning and select the correct set during inference.<n>Previous studies have revealed that learning task-wised prompt sets individually and low selection accuracy pose challenges to the performance of PCL.
arXiv Detail & Related papers (2024-09-27T15:55:13Z) - PECTP: Parameter-Efficient Cross-Task Prompts for Incremental Vision Transformer [76.39111896665585]
Incremental Learning (IL) aims to learn deep models on sequential tasks continually.
Recent vast pre-trained models (PTMs) have achieved outstanding performance by prompt technique in practical IL without the old samples.
arXiv Detail & Related papers (2024-07-04T10:37:58Z) - DiPEx: Dispersing Prompt Expansion for Class-Agnostic Object Detection [45.56930979807214]
Class-agnostic object detection can be a cornerstone or a bottleneck for many downstream vision tasks.<n>We investigate using vision-language models to enhance object detection via a self-supervised prompt learning strategy.<n>We demonstrate the effectiveness of DiPEx through extensive class-agnostic OD and OOD-OD experiments on MS-COCO and LVIS.
arXiv Detail & Related papers (2024-06-21T07:33:37Z) - TemPrompt: Multi-Task Prompt Learning for Temporal Relation Extraction in RAG-based Crowdsourcing Systems [21.312052922118585]
Temporal relation extraction (TRE) aims to grasp the evolution of events or actions, and thus shape the workflow of associated tasks.
We propose a multi-task prompt learning framework for TRE (TemPrompt), incorporating prompt tuning and contrastive learning to tackle these issues.
arXiv Detail & Related papers (2024-06-21T01:52:37Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - Dynamic Prompting: A Unified Framework for Prompt Tuning [33.175097465669374]
We present a unified dynamic prompt (DP) tuning strategy that dynamically determines different factors of prompts based on specific tasks and instances.
Experimental results underscore the significant performance improvement achieved by dynamic prompt tuning across a wide range of tasks.
We establish the universal applicability of our approach under full-data, few-shot, and multitask scenarios.
arXiv Detail & Related papers (2023-03-06T06:04:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.