EW-DETR: Evolving World Object Detection via Incremental Low-Rank DEtection TRansformer
- URL: http://arxiv.org/abs/2602.20985v1
- Date: Tue, 24 Feb 2026 15:06:04 GMT
- Title: EW-DETR: Evolving World Object Detection via Incremental Low-Rank DEtection TRansformer
- Authors: Munish Monga, Vishal Chudasama, Pankaj Wasnik, C. V. Jawahar,
- Abstract summary: We introduce Evolving World Object Detection (EWOD), a paradigm coupling incremental learning, domain adaptation, and unknown detection under exemplar-free constraints.<n>We propose EW-DETR framework that augments DETR-based detectors with three synergistic modules.<n>This framework generalises across DETR-based detectors, enabling state-of-the-art RF-DETR to operate effectively in evolving-world settings.
- Score: 21.633220257780522
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world object detection must operate in evolving environments where new classes emerge, domains shift, and unseen objects must be identified as "unknown": all without accessing prior data. We introduce Evolving World Object Detection (EWOD), a paradigm coupling incremental learning, domain adaptation, and unknown detection under exemplar-free constraints. To tackle EWOD, we propose EW-DETR framework that augments DETR-based detectors with three synergistic modules: Incremental LoRA Adapters for exemplar-free incremental learning under evolving domains; a Query-Norm Objectness Adapter that decouples objectness-aware features from DETR decoder queries; and Entropy-Aware Unknown Mixing for calibrated unknown detection. This framework generalises across DETR-based detectors, enabling state-of-the-art RF-DETR to operate effectively in evolving-world settings. We also introduce FOGS (Forgetting, Openness, Generalisation Score) to holistically evaluate performance across these dimensions. Extensive experiments on Pascal Series and Diverse Weather benchmarks show EW-DETR outperforms other methods, improving FOGS by 57.24%.
Related papers
- From Few-Shot to Zero-Shot: Towards Generalist Graph Anomaly Detection [89.52759572485276]
ARC is a few-shot generalist GAD method that leverages in-context learning and requires only a few labeled normal samples at inference time.<n> ARC and ARC_zero effectively detect anomalies, exhibit strong generalization ability, and perform efficiently under few-shot and zero-shot settings.
arXiv Detail & Related papers (2026-02-21T10:59:00Z) - TransBridge: Boost 3D Object Detection by Scene-Level Completion with Transformer Decoder [66.22997415145467]
This paper presents a joint completion and detection framework that improves the detection feature in sparse areas.<n> Specifically, we propose TransBridge, a novel transformer-based up-sampling block that fuses the features from the detection and completion networks.<n>The results show that our framework consistently improves end-to-end 3D object detection, with the mean average precision (mAP) ranging from 0.7 to 1.5 across multiple methods.
arXiv Detail & Related papers (2025-12-12T00:08:03Z) - Source-Free Object Detection with Detection Transformer [59.33653163035064]
Source-Free Object Detection (SFOD) enables knowledge transfer from a source domain to an unsupervised target domain for object detection without access to source data.<n>Most existing SFOD approaches are either confined to conventional object detection (OD) models like Faster R-CNN or designed as general solutions without tailored adaptations for novel OD architectures, especially Detection Transformer (DETR)<n>In this paper, we introduce Feature Reweighting ANd Contrastive Learning NetworK (FRANCK), a novel SFOD framework specifically designed to perform query-centric feature enhancement for DETRs.
arXiv Detail & Related papers (2025-10-13T07:35:04Z) - RDD: Robust Feature Detector and Descriptor using Deformable Transformer [8.01082121187363]
We present Robust Deformable Detector (RDD), a novel and robust keypoint detector/descriptor.<n>We observed that deformable attention focuses on key locations, effectively reducing the search space complexity.<n>Our proposed method outperforms all state-of-the-art keypoint detection/description methods in sparse matching tasks.
arXiv Detail & Related papers (2025-05-12T19:24:45Z) - Style-Adaptive Detection Transformer for Single-Source Domain Generalized Object Detection [7.768332621617199]
Single-source domain generalization aims to develop a detector using only source domain data that generalizes well to unseen target domains.<n>Existing methods are primarily CNN-based and improve robustness through data augmentation combined with feature alignment.<n>We propose Style-Adaptive DEtection TRansformer (SA-DETR), a DETR-based detector tailored for single-source domain generalization.
arXiv Detail & Related papers (2025-04-29T07:38:37Z) - DG-DETR: Toward Domain Generalized Detection Transformer [8.762314897895175]
We introduce a Domain Generalized DEtection TRansformer (DG-DETR) to enhance the robustness of Transformer-based detectors.<n>Specifically, we propose a novel domain-agnostic query selection strategy that removes domain-induced biases from object queries.<n> Experimental results validate the effectiveness of DG-DETR.
arXiv Detail & Related papers (2025-04-28T08:33:10Z) - LGI-DETR: Local-Global Interaction for UAV Object Detection [0.0]
Most existing end-to-end object detectors are designed for natural scenes.<n>It is not ideal to apply them directly to UAV images.<n>We design an local-global information interaction DETR for UAVs.
arXiv Detail & Related papers (2025-03-24T15:32:49Z) - Oriented Tiny Object Detection: A Dataset, Benchmark, and Dynamic Unbiased Learning [51.170479006249195]
We introduce a new dataset, benchmark, and a dynamic coarse-to-fine learning scheme in this study.<n>Our proposed dataset, AI-TOD-R, features the smallest object sizes among all oriented object detection datasets.<n>We present a benchmark spanning a broad range of detection paradigms, including both fully-supervised and label-efficient approaches.
arXiv Detail & Related papers (2024-12-16T09:14:32Z) - DATR: Unsupervised Domain Adaptive Detection Transformer with Dataset-Level Adaptation and Prototypical Alignment [7.768332621617199]
We introduce a strong DETR-based detector named Domain Adaptive detection TRansformer ( DATR) for unsupervised domain adaptation of object detection.
Our proposed DATR incorporates a mean-teacher based self-training framework, utilizing pseudo-labels generated by the teacher model to further mitigate domain bias.
Experiments demonstrate superior performance and generalization capabilities of our proposed DATR in multiple domain adaptation scenarios.
arXiv Detail & Related papers (2024-05-20T03:48:45Z) - Hausdorff Distance Matching with Adaptive Query Denoising for Rotated Detection Transformer [4.137346786534721]
We introduce a Hausdorff distance-based cost for bipartite matching, which more accurately quantifies the discrepancy between predictions and ground truths.
We propose an adaptive query denoising method that employs bipartite matching to selectively eliminate noised queries that detract from model improvement.
arXiv Detail & Related papers (2023-05-12T16:42:54Z) - OW-DETR: Open-world Detection Transformer [90.56239673123804]
We introduce a novel end-to-end transformer-based framework, OW-DETR, for open-world object detection.
OW-DETR comprises three dedicated components namely, attention-driven pseudo-labeling, novelty classification and objectness scoring.
Our model outperforms the recently introduced OWOD approach, ORE, with absolute gains ranging from 1.8% to 3.3% in terms of unknown recall.
arXiv Detail & Related papers (2021-12-02T18:58:30Z) - DA-DETR: Domain Adaptive Detection Transformer with Information Fusion [53.25930448542148]
DA-DETR is a domain adaptive object detection transformer that introduces information fusion for effective transfer from a labeled source domain to an unlabeled target domain.
We introduce a novel CNN-Transformer Blender (CTBlender) that fuses the CNN features and Transformer features ingeniously for effective feature alignment and knowledge transfer across domains.
CTBlender employs the Transformer features to modulate the CNN features across multiple scales where the high-level semantic information and the low-level spatial information are fused for accurate object identification and localization.
arXiv Detail & Related papers (2021-03-31T13:55:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.