eCDT: Event Clustering for Simultaneous Feature Detection and Tracking-
- URL: http://arxiv.org/abs/2207.09108v2
- Date: Wed, 20 Jul 2022 05:40:43 GMT
- Title: eCDT: Event Clustering for Simultaneous Feature Detection and Tracking-
- Authors: Sumin Hu, Yeeun Kim, Hyungtae Lim, Alex Junho Lee, Hyun Myung
- Abstract summary: Event Clustering-based Detection Tracking (eCDT)
Our method 30% better feature tracking ages compared with the state-of-the-art approach.
- Score: 4.094848360328624
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrary to other standard cameras, event cameras interpret the world in an
entirely different manner; as a collection of asynchronous events. Despite
event camera's unique data output, many event feature detection and tracking
algorithms have shown significant progress by making detours to frame-based
data representations. This paper questions the need to do so and proposes a
novel event data-friendly method that achieve simultaneous feature detection
and tracking, called event Clustering-based Detection and Tracking (eCDT). Our
method employs a novel clustering method, named as k-NN Classifier-based
Spatial Clustering and Applications with Noise (KCSCAN), to cluster adjacent
polarity events to retrieve event trajectories.With the aid of a Head and Tail
Descriptor Matching process, event clusters that reappear in a different
polarity are continually tracked, elongating the feature tracks. Thanks to our
clustering approach in spatio-temporal space, our method automatically solves
feature detection and feature tracking simultaneously. Also, eCDT can extract
feature tracks at any frequency with an adjustable time window, which does not
corrupt the high temporal resolution of the original event data. Our method
achieves 30% better feature tracking ages compared with the state-of-the-art
approach while also having a low error approximately equal to it.
Related papers
- TSdetector: Temporal-Spatial Self-correction Collaborative Learning for Colonoscopy Video Detection [19.00902297385955]
We propose a novel Temporal-Spatial self-correction detector (TSdetector), which integrates temporal-level consistency learning and spatial-level reliability learning to detect objects continuously.
The experimental results on three publicly available polyp video dataset show that TSdetector achieves the highest polyp detection rate and outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-09-30T06:19:29Z) - BlinkTrack: Feature Tracking over 100 FPS via Events and Images [50.98675227695814]
We propose a novel framework, BlinkTrack, which integrates event data with RGB images for high-frequency feature tracking.
Our method extends the traditional Kalman filter into a learning-based framework, utilizing differentiable Kalman filters in both event and image branches.
Experimental results indicate that BlinkTrack significantly outperforms existing event-based methods.
arXiv Detail & Related papers (2024-09-26T15:54:18Z) - Event-ECC: Asynchronous Tracking of Events with Continuous Optimization [1.9446776999250501]
We propose a tracking algorithm that computes a 2D motion warp per single event, called event-ECC (eECC)
The computational burden of event-wise processing is alleviated through a lightweight version that benefits from incremental processing and updating scheme.
We report improvements in tracking accuracy and feature age over state-of-the-art event-based asynchronous trackers.
arXiv Detail & Related papers (2024-09-22T19:03:19Z) - SpikeMOT: Event-based Multi-Object Tracking with Sparse Motion Features [52.213656737672935]
SpikeMOT is an event-based multi-object tracker.
SpikeMOT uses spiking neural networks to extract sparsetemporal features from event streams associated with objects.
arXiv Detail & Related papers (2023-09-29T05:13:43Z) - Graph-based Asynchronous Event Processing for Rapid Object Recognition [59.112755601918074]
Event cameras capture asynchronous events stream in which each event encodes pixel location, trigger time, and the polarity of the brightness changes.
We introduce a novel graph-based framework for event cameras, namely SlideGCN.
Our approach can efficiently process data event-by-event, unlock the low latency nature of events data while still maintaining the graph's structure internally.
arXiv Detail & Related papers (2023-08-28T08:59:57Z) - Asynchronous Blob Tracker for Event Cameras [5.64242497932732]
Event-based cameras are popular for tracking fast-moving objects due to their high temporal resolution, low latency and high range.
We propose a novel algorithm for tracking blobs using raw events asynchronously in real time.
Our algorithm achieves highly accurate blob tracking, velocity estimation, and shape estimation even under challenging lighting conditions.
arXiv Detail & Related papers (2023-07-20T05:15:03Z) - Dual Memory Aggregation Network for Event-Based Object Detection with
Learnable Representation [79.02808071245634]
Event-based cameras are bio-inspired sensors that capture brightness change of every pixel in an asynchronous manner.
Event streams are divided into grids in the x-y-t coordinates for both positive and negative polarity, producing a set of pillars as 3D tensor representation.
Long memory is encoded in the hidden state of adaptive convLSTMs while short memory is modeled by computing spatial-temporal correlation between event pillars.
arXiv Detail & Related papers (2023-03-17T12:12:41Z) - Unsupervised Feature Learning for Event Data: Direct vs Inverse Problem
Formulation [53.850686395708905]
Event-based cameras record an asynchronous stream of per-pixel brightness changes.
In this paper, we focus on single-layer architectures for representation learning from event data.
We show improvements of up to 9 % in the recognition accuracy compared to the state-of-the-art methods.
arXiv Detail & Related papers (2020-09-23T10:40:03Z) - Asynchronous Tracking-by-Detection on Adaptive Time Surfaces for
Event-based Object Tracking [87.0297771292994]
We propose an Event-based Tracking-by-Detection (ETD) method for generic bounding box-based object tracking.
To achieve this goal, we present an Adaptive Time-Surface with Linear Time Decay (ATSLTD) event-to-frame conversion algorithm.
We compare the proposed ETD method with seven popular object tracking methods, that are based on conventional cameras or event cameras, and two variants of ETD.
arXiv Detail & Related papers (2020-02-13T15:58:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.