Online Hybrid Lightweight Representations Learning: Its Application to
Visual Tracking
- URL: http://arxiv.org/abs/2205.11179v1
- Date: Mon, 23 May 2022 10:31:14 GMT
- Title: Online Hybrid Lightweight Representations Learning: Its Application to
Visual Tracking
- Authors: Ilchae Jung, Minji Kim, Eunhyeok Park, Bohyung Han
- Abstract summary: This paper presents a novel hybrid representation learning framework for streaming data.
An image frame in a video is modeled by an ensemble of two distinct deep neural networks.
We incorporate the hybrid representation technique into an online visual tracking task.
- Score: 42.49852446519412
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a novel hybrid representation learning framework for
streaming data, where an image frame in a video is modeled by an ensemble of
two distinct deep neural networks; one is a low-bit quantized network and the
other is a lightweight full-precision network. The former learns coarse primary
information with low cost while the latter conveys residual information for
high fidelity to original representations. The proposed parallel architecture
is effective to maintain complementary information since fixed-point arithmetic
can be utilized in the quantized network and the lightweight model provides
precise representations given by a compact channel-pruned network. We
incorporate the hybrid representation technique into an online visual tracking
task, where deep neural networks need to handle temporal variations of target
appearances in real-time. Compared to the state-of-the-art real-time trackers
based on conventional deep neural networks, our tracking algorithm demonstrates
competitive accuracy on the standard benchmarks with a small fraction of
computational cost and memory footprint.
Related papers
- TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - Efficient and Accurate Hyperspectral Image Demosaicing with Neural Network Architectures [3.386560551295746]
This study investigates the effectiveness of neural network architectures in hyperspectral image demosaicing.
We introduce a range of network models and modifications, and compare them with classical methods and existing reference network approaches.
Results indicate that our networks outperform or match reference models in both datasets demonstrating exceptional performance.
arXiv Detail & Related papers (2023-12-21T08:02:49Z) - Building a Graph-based Deep Learning network model from captured traffic
traces [4.671648049111933]
State of the art network models are based or depend on Discrete Event Simulation (DES)
DES is highly accurate, it is also computationally costly and cumbersome to parallelize, making it unpractical to simulate high performance networks.
We propose a Graph Neural Network (GNN)-based solution specifically designed to better capture the complexities of real network scenarios.
arXiv Detail & Related papers (2023-10-18T11:16:32Z) - Neural Maximum A Posteriori Estimation on Unpaired Data for Motion
Deblurring [87.97330195531029]
We propose a Neural Maximum A Posteriori (NeurMAP) estimation framework for training neural networks to recover blind motion information and sharp content from unpaired data.
The proposed NeurMAP is an approach to existing deblurring neural networks, and is the first framework that enables training image deblurring networks on unpaired datasets.
arXiv Detail & Related papers (2022-04-26T08:09:47Z) - Event Neural Networks [13.207573300016277]
Event Neural Networks (EvNets) leverage repetition to achieve considerable savings for video inference tasks.
We show that it is possible to transform virtually any conventional neural into an EvNet.
We demonstrate the effectiveness of our method on several state-of-the-art neural networks for both high- and low-level visual processing.
arXiv Detail & Related papers (2021-12-02T00:08:48Z) - CondenseNeXt: An Ultra-Efficient Deep Neural Network for Embedded
Systems [0.0]
A Convolutional Neural Network (CNN) is a class of Deep Neural Network (DNN) widely used in the analysis of visual images captured by an image sensor.
In this paper, we propose a neoteric variant of deep convolutional neural network architecture to ameliorate the performance of existing CNN architectures for real-time inference on embedded systems.
arXiv Detail & Related papers (2021-12-01T18:20:52Z) - Neural BRDF Representation and Importance Sampling [79.84316447473873]
We present a compact neural network-based representation of reflectance BRDF data.
We encode BRDFs as lightweight networks, and propose a training scheme with adaptive angular sampling.
We evaluate encoding results on isotropic and anisotropic BRDFs from multiple real-world datasets.
arXiv Detail & Related papers (2021-02-11T12:00:24Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Graph-Based Neural Network Models with Multiple Self-Supervised
Auxiliary Tasks [79.28094304325116]
Graph Convolutional Networks are among the most promising approaches for capturing relationships among structured data points.
We propose three novel self-supervised auxiliary tasks to train graph-based neural network models in a multi-task fashion.
arXiv Detail & Related papers (2020-11-14T11:09:51Z) - Exploring the Connection Between Binary and Spiking Neural Networks [1.329054857829016]
We bridge the recent algorithmic progress in training Binary Neural Networks and Spiking Neural Networks.
We show that training Spiking Neural Networks in the extreme quantization regime results in near full precision accuracies on large-scale datasets.
arXiv Detail & Related papers (2020-02-24T03:46:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.