Temporal Coherent Test-Time Optimization for Robust Video Classification
- URL: http://arxiv.org/abs/2302.14309v1
- Date: Tue, 28 Feb 2023 04:59:23 GMT
- Title: Temporal Coherent Test-Time Optimization for Robust Video Classification
- Authors: Chenyu Yi, Siyuan Yang, Yufei Wang, Haoliang Li, Yap-Peng Tan and Alex
C. Kot
- Abstract summary: Deep neural networks are likely to fail when the test data is corrupted in real-world deployment.
Test-time optimization is an effective way that adapts models to robustness to corrupted data during testing.
We propose a framework to utilize temporal information in test-time optimization for robust classification.
- Score: 55.432935503341064
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks are likely to fail when the test data is corrupted in
real-world deployment (e.g., blur, weather, etc.). Test-time optimization is an
effective way that adapts models to generalize to corrupted data during
testing, which has been shown in the image domain. However, the techniques for
improving video classification corruption robustness remain few. In this work,
we propose a Temporal Coherent Test-time Optimization framework (TeCo) to
utilize spatio-temporal information in test-time optimization for robust video
classification. To exploit information in video with self-supervised learning,
TeCo uses global content from video clips and optimizes models for entropy
minimization. TeCo minimizes the entropy of the prediction based on the global
content from video clips. Meanwhile, it also feeds local content to regularize
the temporal coherence at the feature level. TeCo retains the generalization
ability of various video classification models and achieves significant
improvements in corruption robustness across Mini Kinetics-C and Mini SSV2-C.
Furthermore, TeCo sets a new baseline in video classification corruption
robustness via test-time optimization.
Related papers
- SparseTem: Boosting the Efficiency of CNN-Based Video Encoders by Exploiting Temporal Continuity [15.872209884833977]
We propose a memory-efficient scheduling method to eliminate memory overhead and an online adjustment mechanism to minimize accuracy degradation.
SparseTem achieves speedup of 1.79x for EfficientDet and 4.72x for CRNN, with minimal accuracy drop and no additional memory overhead.
arXiv Detail & Related papers (2024-10-28T07:13:25Z) - Video Dynamics Prior: An Internal Learning Approach for Robust Video
Enhancements [83.5820690348833]
We present a framework for low-level vision tasks that does not require any external training data corpus.
Our approach learns neural modules by optimizing over a corrupted sequence, leveraging the weights of the coherence-temporal test and statistics internal statistics.
arXiv Detail & Related papers (2023-12-13T01:57:11Z) - Neighbourhood Representative Sampling for Efficient End-to-end Video
Quality Assessment [60.57703721744873]
The increased resolution of real-world videos presents a dilemma between efficiency and accuracy for deep Video Quality Assessment (VQA)
In this work, we propose a unified scheme, spatial-temporal grid mini-cube sampling (St-GMS) to get a novel type of sample, named fragments.
With fragments and FANet, the proposed efficient end-to-end FAST-VQA and FasterVQA achieve significantly better performance than existing approaches on all VQA benchmarks.
arXiv Detail & Related papers (2022-10-11T11:38:07Z) - TFCNet: Temporal Fully Connected Networks for Static Unbiased Temporal
Reasoning [3.4570413826505564]
Current video classification benchmarks contain strong biases towards static features, thus cannot accurately reflect the temporal modeling ability.
New video classification benchmarks aiming to eliminate static biases are proposed, with experiments on these new benchmarks showing that the current clip-based 3D CNNs are outperformed by RNN structures and recent video transformers.
With TFC blocks inserted into Video-level 3D CNNs (V3D), our proposed TFCNets establish new state-of-the-art results on synthetic temporal reasoning benchmark, CATER, and real world static-unbiased dataset, Diving48, surpassing all previous methods.
arXiv Detail & Related papers (2022-03-11T13:58:05Z) - Temporal-attentive Covariance Pooling Networks for Video Recognition [52.853765492522655]
existing video architectures usually generate global representation by using a simple global average pooling (GAP) method.
This paper proposes a attentive Covariance Pooling( TCP- TCP), inserted at the end of deep architectures, to produce powerful video representations.
Our TCP is model-agnostic and can be flexibly integrated into any video architectures, resulting in TCPNet for effective video recognition.
arXiv Detail & Related papers (2021-10-27T12:31:29Z) - AuxAdapt: Stable and Efficient Test-Time Adaptation for Temporally
Consistent Video Semantic Segmentation [81.87943324048756]
In video segmentation, generating temporally consistent results across frames is as important as achieving frame-wise accuracy.
Existing methods rely on optical flow regularization or fine-tuning with test data to attain temporal consistency.
This paper presents an efficient, intuitive, and unsupervised online adaptation method, AuxAdapt, for improving the temporal consistency of most neural network models.
arXiv Detail & Related papers (2021-10-24T07:07:41Z) - Multiple Instance-Based Video Anomaly Detection using Deep Temporal
Encoding-Decoding [5.255783459833821]
We propose a weakly supervised deep temporal encoding-decoding solution for anomaly detection in surveillance videos.
The proposed approach uses both abnormal and normal video clips during the training phase.
The results show that the proposed method performs similar to or better than the state-of-the-art solutions for anomaly detection in video surveillance applications.
arXiv Detail & Related papers (2020-07-03T08:22:42Z) - Tent: Fully Test-time Adaptation by Entropy Minimization [77.85911673550851]
A model must adapt itself to generalize to new and different data during testing.
In this setting of fully test-time adaptation the model has only the test data and its own parameters.
We propose to adapt by test entropy minimization (tent): we optimize the model for confidence as measured by the entropy of its predictions.
arXiv Detail & Related papers (2020-06-18T17:55:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.