"Forget" the Forget Gate: Estimating Anomalies in Videos using
Self-contained Long Short-Term Memory Networks
- URL: http://arxiv.org/abs/2104.01478v1
- Date: Sat, 3 Apr 2021 20:43:49 GMT
- Title: "Forget" the Forget Gate: Estimating Anomalies in Videos using
Self-contained Long Short-Term Memory Networks
- Authors: Habtamu Fanta, Zhiwen Shao, Lizhuang Ma
- Abstract summary: We present an approach of detecting anomalies in videos by learning a novel LSTM based self-contained network on normal dense optical flow.
We introduce a bi-gated, light LSTM cell by discarding the forget gate and introducing sigmoid activation.
Removing the forget gate results in a simplified and undemanding LSTM cell with improved performance effectiveness and computational efficiency.
- Score: 20.211951213040937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Abnormal event detection is a challenging task that requires effectively
handling intricate features of appearance and motion. In this paper, we present
an approach of detecting anomalies in videos by learning a novel LSTM based
self-contained network on normal dense optical flow. Due to their sigmoid
implementations, standard LSTM's forget gate is susceptible to overlooking and
dismissing relevant content in long sequence tasks like abnormality detection.
The forget gate mitigates participation of previous hidden state for
computation of cell state prioritizing current input. In addition, the
hyperbolic tangent activation of standard LSTMs sacrifices performance when a
network gets deeper. To tackle these two limitations, we introduce a bi-gated,
light LSTM cell by discarding the forget gate and introducing sigmoid
activation. Specifically, the LSTM architecture we come up with fully sustains
content from previous hidden state thereby enabling the trained model to be
robust and make context-independent decision during evaluation. Removing the
forget gate results in a simplified and undemanding LSTM cell with improved
performance effectiveness and computational efficiency. Empirical evaluations
show that the proposed bi-gated LSTM based network outperforms various LSTM
based models verifying its effectiveness for abnormality detection and
generalization tasks on CUHK Avenue and UCSD datasets.
Related papers
- RigLSTM: Recurrent Independent Grid LSTM for Generalizable Sequence
Learning [75.61681328968714]
We propose recurrent independent Grid LSTM (RigLSTM) to exploit the underlying modular structure of the target task.
Our model adopts cell selection, input feature selection, hidden state selection, and soft state updating to achieve a better generalization ability.
arXiv Detail & Related papers (2023-11-03T07:40:06Z) - Towards Energy-Efficient, Low-Latency and Accurate Spiking LSTMs [1.7969777786551424]
Spiking Neural Networks (SNNs) have emerged as an attractive-temporal computing paradigm vision for complex tasks.
We propose an optimized spiking long short-term memory networks (LSTM) training framework that involves a novel.
rev-to-SNN conversion framework, followed by SNN training.
We evaluate our framework on sequential learning tasks including temporal M, Google Speech Commands (GSC) datasets, and UCI Smartphone on different LSTM architectures.
arXiv Detail & Related papers (2022-10-23T04:10:27Z) - Extreme-Long-short Term Memory for Time-series Prediction [0.0]
Long Short-Term Memory (LSTM) is a new type of Recurrent Neural Networks (RNN)
In this paper, we propose an advanced LSTM algorithm, the Extreme Long Short-Term Memory (E-LSTM)
The new E-LSTM requires only 2 epochs to obtain the results of the 7th epoch traditional LSTM.
arXiv Detail & Related papers (2022-10-15T09:45:48Z) - Self-Supervised Masked Convolutional Transformer Block for Anomaly
Detection [122.4894940892536]
We present a novel self-supervised masked convolutional transformer block (SSMCTB) that comprises the reconstruction-based functionality at a core architectural level.
In this work, we extend our previous self-supervised predictive convolutional attentive block (SSPCAB) with a 3D masked convolutional layer, a transformer for channel-wise attention, as well as a novel self-supervised objective based on Huber loss.
arXiv Detail & Related papers (2022-09-25T04:56:10Z) - Braille Letter Reading: A Benchmark for Spatio-Temporal Pattern
Recognition on Neuromorphic Hardware [50.380319968947035]
Recent deep learning approaches have reached accuracy in such tasks, but their implementation on conventional embedded solutions is still computationally very and energy expensive.
We propose a new benchmark for computing tactile pattern recognition at the edge through letters reading.
We trained and compared feed-forward and recurrent spiking neural networks (SNNs) offline using back-propagation through time with surrogate gradients, then we deployed them on the Intel Loihimorphic chip for efficient inference.
Our results show that the LSTM outperforms the recurrent SNN in terms of accuracy by 14%. However, the recurrent SNN on Loihi is 237 times more energy
arXiv Detail & Related papers (2022-05-30T14:30:45Z) - Working Memory Connections for LSTM [51.742526187978726]
We show that Working Memory Connections constantly improve the performance of LSTMs on a variety of tasks.
Numerical results suggest that the cell state contains useful information that is worth including in the gate structure.
arXiv Detail & Related papers (2021-08-31T18:01:30Z) - Object Tracking through Residual and Dense LSTMs [67.98948222599849]
Deep learning-based trackers based on LSTMs (Long Short-Term Memory) recurrent neural networks have emerged as a powerful alternative.
DenseLSTMs outperform Residual and regular LSTM, and offer a higher resilience to nuisances.
Our case study supports the adoption of residual-based RNNs for enhancing the robustness of other trackers.
arXiv Detail & Related papers (2020-06-22T08:20:17Z) - Unsupervised Online Anomaly Detection On Irregularly Sampled Or Missing
Valued Time-Series Data Using LSTM Networks [0.0]
We study anomaly detection and introduce an algorithm that processes variable length, irregularly sampled sequences or sequences with missing values.
Our algorithm is fully unsupervised, however, can be readily extended to supervised or semisupervised cases.
arXiv Detail & Related papers (2020-05-25T09:41:04Z) - SiTGRU: Single-Tunnelled Gated Recurrent Unit for Abnormality Detection [29.500392184282518]
We propose a novel version of Gated Recurrent Unit (GRU) called Single Tunnelled GRU for abnormality detection.
Our proposed optimized GRU model outperforms standard GRU and Long Short Term Memory (LSTM) networks on most metrics for detection and generalization tasks.
arXiv Detail & Related papers (2020-03-30T14:58:13Z) - Refined Gate: A Simple and Effective Gating Mechanism for Recurrent
Units [68.30422112784355]
We propose a new gating mechanism within general gated recurrent neural networks to handle this issue.
The proposed gates directly short connect the extracted input features to the outputs of vanilla gates.
We verify the proposed gating mechanism on three popular types of gated RNNs including LSTM, GRU and MGU.
arXiv Detail & Related papers (2020-02-26T07:51:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.