From Actions to Events: A Transfer Learning Approach Using Improved Deep
Belief Networks
- URL: http://arxiv.org/abs/2211.17045v1
- Date: Wed, 30 Nov 2022 14:47:10 GMT
- Title: From Actions to Events: A Transfer Learning Approach Using Improved Deep
Belief Networks
- Authors: Mateus Roder, Jurandy Almeida, Gustavo H. de Rosa, Leandro A. Passos,
Andr\'e L. D. Rossi, Jo\~ao P. Papa
- Abstract summary: This paper proposes a novel approach to map the knowledge from action recognition to event recognition using an energy-based model.
Such a model can process all frames simultaneously, carrying spatial and temporal information through the learning process.
- Score: 1.0554048699217669
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the last decade, exponential data growth supplied machine learning-based
algorithms' capacity and enabled their usage in daily-life activities.
Additionally, such an improvement is partially explained due to the advent of
deep learning techniques, i.e., stacks of simple architectures that end up in
more complex models. Although both factors produce outstanding results, they
also pose drawbacks regarding the learning process as training complex models
over large datasets are expensive and time-consuming. Such a problem is even
more evident when dealing with video analysis. Some works have considered
transfer learning or domain adaptation, i.e., approaches that map the knowledge
from one domain to another, to ease the training burden, yet most of them
operate over individual or small blocks of frames. This paper proposes a novel
approach to map the knowledge from action recognition to event recognition
using an energy-based model, denoted as Spectral Deep Belief Network. Such a
model can process all frames simultaneously, carrying spatial and temporal
information through the learning process. The experimental results conducted
over two public video dataset, the HMDB-51 and the UCF-101, depict the
effectiveness of the proposed model and its reduced computational burden when
compared to traditional energy-based models, such as Restricted Boltzmann
Machines and Deep Belief Networks.
Related papers
- Accelerating Deep Learning with Fixed Time Budget [2.190627491782159]
This paper proposes an effective technique for training arbitrary deep learning models within fixed time constraints.
The proposed method is extensively evaluated in both classification and regression tasks in computer vision.
arXiv Detail & Related papers (2024-10-03T21:18:04Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - Optimizing Dense Feed-Forward Neural Networks [0.0]
We propose a novel feed-forward neural network constructing method based on pruning and transfer learning.
Our approach can compress the number of parameters by more than 70%.
We also evaluate the transfer learning level comparing the refined model and the original one training from scratch a neural network.
arXiv Detail & Related papers (2023-12-16T23:23:16Z) - The Trifecta: Three simple techniques for training deeper
Forward-Forward networks [0.0]
We propose a collection of three techniques that synergize exceptionally well and drastically improve the Forward-Forward algorithm on deeper networks.
Our experiments demonstrate that our models are on par with similarly structured, backpropagation-based models in both training speed and test accuracy on simple datasets.
arXiv Detail & Related papers (2023-11-29T22:44:32Z) - Breaking the Curse of Dimensionality in Deep Neural Networks by Learning
Invariant Representations [1.9580473532948401]
This thesis explores the theoretical foundations of deep learning by studying the relationship between the architecture of these models and the inherent structures found within the data they process.
We ask What drives the efficacy of deep learning algorithms and allows them to beat the so-called curse of dimensionality.
Our methodology takes an empirical approach to deep learning, combining experimental studies with physics-inspired toy models.
arXiv Detail & Related papers (2023-10-24T19:50:41Z) - PILOT: A Pre-Trained Model-Based Continual Learning Toolbox [71.63186089279218]
This paper introduces a pre-trained model-based continual learning toolbox known as PILOT.
On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt.
On the other hand, PILOT fits typical class-incremental learning algorithms within the context of pre-trained models to evaluate their effectiveness.
arXiv Detail & Related papers (2023-09-13T17:55:11Z) - STAR: Sparse Transformer-based Action Recognition [61.490243467748314]
This work proposes a novel skeleton-based human action recognition model with sparse attention on the spatial dimension and segmented linear attention on the temporal dimension of data.
Experiments show that our model can achieve comparable performance while utilizing much less trainable parameters and achieve high speed in training and inference.
arXiv Detail & Related papers (2021-07-15T02:53:11Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z) - Collaborative Distillation in the Parameter and Spectrum Domains for
Video Action Recognition [79.60708268515293]
This paper explores how to train small and efficient networks for action recognition.
We propose two distillation strategies in the frequency domain, namely the feature spectrum and parameter distribution distillations respectively.
Our method can achieve higher performance than state-of-the-art methods with the same backbone.
arXiv Detail & Related papers (2020-09-15T07:29:57Z) - Knowledge Distillation: A Survey [87.51063304509067]
Deep neural networks have been successful in both industry and academia, especially for computer vision tasks.
It is a challenge to deploy these cumbersome deep models on devices with limited resources.
Knowledge distillation effectively learns a small student model from a large teacher model.
arXiv Detail & Related papers (2020-06-09T21:47:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.