Deep hierarchical pooling design for cross-granularity action
recognition
- URL: http://arxiv.org/abs/2006.04473v1
- Date: Mon, 8 Jun 2020 11:03:54 GMT
- Title: Deep hierarchical pooling design for cross-granularity action
recognition
- Authors: Ahmed Mazari and Hichem Sahbi
- Abstract summary: We introduce a novel hierarchical aggregation design that captures different levels of temporal granularity in action recognition.
Learning the combination of operations in this network -- which best fits a given ground-truth -- is obtained by solving a constrained minimization problem.
Besides being principled and well grounded, the proposed hierarchical pooling is also video-length and resilient to misalignments in actions.
- Score: 14.696233190562936
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we introduce a novel hierarchical aggregation design that
captures different levels of temporal granularity in action recognition. Our
design principle is coarse-to-fine and achieved using a tree-structured
network; as we traverse this network top-down, pooling operations are getting
less invariant but timely more resolute and well localized. Learning the
combination of operations in this network -- which best fits a given
ground-truth -- is obtained by solving a constrained minimization problem whose
solution corresponds to the distribution of weights that capture the
contribution of each level (and thereby temporal granularity) in the global
hierarchical pooling process. Besides being principled and well grounded, the
proposed hierarchical pooling is also video-length agnostic and resilient to
misalignments in actions. Extensive experiments conducted on the challenging
UCF-101 database corroborate these statements.
Related papers
- Component-based Sketching for Deep ReLU Nets [55.404661149594375]
We develop a sketching scheme based on deep net components for various tasks.
We transform deep net training into a linear empirical risk minimization problem.
We show that the proposed component-based sketching provides almost optimal rates in approximating saturated functions.
arXiv Detail & Related papers (2024-09-21T15:30:43Z) - Miniaturized Graph Convolutional Networks with Topologically Consistent
Pruning [12.18340575383456]
We devise a novel magnitude pruning method that allows extractingworks while guarantying their topological consistency.
The latter ensures that only accessible and co-accessible connections are kept in the resulting lightweight networks.
Our solution is based on a novel reparametrization and two supervisory bi-directional networks which implement accessibility/co-accessibility.
arXiv Detail & Related papers (2023-06-30T12:09:22Z) - Augmenting Convolutional networks with attention-based aggregation [55.97184767391253]
We show how to augment any convolutional network with an attention-based global map to achieve non-local reasoning.
We plug this learned aggregation layer with a simplistic patch-based convolutional network parametrized by 2 parameters (width and depth)
It yields surprisingly competitive trade-offs between accuracy and complexity, in particular in terms of memory consumption.
arXiv Detail & Related papers (2021-12-27T14:05:41Z) - Clustering-Based Interpretation of Deep ReLU Network [17.234442722611803]
We recognize that the non-linear behavior of the ReLU function gives rise to a natural clustering.
We propose a method to increase the level of interpretability of a fully connected feedforward ReLU neural network.
arXiv Detail & Related papers (2021-10-13T09:24:11Z) - SIRe-Networks: Skip Connections over Interlaced Multi-Task Learning and
Residual Connections for Structure Preserving Object Classification [28.02302915971059]
In this paper, we introduce an interlaced multi-task learning strategy, defined SIRe, to reduce the vanishing gradient in relation to the object classification task.
The presented methodology directly improves a convolutional neural network (CNN) by enforcing the input image structure preservation through auto-encoders.
To validate the presented methodology, a simple CNN and various implementations of famous networks are extended via the SIRe strategy and extensively tested on the CIFAR100 dataset.
arXiv Detail & Related papers (2021-10-06T13:54:49Z) - Unsupervised Monocular Depth Learning with Integrated Intrinsics and
Spatio-Temporal Constraints [61.46323213702369]
This work presents an unsupervised learning framework that is able to predict at-scale depth maps and egomotion.
Our results demonstrate strong performance when compared to the current state-of-the-art on multiple sequences of the KITTI driving dataset.
arXiv Detail & Related papers (2020-11-02T22:26:58Z) - The Heterogeneity Hypothesis: Finding Layer-Wise Differentiated Network
Architectures [179.66117325866585]
We investigate a design space that is usually overlooked, i.e. adjusting the channel configurations of predefined networks.
We find that this adjustment can be achieved by shrinking widened baseline networks and leads to superior performance.
Experiments are conducted on various networks and datasets for image classification, visual tracking and image restoration.
arXiv Detail & Related papers (2020-06-29T17:59:26Z) - Action Recognition with Deep Multiple Aggregation Networks [14.696233190562936]
We introduce a novel hierarchical pooling design that captures different levels of temporal granularity in action recognition.
Our design principle is coarse-to-fine and achieved using a tree-structured network.
Besides being principled and well grounded, the proposed hierarchical pooling is also video-length and resolution agnostic.
arXiv Detail & Related papers (2020-06-08T11:37:38Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z) - Local Propagation in Constraint-based Neural Network [77.37829055999238]
We study a constraint-based representation of neural network architectures.
We investigate a simple optimization procedure that is well suited to fulfil the so-called architectural constraints.
arXiv Detail & Related papers (2020-02-18T16:47:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.