Deep Convolutional Neural Network-based Bernoulli Heatmap for Head Pose
Estimation
- URL: http://arxiv.org/abs/2005.11780v1
- Date: Sun, 24 May 2020 15:36:29 GMT
- Title: Deep Convolutional Neural Network-based Bernoulli Heatmap for Head Pose
Estimation
- Authors: Zhongxu Hu, Yang Xing, Chen Lv, Peng Hang, Jie Liu
- Abstract summary: This paper proposes a novel Bernoulli heatmap for head pose estimation from a single RGB image.
Our method can achieve the positioning of the head area while estimating the angles of the head.
A deep convolutional neural network (CNN) structure with multiscale representations is adopted to maintain high-resolution information.
- Score: 11.676928225717337
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Head pose estimation is a crucial problem for many tasks, such as driver
attention, fatigue detection, and human behaviour analysis. It is well known
that neural networks are better at handling classification problems than
regression problems. It is an extremely nonlinear process to let the network
output the angle value directly for optimization learning, and the weight
constraint of the loss function will be relatively weak. This paper proposes a
novel Bernoulli heatmap for head pose estimation from a single RGB image. Our
method can achieve the positioning of the head area while estimating the angles
of the head. The Bernoulli heatmap makes it possible to construct fully
convolutional neural networks without fully connected layers and provides a new
idea for the output form of head pose estimation. A deep convolutional neural
network (CNN) structure with multiscale representations is adopted to maintain
high-resolution information and low-resolution information in parallel. This
kind of structure can maintain rich, high-resolution representations. In
addition, channelwise fusion is adopted to make the fusion weights learnable
instead of simple addition with equal weights. As a result, the estimation is
spatially more precise and potentially more accurate. The effectiveness of the
proposed method is empirically demonstrated by comparing it with other
state-of-the-art methods on public datasets.
Related papers
- Addressing caveats of neural persistence with deep graph persistence [54.424983583720675]
We find that the variance of network weights and spatial concentration of large weights are the main factors that impact neural persistence.
We propose an extension of the filtration underlying neural persistence to the whole neural network instead of single layers.
This yields our deep graph persistence measure, which implicitly incorporates persistent paths through the network and alleviates variance-related issues.
arXiv Detail & Related papers (2023-07-20T13:34:11Z) - Hidden Classification Layers: Enhancing linear separability between
classes in neural networks layers [0.0]
We investigate the impact on deep network performances of a training approach.
We propose a neural network architecture which induces an error function involving the outputs of all the network layers.
arXiv Detail & Related papers (2023-06-09T10:52:49Z) - Look beyond labels: Incorporating functional summary information in
Bayesian neural networks [11.874130244353253]
We present a simple approach to incorporate summary information about the predicted probability.
The available summary information is incorporated as augmented data and modeled with a Dirichlet process.
We show how the method can inform the model about task difficulty or class imbalance.
arXiv Detail & Related papers (2022-07-04T07:06:45Z) - Neural Maximum A Posteriori Estimation on Unpaired Data for Motion
Deblurring [87.97330195531029]
We propose a Neural Maximum A Posteriori (NeurMAP) estimation framework for training neural networks to recover blind motion information and sharp content from unpaired data.
The proposed NeurMAP is an approach to existing deblurring neural networks, and is the first framework that enables training image deblurring networks on unpaired datasets.
arXiv Detail & Related papers (2022-04-26T08:09:47Z) - Optimization-Based Separations for Neural Networks [57.875347246373956]
We show that gradient descent can efficiently learn ball indicator functions using a depth 2 neural network with two layers of sigmoidal activations.
This is the first optimization-based separation result where the approximation benefits of the stronger architecture provably manifest in practice.
arXiv Detail & Related papers (2021-12-04T18:07:47Z) - Non-Gradient Manifold Neural Network [79.44066256794187]
Deep neural network (DNN) generally takes thousands of iterations to optimize via gradient descent.
We propose a novel manifold neural network based on non-gradient optimization.
arXiv Detail & Related papers (2021-06-15T06:39:13Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - WoodFisher: Efficient Second-Order Approximation for Neural Network
Compression [35.45199662813043]
We develop a method to compute a faithful and efficient estimate of the inverse Hessian.
Our main application is to neural network compression.
We show how our method can be extended to take into account first-order information.
arXiv Detail & Related papers (2020-04-29T17:14:23Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z) - Mixed-Precision Quantized Neural Network with Progressively Decreasing
Bitwidth For Image Classification and Object Detection [21.48875255723581]
A mixed-precision quantized neural network with progressively ecreasing bitwidth is proposed to improve the trade-off between accuracy and compression.
Experiments on typical network architectures and benchmark datasets demonstrate that the proposed method could achieve better or comparable results.
arXiv Detail & Related papers (2019-12-29T14:11:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.