Incorporating Exemplar Optimization into Training with Dual Networks for
Human Mesh Recovery
- URL: http://arxiv.org/abs/2401.14121v1
- Date: Thu, 25 Jan 2024 12:04:53 GMT
- Title: Incorporating Exemplar Optimization into Training with Dual Networks for
Human Mesh Recovery
- Authors: Yongwei Nie, Mingxian Fan, Chengjiang Long, Qing Zhang, Jian Zhu,
Xuemiao Xu
- Abstract summary: We propose a novel optimization-based human mesh recovery method from a single image.
We incorporate exemplar optimization into the training stage.
Experiments demonstrate that our exemplar optimization after the novel training scheme significantly outperforms state-of-the-art approaches.
- Score: 37.233112051529574
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel optimization-based human mesh recovery method from a
single image. Given a test exemplar, previous approaches optimize the
pre-trained regression network to minimize the 2D re-projection loss, which
however suffer from over-/under-fitting problems. This is because the
``exemplar optimization'' at testing time has too weak relation to the
pre-training process, and the exemplar optimization loss function is different
from the training loss function. (1) We incorporate exemplar optimization into
the training stage. During training, our method first executes exemplar
optimization and subsequently proceeds with training-time optimization. The
exemplar optimization may run into a wrong direction, while the subsequent
training optimization serves to correct the deviation. Involved in training,
the exemplar optimization learns to adapt its behavior to training data,
thereby acquires generalibility to test exemplars. (2) We devise a dual-network
architecture to convey the novel training paradigm, which is composed of a main
regression network and an auxiliary network, in which we can formulate the
exemplar optimization loss function in the same form as the training loss
function. This further enhances the compatibility between the exemplar and
training optimizations. Experiments demonstrate that our exemplar optimization
after the novel training scheme significantly outperforms state-of-the-art
approaches.
Related papers
- Fast Propagation is Better: Accelerating Single-Step Adversarial
Training via Sampling Subnetworks [69.54774045493227]
A drawback of adversarial training is the computational overhead introduced by the generation of adversarial examples.
We propose to exploit the interior building blocks of the model to improve efficiency.
Compared with previous methods, our method not only reduces the training cost but also achieves better model robustness.
arXiv Detail & Related papers (2023-10-24T01:36:20Z) - Transformer-Based Learned Optimization [37.84626515073609]
We propose a new approach to learned optimization where we represent the computation's update step using a neural network.
Our innovation is a new neural network architecture inspired by the classic BFGS algorithm.
We demonstrate the advantages of our approach on a benchmark composed of objective functions traditionally used for the evaluation of optimization algorithms.
arXiv Detail & Related papers (2022-12-02T09:47:08Z) - Self-Supervised Primal-Dual Learning for Constrained Optimization [19.965556179096385]
This paper studies how to train machine-learning models that directly approximate the optimal solutions of constrained optimization problems.
It proposes the idea of Primal-Dual Learning (PDL), a self-supervised training method that does not require a set of pre-solved instances or an optimization solver for training and inference.
arXiv Detail & Related papers (2022-08-18T20:07:10Z) - Optimization-Derived Learning with Essential Convergence Analysis of
Training and Hyper-training [52.39882976848064]
We design a Generalized Krasnoselskii-Mann (GKM) scheme based on fixed-point iterations as our fundamental ODL module.
Under the GKM scheme, a Bilevel Meta Optimization (BMO) algorithmic framework is constructed to solve the optimal training and hyper-training variables together.
arXiv Detail & Related papers (2022-06-16T01:50:25Z) - Joint inference and input optimization in equilibrium networks [68.63726855991052]
deep equilibrium model is a class of models that foregoes traditional network depth and instead computes the output of a network by finding the fixed point of a single nonlinear layer.
We show that there is a natural synergy between these two settings.
We demonstrate this strategy on various tasks such as training generative models while optimizing over latent codes, training models for inverse problems like denoising and inpainting, adversarial training and gradient based meta-learning.
arXiv Detail & Related papers (2021-11-25T19:59:33Z) - Optimizer Fusion: Efficient Training with Better Locality and
Parallelism [11.656318345362804]
Experimental results show that we can achieve an up to 20% training time reduction on various configurations.
Since our methods do not alter the algorithm, they can be used as a general "plug-in" technique to the training process.
arXiv Detail & Related papers (2021-04-01T03:44:13Z) - Training Learned Optimizers with Randomly Initialized Learned Optimizers [49.67678615506608]
We show that a population of randomly learneds can be used to train themselves from scratch in an online fashion.
A form of population based training is used to orchestrate this self-training.
We believe feedback loops of this type will be important and powerful in the future of machine learning.
arXiv Detail & Related papers (2021-01-14T19:07:17Z) - Efficient Robust Training via Backward Smoothing [125.91185167854262]
Adversarial training is the most effective strategy in defending against adversarial examples.
It suffers from high computational costs due to the iterative adversarial attacks in each training step.
Recent studies show that it is possible to achieve fast Adversarial Training by performing a single-step attack.
arXiv Detail & Related papers (2020-10-03T04:37:33Z) - A Novel DNN Training Framework via Data Sampling and Multi-Task
Optimization [7.001799696806368]
We propose a novel framework to train DNN models.
It generates multiple pairs of training and validation sets from the gross training set via random splitting.
It outputs the best, among all trained models, which has the overall best performance across the validation sets from all pairs.
arXiv Detail & Related papers (2020-07-02T10:58:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.