Incorporating Exemplar Optimization into Training with Dual Networks for
Human Mesh Recovery
- URL: http://arxiv.org/abs/2401.14121v1
- Date: Thu, 25 Jan 2024 12:04:53 GMT
- Title: Incorporating Exemplar Optimization into Training with Dual Networks for
Human Mesh Recovery
- Authors: Yongwei Nie, Mingxian Fan, Chengjiang Long, Qing Zhang, Jian Zhu,
Xuemiao Xu
- Abstract summary: We propose a novel optimization-based human mesh recovery method from a single image.
We incorporate exemplar optimization into the training stage.
Experiments demonstrate that our exemplar optimization after the novel training scheme significantly outperforms state-of-the-art approaches.
- Score: 37.233112051529574
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel optimization-based human mesh recovery method from a
single image. Given a test exemplar, previous approaches optimize the
pre-trained regression network to minimize the 2D re-projection loss, which
however suffer from over-/under-fitting problems. This is because the
``exemplar optimization'' at testing time has too weak relation to the
pre-training process, and the exemplar optimization loss function is different
from the training loss function. (1) We incorporate exemplar optimization into
the training stage. During training, our method first executes exemplar
optimization and subsequently proceeds with training-time optimization. The
exemplar optimization may run into a wrong direction, while the subsequent
training optimization serves to correct the deviation. Involved in training,
the exemplar optimization learns to adapt its behavior to training data,
thereby acquires generalibility to test exemplars. (2) We devise a dual-network
architecture to convey the novel training paradigm, which is composed of a main
regression network and an auxiliary network, in which we can formulate the
exemplar optimization loss function in the same form as the training loss
function. This further enhances the compatibility between the exemplar and
training optimizations. Experiments demonstrate that our exemplar optimization
after the novel training scheme significantly outperforms state-of-the-art
approaches.
Related papers
- AdaLRS: Loss-Guided Adaptive Learning Rate Search for Efficient Foundation Model Pretraining [12.630306478872043]
We propose textbfAdaLRS, a plug-in-and-play adaptive learning rate search algorithm that conducts online optimal learning rate search.<n>Experiments show that AdaLRS adjusts suboptimal learning rates to the neighborhood of optimum with marked efficiency and effectiveness.
arXiv Detail & Related papers (2025-06-16T09:14:01Z) - Optimizing ML Training with Metagradient Descent [69.89631748402377]
We introduce an algorithm for efficiently calculating metagradients -- gradients through model training -- at scale.
We then introduce a "smooth model training" framework that enables effective optimization using metagradients.
arXiv Detail & Related papers (2025-03-17T22:18:24Z) - Learning to Generate Gradients for Test-Time Adaptation via Test-Time Training Layers [18.921532965557475]
Test-time adaptation aims to fine-tune a trained model online using unlabeled testing data.
In this optimization process, unsupervised learning objectives like entropy frequently encounter noisy learning signals.
We employ a learning-to-optimize approach to automatically learn an entropy generator called Meta Gradient Generator.
arXiv Detail & Related papers (2024-12-22T07:24:09Z) - A hybrid framework for effective and efficient machine unlearning [12.499101994047862]
Machine unlearning (MU) is proposed to remove the imprints of revoked samples from the already trained model parameters.
We present a novel hybrid strategy on top of them to achieve an overall success.
arXiv Detail & Related papers (2024-12-19T03:59:26Z) - What Do Learning Dynamics Reveal About Generalization in LLM Reasoning? [83.83230167222852]
We find that a model's generalization behavior can be effectively characterized by a training metric we call pre-memorization train accuracy.
By connecting a model's learning behavior to its generalization, pre-memorization train accuracy can guide targeted improvements to training strategies.
arXiv Detail & Related papers (2024-11-12T09:52:40Z) - FREE: Faster and Better Data-Free Meta-Learning [77.90126669914324]
Data-Free Meta-Learning (DFML) aims to extract knowledge from a collection of pre-trained models without requiring the original data.
We introduce the Faster and Better Data-Free Meta-Learning framework, which contains: (i) a meta-generator for rapidly recovering training tasks from pre-trained models; and (ii) a meta-learner for generalizing to new unseen tasks.
arXiv Detail & Related papers (2024-05-02T03:43:19Z) - Architecture, Dataset and Model-Scale Agnostic Data-free Meta-Learning [119.70303730341938]
We propose ePisode cUrriculum inveRsion (ECI) during data-free meta training and invErsion calibRation following inner loop (ICFIL) during meta testing.
ECI adaptively increases the difficulty level of pseudo episodes according to the real-time feedback of the meta model.
We formulate the optimization process of meta training with ECI as an adversarial form in an end-to-end manner.
arXiv Detail & Related papers (2023-03-20T15:10:41Z) - Meta-Registration: Learning Test-Time Optimization for Single-Pair Image
Registration [0.37501702548174964]
This work formulates image registration as a meta-learning algorithm.
Experiments are presented in this paper using clinical transrectal ultrasound image data from 108 prostate cancer patients.
arXiv Detail & Related papers (2022-07-22T10:30:00Z) - MT3: Meta Test-Time Training for Self-Supervised Test-Time Adaption [69.76837484008033]
An unresolved problem in Deep Learning is the ability of neural networks to cope with domain shifts during test-time.
We combine meta-learning, self-supervision and test-time training to learn to adapt to unseen test distributions.
Our approach significantly improves the state-of-the-art results on the CIFAR-10-Corrupted image classification benchmark.
arXiv Detail & Related papers (2021-03-30T09:33:38Z) - How much progress have we made in neural network training? A New
Evaluation Protocol for Benchmarking Optimizers [86.36020260204302]
We propose a new benchmarking protocol to evaluate both end-to-end efficiency and data-addition training efficiency.
A human study is conducted to show that our evaluation protocol matches human tuning behavior better than the random search.
We then apply the proposed benchmarking framework to 7s and various tasks, including computer vision, natural language processing, reinforcement learning, and graph mining.
arXiv Detail & Related papers (2020-10-19T21:46:39Z) - Neural Model-based Optimization with Right-Censored Observations [42.530925002607376]
Neural networks (NNs) have been demonstrated to work well at the core of model-based optimization procedures.
We show that our trained regression models achieve a better predictive quality than several baselines.
arXiv Detail & Related papers (2020-09-29T07:32:30Z) - AutoSimulate: (Quickly) Learning Synthetic Data Generation [70.82315853981838]
We propose an efficient alternative for optimal synthetic data generation based on a novel differentiable approximation of the objective.
We demonstrate that the proposed method finds the optimal data distribution faster (up to $50times$), with significantly reduced training data generation (up to $30times$) and better accuracy ($+8.7%$) on real-world test datasets than previous methods.
arXiv Detail & Related papers (2020-08-16T11:36:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.