Localization with Sampling-Argmax
- URL: http://arxiv.org/abs/2110.08825v1
- Date: Sun, 17 Oct 2021 13:56:25 GMT
- Title: Localization with Sampling-Argmax
- Authors: Jiefeng Li, Tong Chen, Ruiqi Shi, Yujing Lou, Yong-Lu Li, Cewu Lu
- Abstract summary: We propose sampling-argmax, a differentiable training method that imposes implicit constraints to the shape of the probability map.
We show that sampling-argmax can seamlessly replace the conventional soft-argmax operation on various localization tasks.
- Score: 45.408767601013786
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Soft-argmax operation is commonly adopted in detection-based methods to
localize the target position in a differentiable manner. However, training the
neural network with soft-argmax makes the shape of the probability map
unconstrained. Consequently, the model lacks pixel-wise supervision through the
map during training, leading to performance degradation. In this work, we
propose sampling-argmax, a differentiable training method that imposes implicit
constraints to the shape of the probability map by minimizing the expectation
of the localization error. To approximate the expectation, we introduce a
continuous formulation of the output distribution and develop a differentiable
sampling process. The expectation can be approximated by calculating the
average error of all samples drawn from the output distribution. We show that
sampling-argmax can seamlessly replace the conventional soft-argmax operation
on various localization tasks. Comprehensive experiments demonstrate the
effectiveness and flexibility of the proposed method. Code is available at
https://github.com/Jeff-sjtu/sampling-argmax
Related papers
- Controllable Generation via Locally Constrained Resampling [77.48624621592523]
We propose a tractable probabilistic approach that performs Bayesian conditioning to draw samples subject to a constraint.
Our approach considers the entire sequence, leading to a more globally optimal constrained generation than current greedy methods.
We show that our approach is able to steer the model's outputs away from toxic generations, outperforming similar approaches to detoxification.
arXiv Detail & Related papers (2024-10-17T00:49:53Z) - The Sampling-Gaussian for stereo matching [7.9898209414259425]
The soft-argmax operation is widely adopted in neural network-based stereo matching methods.
Previous methods failed to effectively improve the accuracy and even compromises the efficiency of the network.
We propose a novel supervision method for stereo matching, Sampling-Gaussian.
arXiv Detail & Related papers (2024-10-09T03:57:13Z) - DistPred: A Distribution-Free Probabilistic Inference Method for Regression and Forecasting [14.390842560217743]
We propose a novel approach called DistPred for regression and forecasting tasks.
We transform proper scoring rules that measure the discrepancy between the predicted distribution and the target distribution into a differentiable discrete form.
This allows the model to sample numerous samples in a single forward pass to estimate the potential distribution of the response variable.
arXiv Detail & Related papers (2024-06-17T10:33:00Z) - Dirichlet-Based Prediction Calibration for Learning with Noisy Labels [40.78497779769083]
Learning with noisy labels can significantly hinder the generalization performance of deep neural networks (DNNs)
Existing approaches address this issue through loss correction or example selection methods.
We propose the textitDirichlet-based Prediction (DPC) method as a solution.
arXiv Detail & Related papers (2024-01-13T12:33:04Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Efficient Multimodal Sampling via Tempered Distribution Flow [11.36635610546803]
We develop a new type of transport-based sampling method called TemperFlow.
Various experiments demonstrate the superior performance of this novel sampler compared to traditional methods.
We show its applications in modern deep learning tasks such as image generation.
arXiv Detail & Related papers (2023-04-08T06:40:06Z) - Generalized Differentiable RANSAC [95.95627475224231]
$nabla$-RANSAC is a differentiable RANSAC that allows learning the entire randomized robust estimation pipeline.
$nabla$-RANSAC is superior to the state-of-the-art in terms of accuracy while running at a similar speed to its less accurate alternatives.
arXiv Detail & Related papers (2022-12-26T15:13:13Z) - Diverse Human Motion Prediction via Gumbel-Softmax Sampling from an
Auxiliary Space [34.83587750498361]
Diverse human motion prediction aims at predicting multiple possible future pose sequences from a sequence of observed poses.
Previous approaches usually employ deep generative networks to model the conditional distribution of data, and then randomly sample outcomes from the distribution.
We propose a novel sampling strategy for sampling very diverse results from an imbalanced multimodal distribution.
arXiv Detail & Related papers (2022-07-15T09:03:57Z) - Improving Maximum Likelihood Training for Text Generation with Density
Ratio Estimation [51.091890311312085]
We propose a new training scheme for auto-regressive sequence generative models, which is effective and stable when operating at large sample space encountered in text generation.
Our method stably outperforms Maximum Likelihood Estimation and other state-of-the-art sequence generative models in terms of both quality and diversity.
arXiv Detail & Related papers (2020-07-12T15:31:24Z) - Spatially Adaptive Inference with Stochastic Feature Sampling and
Interpolation [72.40827239394565]
We propose to compute features only at sparsely sampled locations.
We then densely reconstruct the feature map with an efficient procedure.
The presented network is experimentally shown to save substantial computation while maintaining accuracy over a variety of computer vision tasks.
arXiv Detail & Related papers (2020-03-19T15:36:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.