CONSAC: Robust Multi-Model Fitting by Conditional Sample Consensus
- URL: http://arxiv.org/abs/2001.02643v3
- Date: Wed, 25 Mar 2020 11:48:39 GMT
- Title: CONSAC: Robust Multi-Model Fitting by Conditional Sample Consensus
- Authors: Florian Kluger, Eric Brachmann, Hanno Ackermann, Carsten Rother,
Michael Ying Yang, Bodo Rosenhahn
- Abstract summary: We present a robust estimator for fitting multiple parametric models of the same form to noisy measurements.
In contrast to previous works, which resorted to hand-crafted search strategies for multiple model detection, we learn the search strategy from data.
For self-supervised learning of the search, we evaluate the proposed algorithm on multi-homography estimation and demonstrate an accuracy that is superior to state-of-the-art methods.
- Score: 62.86856923633923
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a robust estimator for fitting multiple parametric models of the
same form to noisy measurements. Applications include finding multiple
vanishing points in man-made scenes, fitting planes to architectural imagery,
or estimating multiple rigid motions within the same sequence. In contrast to
previous works, which resorted to hand-crafted search strategies for multiple
model detection, we learn the search strategy from data. A neural network
conditioned on previously detected models guides a RANSAC estimator to
different subsets of all measurements, thereby finding model instances one
after another. We train our method supervised as well as self-supervised. For
supervised training of the search strategy, we contribute a new dataset for
vanishing point estimation. Leveraging this dataset, the proposed algorithm is
superior with respect to other robust estimators as well as to designated
vanishing point estimation algorithms. For self-supervised learning of the
search, we evaluate the proposed algorithm on multi-homography estimation and
demonstrate an accuracy that is superior to state-of-the-art methods.
Related papers
- RHiOTS: A Framework for Evaluating Hierarchical Time Series Forecasting Algorithms [0.393259574660092]
RHiOTS is designed to assess the robustness of hierarchical time series forecasting models and algorithms on real-world datasets.
RHiOTS incorporates an innovative visualization component, turning complex, multidimensional robustness evaluation results into intuitive, easily interpretable visuals.
Our findings show that traditional statistical methods are more robust than state-of-the-art deep learning algorithms, except when the transformation effect is highly disruptive.
arXiv Detail & Related papers (2024-08-06T18:52:15Z) - Horseshoe-type Priors for Independent Component Estimation [0.4987670632802289]
Independent Component Estimation (ICE) has many applications in modern day machine learning.
Horseshoe-type priors are used to provide scalable algorithms.
We show how to implement conditional posteriors and envelope-based methods for optimization.
arXiv Detail & Related papers (2024-06-24T18:18:58Z) - Piecewise-Linear Manifolds for Deep Metric Learning [8.670873561640903]
Unsupervised deep metric learning focuses on learning a semantic representation space using only unlabeled data.
We propose to model the high-dimensional data manifold using a piecewise-linear approximation, with each low-dimensional linear piece approximating the data manifold in a small neighborhood of a point.
We empirically show that this similarity estimate correlates better with the ground truth than the similarity estimates of current state-of-the-art techniques.
arXiv Detail & Related papers (2024-03-22T06:22:20Z) - Consensus-Adaptive RANSAC [104.87576373187426]
We propose a new RANSAC framework that learns to explore the parameter space by considering the residuals seen so far via a novel attention layer.
The attention mechanism operates on a batch of point-to-model residuals, and updates a per-point estimation state to take into account the consensus found through a lightweight one-step transformer.
arXiv Detail & Related papers (2023-07-26T08:25:46Z) - Towards Deterministic Diverse Subset Sampling [14.236193187116049]
In this paper, we discuss a greedy deterministic adaptation of k-DPP.
We demonstrate the usefulness of the model on an image search task.
arXiv Detail & Related papers (2021-05-28T16:05:58Z) - Estimating leverage scores via rank revealing methods and randomization [50.591267188664666]
We study algorithms for estimating the statistical leverage scores of rectangular dense or sparse matrices of arbitrary rank.
Our approach is based on combining rank revealing methods with compositions of dense and sparse randomized dimensionality reduction transforms.
arXiv Detail & Related papers (2021-05-23T19:21:55Z) - Finding Geometric Models by Clustering in the Consensus Space [61.65661010039768]
We propose a new algorithm for finding an unknown number of geometric models, e.g., homographies.
We present a number of applications where the use of multiple geometric models improves accuracy.
These include pose estimation from multiple generalized homographies; trajectory estimation of fast-moving objects.
arXiv Detail & Related papers (2021-03-25T14:35:07Z) - Evaluating the Disentanglement of Deep Generative Models through
Manifold Topology [66.06153115971732]
We present a method for quantifying disentanglement that only uses the generative model.
We empirically evaluate several state-of-the-art models across multiple datasets.
arXiv Detail & Related papers (2020-06-05T20:54:11Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z) - PAC-Bayes meta-learning with implicit task-specific posteriors [37.32107678838193]
We introduce a new and rigorously-formulated PAC-Bayes meta-learning algorithm that solves few-shot learning.
We show that the models trained with our proposed meta-learning algorithm are well calibrated and accurate.
arXiv Detail & Related papers (2020-03-05T06:56:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.