On Sinkhorn's Algorithm and Choice Modeling
- URL: http://arxiv.org/abs/2310.00260v1
- Date: Sat, 30 Sep 2023 05:20:23 GMT
- Title: On Sinkhorn's Algorithm and Choice Modeling
- Authors: Zhaonan Qu, Alfred Galichon, Johan Ugander
- Abstract summary: We show that the associated maximum likelihood estimation problems are equivalent to a classic matrix balancing problem with target row and column sums.
This perspective opens doors between two seemingly unrelated research areas.
We draw inspirations from these connections and resolve important open problems on the study of Sinkhorn's algorithm.
- Score: 6.43826005042477
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For a broad class of choice and ranking models based on Luce's choice axiom,
including the Bradley--Terry--Luce and Plackett--Luce models, we show that the
associated maximum likelihood estimation problems are equivalent to a classic
matrix balancing problem with target row and column sums. This perspective
opens doors between two seemingly unrelated research areas, and allows us to
unify existing algorithms in the choice modeling literature as special
instances or analogs of Sinkhorn's celebrated algorithm for matrix balancing.
We draw inspirations from these connections and resolve important open problems
on the study of Sinkhorn's algorithm. We first prove the global linear
convergence of Sinkhorn's algorithm for non-negative matrices whenever finite
solutions to the matrix balancing problem exist. We characterize this global
rate of convergence in terms of the algebraic connectivity of the bipartite
graph constructed from data. Next, we also derive the sharp asymptotic rate of
linear convergence, which generalizes a classic result of Knight (2008), but
with a more explicit analysis that exploits an intrinsic orthogonality
structure. To our knowledge, these are the first quantitative linear
convergence results for Sinkhorn's algorithm for general non-negative matrices
and positive marginals. The connections we establish in this paper between
matrix balancing and choice modeling could help motivate further transmission
of ideas and interesting results in both directions.
Related papers
- Doubly Stochastic Adaptive Neighbors Clustering via the Marcus Mapping [56.57574396804837]
Clustering is a fundamental task in machine learning and data science, and similarity graph-based clustering is an important approach within this domain.
We study the relationship between the Marcus mapping and optimal transport.
We prove that the Marcus mapping solves a specific type of optimal transport problem and demonstrate that solving this problem through Marcus mapping is more efficient than directly applying optimal transport methods.
arXiv Detail & Related papers (2024-08-06T03:34:43Z) - An SDP-based Branch-and-Cut Algorithm for Biclustering [0.0]
We present a tailored branch-and-cut algorithm for biclustering problems.
We show that the proposed algorithm can solve instances 20 times larger than those handled by general-purpose solvers.
arXiv Detail & Related papers (2024-03-17T21:43:19Z) - Data Assimilation for Sign-indefinite Priors: A generalization of
Sinkhorn's algorithm [0.0]
We seek to revise sign-indefinite multi-dimensional arrays in a way that the updated values agree with specified marginals.
Our approach follows the rationale in Schr"odinger's problem, aimed at updating a "prior" probability measure to agree with marginal distributions.
arXiv Detail & Related papers (2023-08-22T21:13:39Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Semi-Supervised Subspace Clustering via Tensor Low-Rank Representation [64.49871502193477]
We propose a novel semi-supervised subspace clustering method, which is able to simultaneously augment the initial supervisory information and construct a discriminative affinity matrix.
Comprehensive experimental results on six commonly-used benchmark datasets demonstrate the superiority of our method over state-of-the-art methods.
arXiv Detail & Related papers (2022-05-21T01:47:17Z) - Sparse Quadratic Optimisation over the Stiefel Manifold with Application
to Permutation Synchronisation [71.27989298860481]
We address the non- optimisation problem of finding a matrix on the Stiefel manifold that maximises a quadratic objective function.
We propose a simple yet effective sparsity-promoting algorithm for finding the dominant eigenspace matrix.
arXiv Detail & Related papers (2021-09-30T19:17:35Z) - Near-Optimal Algorithms for Linear Algebra in the Current Matrix
Multiplication Time [46.31710224483631]
We show how to bypass the main open question of Nelson and Nguyen (FOCS, 2013) regarding the logarithmic factors in the sketching dimension for existing constant factor approximation oblivious subspace embeddings.
A key technique we use is an explicit mapping of Indyk based on uncertainty principles and extractors.
For the fundamental problems of rank computation and finding a linearly independent subset of columns, our algorithms improve Cheung, Kwok, and Lau (JACM, 2013) and are optimal to within a constant factor and a $loglog(n)$-factor, respectively.
arXiv Detail & Related papers (2021-07-16T19:34:10Z) - Joint Network Topology Inference via Structured Fusion Regularization [70.30364652829164]
Joint network topology inference represents a canonical problem of learning multiple graph Laplacian matrices from heterogeneous graph signals.
We propose a general graph estimator based on a novel structured fusion regularization.
We show that the proposed graph estimator enjoys both high computational efficiency and rigorous theoretical guarantee.
arXiv Detail & Related papers (2021-03-05T04:42:32Z) - Optimal Iterative Sketching with the Subsampled Randomized Hadamard
Transform [64.90148466525754]
We study the performance of iterative sketching for least-squares problems.
We show that the convergence rate for Haar and randomized Hadamard matrices are identical, andally improve upon random projections.
These techniques may be applied to other algorithms that employ randomized dimension reduction.
arXiv Detail & Related papers (2020-02-03T16:17:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.