Scalable Sampling for Nonsymmetric Determinantal Point Processes
- URL: http://arxiv.org/abs/2201.08417v1
- Date: Thu, 20 Jan 2022 19:21:59 GMT
- Title: Scalable Sampling for Nonsymmetric Determinantal Point Processes
- Authors: Insu Han, Mike Gartrell, Jennifer Gillenwater, Elvis Dohmatob, Amin
Karbasi
- Abstract summary: A determinantal point process (DPP) on a collection of $M$ items is a model, parameterized by a symmetric kernel matrix.
Recent work shows that removing the kernel symmetry constraint, yielding nonsymmetric DPPs (NDPPs), can lead to significant predictive performance gains for machine learning applications.
There is only one known DPP sampling algorithm, based on Cholesky decomposition, that can directly apply to NDPPs as well.
We show that imposing certain structural constraints on the NDPP kernel enables us to bound the rejection rate in a way that depends only on the kernel rank.
- Score: 47.18450267217498
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A determinantal point process (DPP) on a collection of $M$ items is a model,
parameterized by a symmetric kernel matrix, that assigns a probability to every
subset of those items. Recent work shows that removing the kernel symmetry
constraint, yielding nonsymmetric DPPs (NDPPs), can lead to significant
predictive performance gains for machine learning applications. However,
existing work leaves open the question of scalable NDPP sampling. There is only
one known DPP sampling algorithm, based on Cholesky decomposition, that can
directly apply to NDPPs as well. Unfortunately, its runtime is cubic in $M$,
and thus does not scale to large item collections. In this work, we first note
that this algorithm can be transformed into a linear-time one for kernels with
low-rank structure. Furthermore, we develop a scalable sublinear-time rejection
sampling algorithm by constructing a novel proposal distribution. Additionally,
we show that imposing certain structural constraints on the NDPP kernel enables
us to bound the rejection rate in a way that depends only on the kernel rank.
In our experiments we compare the speed of all of these samplers for a variety
of real-world tasks.
Related papers
- Approximating a RUM from Distributions on k-Slates [88.32814292632675]
We find a generalization-time algorithm that finds the RUM that best approximates the given distribution on average.
Our theoretical result can also be made practical: we obtain a that is effective and scales to real-world datasets.
arXiv Detail & Related papers (2023-05-22T17:43:34Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv Detail & Related papers (2022-08-11T04:12:50Z) - Scalable MCMC Sampling for Nonsymmetric Determinantal Point Processes [45.40646054226403]
A determinantal point process (DPP) assigns a probability to every subset of $n$ items.
Recent work has studied an approximate chain Monte Carlo sampling algorithm for NDPPs restricted to size-$k$ subsets.
We develop a scalable MCMC sampling algorithm for $k$-NDPPs with low-rank kernels.
arXiv Detail & Related papers (2022-07-01T15:22:12Z) - Simple and Near-Optimal MAP Inference for Nonsymmetric DPPs [3.3504365823045044]
We study the problem of maximum a posteriori (MAP) inference for determinantal point processes defined by a nonsymmetric kernel matrix.
We obtain the first multiplicative approximation guarantee for this problem using local search.
Our approximation factor of $kO(k)$ is nearly tight, and we show theoretically and experimentally that it compares favorably to the state-of-the-art methods for this problem.
arXiv Detail & Related papers (2021-02-10T09:34:44Z) - Learning from DPPs via Sampling: Beyond HKPV and symmetry [2.0305676256390934]
We show how to approximate the distribution function of a linear statistic of a Determinantal point processes (DPP)
Our approach is scalable and applies to very general DPPs, beyond traditional symmetric kernels.
arXiv Detail & Related papers (2020-07-08T17:33:45Z) - Sampling from a $k$-DPP without looking at all items [58.30573872035083]
Given a kernel function and a subset size $k$, our goal is to sample $k$ out of $n$ items with probability proportional to the determinant of the kernel matrix induced by the subset (a.k.a. $k$-DPP)
Existing $k$-DPP sampling algorithms require an expensive preprocessing step which involves multiple passes over all $n$ items, making it infeasible for large datasets.
We develop an algorithm which adaptively builds a sufficiently large uniform sample of data that is then used to efficiently generate a smaller set of $k$ items.
arXiv Detail & Related papers (2020-06-30T16:40:44Z) - Scalable Learning and MAP Inference for Nonsymmetric Determinantal Point
Processes [29.56615511158156]
We develop a learning algorithm with space and time requirements linear in $M$ by introducing a new NDPP kernel decomposition.
We also derive a linear-complexity NDPP maximum a posteriori (MAP) inference algorithm that applies not only to our new kernel but also to that of prior work.
arXiv Detail & Related papers (2020-06-17T13:42:09Z) - Determinantal Point Processes in Randomized Numerical Linear Algebra [80.27102478796613]
Numerical Linear Algebra (RandNLA) uses randomness to develop improved algorithms for matrix problems that arise in scientific computing, data science, machine learning, etc.
Recent work has uncovered deep and fruitful connections between DPPs and RandNLA which lead to new guarantees and improved algorithms.
arXiv Detail & Related papers (2020-05-07T00:39:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.