Scalable MCMC Sampling for Nonsymmetric Determinantal Point Processes
- URL: http://arxiv.org/abs/2207.00486v1
- Date: Fri, 1 Jul 2022 15:22:12 GMT
- Title: Scalable MCMC Sampling for Nonsymmetric Determinantal Point Processes
- Authors: Insu Han, Mike Gartrell, Elvis Dohmatob, Amin Karbasi
- Abstract summary: A determinantal point process (DPP) assigns a probability to every subset of $n$ items.
Recent work has studied an approximate chain Monte Carlo sampling algorithm for NDPPs restricted to size-$k$ subsets.
We develop a scalable MCMC sampling algorithm for $k$-NDPPs with low-rank kernels.
- Score: 45.40646054226403
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A determinantal point process (DPP) is an elegant model that assigns a
probability to every subset of a collection of $n$ items. While conventionally
a DPP is parameterized by a symmetric kernel matrix, removing this symmetry
constraint, resulting in nonsymmetric DPPs (NDPPs), leads to significant
improvements in modeling power and predictive performance. Recent work has
studied an approximate Markov chain Monte Carlo (MCMC) sampling algorithm for
NDPPs restricted to size-$k$ subsets (called $k$-NDPPs). However, the runtime
of this approach is quadratic in $n$, making it infeasible for large-scale
settings. In this work, we develop a scalable MCMC sampling algorithm for
$k$-NDPPs with low-rank kernels, thus enabling runtime that is sublinear in
$n$. Our method is based on a state-of-the-art NDPP rejection sampling
algorithm, which we enhance with a novel approach for efficiently constructing
the proposal distribution. Furthermore, we extend our scalable $k$-NDPP
sampling algorithm to NDPPs without size constraints. Our resulting sampling
method has polynomial time complexity in the rank of the kernel, while the
existing approach has runtime that is exponential in the rank. With both a
theoretical analysis and experiments on real-world datasets, we verify that our
scalable approximate sampling algorithms are orders of magnitude faster than
existing sampling approaches for $k$-NDPPs and NDPPs.
Related papers
- Reward-Free RL is No Harder Than Reward-Aware RL in Linear Markov
Decision Processes [61.11090361892306]
Reward-free reinforcement learning (RL) considers the setting where the agent does not have access to a reward function during exploration.
We show that this separation does not exist in the setting of linear MDPs.
We develop a computationally efficient algorithm for reward-free RL in a $d$-dimensional linear MDP.
arXiv Detail & Related papers (2022-01-26T22:09:59Z) - Scalable Sampling for Nonsymmetric Determinantal Point Processes [47.18450267217498]
A determinantal point process (DPP) on a collection of $M$ items is a model, parameterized by a symmetric kernel matrix.
Recent work shows that removing the kernel symmetry constraint, yielding nonsymmetric DPPs (NDPPs), can lead to significant predictive performance gains for machine learning applications.
There is only one known DPP sampling algorithm, based on Cholesky decomposition, that can directly apply to NDPPs as well.
We show that imposing certain structural constraints on the NDPP kernel enables us to bound the rejection rate in a way that depends only on the kernel rank.
arXiv Detail & Related papers (2022-01-20T19:21:59Z) - Towards Tight Bounds on the Sample Complexity of Average-reward MDPs [39.01663172393174]
We find an optimal policy of an infinite-horizon average-reward Markov decision process given access to a generative model.
We provide an algorithm that solves the problem using $widetildeO(t_mathrmmix epsilon-3)$ (oblivious) samples per state-action pair.
arXiv Detail & Related papers (2021-06-13T17:18:11Z) - Simple and Near-Optimal MAP Inference for Nonsymmetric DPPs [3.3504365823045044]
We study the problem of maximum a posteriori (MAP) inference for determinantal point processes defined by a nonsymmetric kernel matrix.
We obtain the first multiplicative approximation guarantee for this problem using local search.
Our approximation factor of $kO(k)$ is nearly tight, and we show theoretically and experimentally that it compares favorably to the state-of-the-art methods for this problem.
arXiv Detail & Related papers (2021-02-10T09:34:44Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - Sampling from a $k$-DPP without looking at all items [58.30573872035083]
Given a kernel function and a subset size $k$, our goal is to sample $k$ out of $n$ items with probability proportional to the determinant of the kernel matrix induced by the subset (a.k.a. $k$-DPP)
Existing $k$-DPP sampling algorithms require an expensive preprocessing step which involves multiple passes over all $n$ items, making it infeasible for large datasets.
We develop an algorithm which adaptively builds a sufficiently large uniform sample of data that is then used to efficiently generate a smaller set of $k$ items.
arXiv Detail & Related papers (2020-06-30T16:40:44Z) - Scalable Learning and MAP Inference for Nonsymmetric Determinantal Point
Processes [29.56615511158156]
We develop a learning algorithm with space and time requirements linear in $M$ by introducing a new NDPP kernel decomposition.
We also derive a linear-complexity NDPP maximum a posteriori (MAP) inference algorithm that applies not only to our new kernel but also to that of prior work.
arXiv Detail & Related papers (2020-06-17T13:42:09Z) - Determinantal Point Processes in Randomized Numerical Linear Algebra [80.27102478796613]
Numerical Linear Algebra (RandNLA) uses randomness to develop improved algorithms for matrix problems that arise in scientific computing, data science, machine learning, etc.
Recent work has uncovered deep and fruitful connections between DPPs and RandNLA which lead to new guarantees and improved algorithms.
arXiv Detail & Related papers (2020-05-07T00:39:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.