Testing Determinantal Point Processes
- URL: http://arxiv.org/abs/2008.03650v1
- Date: Sun, 9 Aug 2020 04:45:16 GMT
- Title: Testing Determinantal Point Processes
- Authors: Khashayar Gatmiry (1), Maryam Aliakbarpour (1), Stefanie Jegelka (1)
((1) Massachusetts Institute of Technology)
- Abstract summary: We investigate DPPs from a new perspective: property testing of distributions.
We propose the first algorithm for testing DPPs.
We show a new hardness result for the problem of testing the more general class of log-submodular distributions.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Determinantal point processes (DPPs) are popular probabilistic models of
diversity. In this paper, we investigate DPPs from a new perspective: property
testing of distributions. Given sample access to an unknown distribution $q$
over the subsets of a ground set, we aim to distinguish whether $q$ is a DPP
distribution, or $\epsilon$-far from all DPP distributions in
$\ell_1$-distance. In this work, we propose the first algorithm for testing
DPPs. Furthermore, we establish a matching lower bound on the sample complexity
of DPP testing. This lower bound also extends to showing a new hardness result
for the problem of testing the more general class of log-submodular
distributions.
Related papers
- Stochastic Approximation Approaches to Group Distributionally Robust Optimization and Beyond [89.72693227960274]
This paper investigates group distributionally robust optimization (GDRO) with the goal of learning a model that performs well over $m$ different distributions.
To reduce the number of samples in each round from $m$ to 1, we cast GDRO as a two-player game, where one player conducts and the other executes an online algorithm for non-oblivious multi-armed bandits.
In the second scenario, we propose to optimize the average top-$k$ risk instead of the maximum risk, thereby mitigating the impact of distributions.
arXiv Detail & Related papers (2023-02-18T09:24:15Z) - Differentially-Private Bayes Consistency [70.92545332158217]
We construct a Bayes consistent learning rule that satisfies differential privacy (DP)
We prove that any VC class can be privately learned in a semi-supervised setting with a near-optimal sample complexity.
arXiv Detail & Related papers (2022-12-08T11:57:30Z) - A Faster Sampler for Discrete Determinantal Point Processes [10.355938901584567]
Discrete Determinantal Point Processes (DPPs) have a wide array of potential applications for subsampling datasets.
In the worst-case scenario, the sampling cost scales as $O(n3)$ where n is the number of elements of the ground set.
A popular workaround to this prohibitive cost is to sample DPPs defined by low-rank kernels.
arXiv Detail & Related papers (2022-10-31T14:37:54Z) - Scalable MCMC Sampling for Nonsymmetric Determinantal Point Processes [45.40646054226403]
A determinantal point process (DPP) assigns a probability to every subset of $n$ items.
Recent work has studied an approximate chain Monte Carlo sampling algorithm for NDPPs restricted to size-$k$ subsets.
We develop a scalable MCMC sampling algorithm for $k$-NDPPs with low-rank kernels.
arXiv Detail & Related papers (2022-07-01T15:22:12Z) - Scalable Sampling for Nonsymmetric Determinantal Point Processes [47.18450267217498]
A determinantal point process (DPP) on a collection of $M$ items is a model, parameterized by a symmetric kernel matrix.
Recent work shows that removing the kernel symmetry constraint, yielding nonsymmetric DPPs (NDPPs), can lead to significant predictive performance gains for machine learning applications.
There is only one known DPP sampling algorithm, based on Cholesky decomposition, that can directly apply to NDPPs as well.
We show that imposing certain structural constraints on the NDPP kernel enables us to bound the rejection rate in a way that depends only on the kernel rank.
arXiv Detail & Related papers (2022-01-20T19:21:59Z) - Robust Learning of Optimal Auctions [84.13356290199603]
We study the problem of learning revenue-optimal multi-bidder auctions from samples when the samples of bidders' valuations can be adversarially corrupted or drawn from distributions that are adversarially perturbed.
We propose new algorithms that can learn a mechanism whose revenue is nearly optimal simultaneously for all true distributions'' that are $alpha$-close to the original distribution in Kolmogorov-Smirnov distance.
arXiv Detail & Related papers (2021-07-13T17:37:21Z) - Optimal Testing of Discrete Distributions with High Probability [49.19942805582874]
We study the problem of testing discrete distributions with a focus on the high probability regime.
We provide the first algorithms for closeness and independence testing that are sample-optimal, within constant factors.
arXiv Detail & Related papers (2020-09-14T16:09:17Z) - Learning from DPPs via Sampling: Beyond HKPV and symmetry [2.0305676256390934]
We show how to approximate the distribution function of a linear statistic of a Determinantal point processes (DPP)
Our approach is scalable and applies to very general DPPs, beyond traditional symmetric kernels.
arXiv Detail & Related papers (2020-07-08T17:33:45Z) - Sampling from a $k$-DPP without looking at all items [58.30573872035083]
Given a kernel function and a subset size $k$, our goal is to sample $k$ out of $n$ items with probability proportional to the determinant of the kernel matrix induced by the subset (a.k.a. $k$-DPP)
Existing $k$-DPP sampling algorithms require an expensive preprocessing step which involves multiple passes over all $n$ items, making it infeasible for large datasets.
We develop an algorithm which adaptively builds a sufficiently large uniform sample of data that is then used to efficiently generate a smaller set of $k$ items.
arXiv Detail & Related papers (2020-06-30T16:40:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.