Faster Maximum Inner Product Search in High Dimensions
- URL: http://arxiv.org/abs/2212.07551v3
- Date: Tue, 27 Jun 2023 03:36:30 GMT
- Title: Faster Maximum Inner Product Search in High Dimensions
- Authors: Mo Tiwari, Ryan Kang, Je-Yong Lee, Donghyun Lee, Chris Piech,
Sebastian Thrun, Ilan Shomorony, Martin Jinye Zhang
- Abstract summary: Maximum Inner Product Search (MIPS) is a ubiquitous task in machine learning applications such as recommendation systems.
We present BanditMIPS, a novel randomized MIPS algorithm whose complexity is independent of $d$.
We provide theoretical guarantees that BanditMIPS returns the correct answer with high probability, while improving the complexity in $d$ from $O(sqrtd)$ to $O(1)$.
- Score: 17.040520467777295
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Maximum Inner Product Search (MIPS) is a ubiquitous task in machine learning
applications such as recommendation systems. Given a query vector and $n$ atom
vectors in $d$-dimensional space, the goal of MIPS is to find the atom that has
the highest inner product with the query vector. Existing MIPS algorithms scale
at least as $O(\sqrt{d})$, which becomes computationally prohibitive in
high-dimensional settings. In this work, we present BanditMIPS, a novel
randomized MIPS algorithm whose complexity is independent of $d$. BanditMIPS
estimates the inner product for each atom by subsampling coordinates and
adaptively evaluates more coordinates for more promising atoms. The specific
adaptive sampling strategy is motivated by multi-armed bandits. We provide
theoretical guarantees that BanditMIPS returns the correct answer with high
probability, while improving the complexity in $d$ from $O(\sqrt{d})$ to
$O(1)$. We also perform experiments on four synthetic and real-world datasets
and demonstrate that BanditMIPS outperforms prior state-of-the-art algorithms.
For example, in the Movie Lens dataset ($n$=4,000, $d$=6,000), BanditMIPS is
20$\times$ faster than the next best algorithm while returning the same answer.
BanditMIPS requires no preprocessing of the data and includes a hyperparameter
that practitioners may use to trade off accuracy and runtime. We also propose a
variant of our algorithm, named BanditMIPS-$\alpha$, which achieves further
speedups by employing non-uniform sampling across coordinates. Finally, we
demonstrate how known preprocessing techniques can be used to further
accelerate BanditMIPS, and discuss applications to Matching Pursuit and Fourier
analysis.
Related papers
- Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - On the Sample Complexity of Representation Learning in Multi-task
Bandits with Global and Local structure [77.60508571062958]
We investigate the sample complexity of learning the optimal arm for multi-task bandit problems.
Arms consist of two components: one that is shared across tasks (that we call representation) and one that is task-specific (that we call predictor)
We devise an algorithm OSRL-SC whose sample complexity approaches the lower bound, and scales at most as $H(Glog(delta_G)+ Xlog(delta_H))$, with $X,G,H$ being, respectively, the number of tasks, representations and predictors.
arXiv Detail & Related papers (2022-11-28T08:40:12Z) - Scalable MCMC Sampling for Nonsymmetric Determinantal Point Processes [45.40646054226403]
A determinantal point process (DPP) assigns a probability to every subset of $n$ items.
Recent work has studied an approximate chain Monte Carlo sampling algorithm for NDPPs restricted to size-$k$ subsets.
We develop a scalable MCMC sampling algorithm for $k$-NDPPs with low-rank kernels.
arXiv Detail & Related papers (2022-07-01T15:22:12Z) - Matching Pursuit Based Scheduling for Over-the-Air Federated Learning [67.59503935237676]
This paper develops a class of low-complexity device scheduling algorithms for over-the-air learning via the method of federated learning.
Compared to the state-of-the-art proposed scheme, the proposed scheme poses a drastically lower efficiency system.
The efficiency of the proposed scheme is confirmed via experiments on the CIFAR dataset.
arXiv Detail & Related papers (2022-06-14T08:14:14Z) - Kernel Packet: An Exact and Scalable Algorithm for Gaussian Process
Regression with Mat\'ern Correlations [23.560067934682294]
We develop an exact and scalable algorithm for one-dimensional Gaussian process regression with Mat'ern correlations.
The proposed algorithm is significantly superior to the existing alternatives in both the computational time and predictive accuracy.
arXiv Detail & Related papers (2022-03-07T03:30:35Z) - Quantum Algorithms for Reinforcement Learning with a Generative Model [16.168901236223117]
Reinforcement learning studies how an agent should interact with an environment to maximize its cumulative reward.
We design quantum algorithms that approximate an optimal policy ($pi*$), the optimal value function ($v*$), and the optimal $Q$-function ($q*$)
We show that our quantum algorithm for computing $q*$ is optimal by proving a matching quantum lower bound.
arXiv Detail & Related papers (2021-12-15T19:51:49Z) - Linear Bandit Algorithms with Sublinear Time Complexity [67.21046514005029]
We propose to accelerate existing linear bandit algorithms to achieve per-step time complexity sublinear in the number of arms $K$.
We show that our proposed algorithms can achieve $O(K1-alpha(T))$ per-step complexity for some $alpha(T) > 0$ and $widetilde O(stT)$ regret, where $T$ is the time horizon.
arXiv Detail & Related papers (2021-03-03T22:42:15Z) - Hutch++: Optimal Stochastic Trace Estimation [75.45968495410048]
We introduce a new randomized algorithm, Hutch++, which computes a $(1 pm epsilon)$ approximation to $tr(A)$ for any positive semidefinite (PSD) $A$.
We show that it significantly outperforms Hutchinson's method in experiments.
arXiv Detail & Related papers (2020-10-19T16:45:37Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - BanditPAM: Almost Linear Time $k$-Medoids Clustering via Multi-Armed
Bandits [16.1767275655842]
Current $k$-medoids clustering algorithms, such as Partitioning Around Medoids (PAM), are iterative and are in the dataset size $n$ for each iteration, being prohibitively expensive for large datasets.
We propose BanditPAM, a randomized algorithm inspired by techniques from multi-armed bandits, that reduces the complexity of each PAM iteration from $O(n2)$ to $O(n log n)$ and returns the same results with high probability, under assumptions on the data that often hold in practice.
We empirically validate our results on several large real-world datasets, including a coding
arXiv Detail & Related papers (2020-06-11T22:17:16Z) - Low Complexity Sequential Search with Size-Dependent Measurement Noise [19.201555109949677]
This paper considers a target localization problem where at any given time an agent can choose a region to query for the presence of the target in that region.
Motivated by practical applications such as initial beam alignment in array processing, heavy hitter detection in networking, and visual search in robotics, we consider practically important complexity constraints/metrics.
Two novel search strategy, $dyaPM$ and $hiePM$, are proposed.
arXiv Detail & Related papers (2020-05-15T22:40:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.