Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data
- URL: http://arxiv.org/abs/2106.07636v1
- Date: Mon, 14 Jun 2021 17:52:50 GMT
- Title: Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data
- Authors: Feng Liu and Wenkai Xu and Jie Lu and Danica J. Sutherland
- Abstract summary: We introduce the problem of meta two-sample testing (M2ST)
M2ST aims to exploit (abundant) auxiliary data on related tasks to find an algorithm that can quickly identify a powerful test on new target tasks.
We provide both theoretical justification and empirical evidence that our proposed meta-testing schemes out-perform learning kernel-based tests directly from scarce observations.
- Score: 21.596650236820377
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern kernel-based two-sample tests have shown great success in
distinguishing complex, high-dimensional distributions with appropriate learned
kernels. Previous work has demonstrated that this kernel learning procedure
succeeds, assuming a considerable number of observed samples from each
distribution. In realistic scenarios with very limited numbers of data samples,
however, it can be challenging to identify a kernel powerful enough to
distinguish complex distributions. We address this issue by introducing the
problem of meta two-sample testing (M2ST), which aims to exploit (abundant)
auxiliary data on related tasks to find an algorithm that can quickly identify
a powerful test on new target tasks. We propose two specific algorithms for
this task: a generic scheme which improves over baselines and amore tailored
approach which performs even better. We provide both theoretical justification
and empirical evidence that our proposed meta-testing schemes out-perform
learning kernel-based tests directly from scarce observations, and identify
when such schemes will be successful.
Related papers
- A Weighted K-Center Algorithm for Data Subset Selection [70.49696246526199]
Subset selection is a fundamental problem that can play a key role in identifying smaller portions of the training data.
We develop a novel factor 3-approximation algorithm to compute subsets based on the weighted sum of both k-center and uncertainty sampling objective functions.
arXiv Detail & Related papers (2023-12-17T04:41:07Z) - Multimodal Multi-User Surface Recognition with the Kernel Two-Sample
Test [15.051737123188174]
We propose a framework that can handle heterogeneous data sources for classification tasks.
Our data-versus-data approach automatically quantifies distinctive differences in distributions in a high-dimensional space.
We achieve 97.2% accuracy on a standard multi-user dataset with 108 surface classes, outperforming the state-of-the-art machine-learning algorithm by 6% on a more difficult version of the task.
arXiv Detail & Related papers (2023-03-08T22:58:55Z) - Boosting the Power of Kernel Two-Sample Tests [4.07125466598411]
A kernel two-sample test based on the maximum mean discrepancy (MMD) is one of the most popular methods for detecting differences between two distributions over general metric spaces.
We propose a method to boost the power of the kernel test by combining MMD estimates over multiple kernels using their Mahalanobis distance.
arXiv Detail & Related papers (2023-02-21T14:14:30Z) - Multi-task Representation Learning for Pure Exploration in Linear
Bandits [34.67303292713379]
We study multi-task representation learning for best arm identification in linear bandits (RepBAI-LB) and best policy identification in contextual linear bandits (RepBPI-CLB)
In these two problems, all tasks share a common low-dimensional linear representation, and our goal is to leverage this feature to accelerate the best arm (policy) identification process for all tasks.
We show that by learning the common representation among tasks, our sample complexity is significantly better than that of the native approach which solves tasks independently.
arXiv Detail & Related papers (2023-02-09T05:14:48Z) - A Moment-Matching Approach to Testable Learning and a New
Characterization of Rademacher Complexity [15.746613321517282]
We give a powerful new approach for developing algorithms for testable learning using tools from moment matching and metric agnostic in probability.
Surprisingly, we show that the information-theoretic complexity of testable learning is tightly characterized by the Rademacher complexity of the concept class.
arXiv Detail & Related papers (2022-11-23T21:29:51Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - MetaKernel: Learning Variational Random Features with Limited Labels [120.90737681252594]
Few-shot learning deals with the fundamental and challenging problem of learning from a few annotated samples, while being able to generalize well on new tasks.
We propose meta-learning kernels with random Fourier features for few-shot learning, we call Meta Kernel.
arXiv Detail & Related papers (2021-05-08T21:24:09Z) - Distribution Alignment: A Unified Framework for Long-tail Visual
Recognition [52.36728157779307]
We propose a unified distribution alignment strategy for long-tail visual recognition.
We then introduce a generalized re-weight method in the two-stage learning to balance the class prior.
Our approach achieves the state-of-the-art results across all four recognition tasks with a simple and unified framework.
arXiv Detail & Related papers (2021-03-30T14:09:53Z) - Expert Training: Task Hardness Aware Meta-Learning for Few-Shot
Classification [62.10696018098057]
We propose an easy-to-hard expert meta-training strategy to arrange the training tasks properly.
A task hardness aware module is designed and integrated into the training procedure to estimate the hardness of a task.
Experimental results on the miniImageNet and tieredImageNetSketch datasets show that the meta-learners can obtain better results with our expert training strategy.
arXiv Detail & Related papers (2020-07-13T08:49:00Z) - Learning Deep Kernels for Non-Parametric Two-Sample Tests [50.92621794426821]
We propose a class of kernel-based two-sample tests, which aim to determine whether two sets of samples are drawn from the same distribution.
Our tests are constructed from kernels parameterized by deep neural nets, trained to maximize test power.
arXiv Detail & Related papers (2020-02-21T03:54:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.