Bayesian Kernelised Test of (In)dependence with Mixed-type Variables
- URL: http://arxiv.org/abs/2105.04001v1
- Date: Sun, 9 May 2021 19:21:43 GMT
- Title: Bayesian Kernelised Test of (In)dependence with Mixed-type Variables
- Authors: Alessio Benavoli and Cassio de Campos
- Abstract summary: A fundamental task in AI is to assess (in)dependence between mixed-type variables (text, image, sound)
We propose a Bayesian kernelised correlation test of (in)dependence using a Dirichlet process model.
We show the properties of the approach, as well as algorithms for fast computation with it.
- Score: 1.2691047660244332
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A fundamental task in AI is to assess (in)dependence between mixed-type
variables (text, image, sound). We propose a Bayesian kernelised correlation
test of (in)dependence using a Dirichlet process model. The new measure of
(in)dependence allows us to answer some fundamental questions: Based on data,
are (mixed-type) variables independent? How likely is dependence/independence
to hold? How high is the probability that two mixed-type variables are more
than just weakly dependent? We theoretically show the properties of the
approach, as well as algorithms for fast computation with it. We empirically
demonstrate the effectiveness of the proposed method by analysing its
performance and by comparing it with other frequentist and Bayesian approaches
on a range of datasets and tasks with mixed-type variables.
Related papers
- Estimation of mutual information via quantum kernel method [0.0]
Estimating mutual information (MI) plays a critical role to investigate the relationship among multiple random variables with a nonlinear correlation.
We propose a method for estimating mutual information using the quantum kernel.
arXiv Detail & Related papers (2023-10-19T00:53:16Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Counterfactual Invariance to Spurious Correlations: Why and How to Pass
Stress Tests [87.60900567941428]
A spurious correlation' is the dependence of a model on some aspect of the input data that an analyst thinks shouldn't matter.
In machine learning, these have a know-it-when-you-see-it character.
We study stress testing using the tools of causal inference.
arXiv Detail & Related papers (2021-05-31T14:39:38Z) - Disentangling Observed Causal Effects from Latent Confounders using
Method of Moments [67.27068846108047]
We provide guarantees on identifiability and learnability under mild assumptions.
We develop efficient algorithms based on coupled tensor decomposition with linear constraints to obtain scalable and guaranteed solutions.
arXiv Detail & Related papers (2021-01-17T07:48:45Z) - Inferring Causal Direction from Observational Data: A Complexity
Approach [0.3553493344868413]
We propose several criteria for distinguishing cause and effect in pairs of discrete or continuous random variables.
We demonstrate the accuracy of the criteria on synthetic data generated under a broad family of causal mechanisms and types of noise.
arXiv Detail & Related papers (2020-10-12T12:10:37Z) - A Nonparametric Test of Dependence Based on Ensemble of Decision Trees [0.0]
The proposed coefficient is a permutation-like statistic that quantifies how much the observed sample S_n : (X_i, Y_i), i = 1.
n is discriminable from the permutated sample S_nn : (X_i, Y_j), i, j = 1.
n, where the two variables are independent.
arXiv Detail & Related papers (2020-07-24T02:48:33Z) - Neural Methods for Point-wise Dependency Estimation [129.93860669802046]
We focus on estimating point-wise dependency (PD), which quantitatively measures how likely two outcomes co-occur.
We demonstrate the effectiveness of our approaches in 1) MI estimation, 2) self-supervised representation learning, and 3) cross-modal retrieval task.
arXiv Detail & Related papers (2020-06-09T23:26:15Z) - Stable Prediction via Leveraging Seed Variable [73.9770220107874]
Previous machine learning methods might exploit subtly spurious correlations in training data induced by non-causal variables for prediction.
We propose a conditional independence test based algorithm to separate causal variables with a seed variable as priori, and adopt them for stable prediction.
Our algorithm outperforms state-of-the-art methods for stable prediction.
arXiv Detail & Related papers (2020-06-09T06:56:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.