Learning Distributional Programs for Relational Autocompletion
- URL: http://arxiv.org/abs/2001.08603v5
- Date: Mon, 5 Jul 2021 14:35:00 GMT
- Title: Learning Distributional Programs for Relational Autocompletion
- Authors: Kumar Nitesh, Kuzelka Ondrej and De Raedt Luc
- Abstract summary: We introduce DiceML an approach to learn both the structure and the parameters of programs from relational data (with possibly missing data)
DiceML integrates statistical modeling and distributional clauses with rule learning.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Relational autocompletion is the problem of automatically filling out some
missing values in multi-relational data. We tackle this problem within the
probabilistic logic programming framework of Distributional Clauses (DC), which
supports both discrete and continuous probability distributions. Within this
framework, we introduce DiceML { an approach to learn both the structure and
the parameters of DC programs from relational data (with possibly missing
data). To realize this, DiceML integrates statistical modeling and
distributional clauses with rule learning. The distinguishing features of
DiceML are that it 1) tackles autocompletion in relational data, 2) learns
distributional clauses extended with statistical models, 3) deals with both
discrete and continuous distributions, 4) can exploit background knowledge, and
5) uses an expectation-maximization based algorithm to cope with missing data.
The empirical results show the promise of the approach, even when there is
missing data.
Related papers
- Dataset Condensation with Latent Quantile Matching [5.466962214217334]
Current distribution matching (DM) based DC methods learn a synthesized dataset by matching the mean of the latent embeddings between the synthetic and the real outliers.
We propose Latent Quantile Matching (LQM) which matches the quantiles of the latent embeddings to minimize the goodness of fit test statistic between two distributions.
arXiv Detail & Related papers (2024-06-14T09:20:44Z) - Universal Metric Learning with Parameter-Efficient Transfer Learning [40.85295050164728]
A common practice in metric learning is to train and test an embedding model for each dataset.
This dataset-specific approach fails to simulate real-world scenarios that involve multiple heterogeneous distributions of data.
We introduce a novel metric learning paradigm, called Universal Metric Learning (UML), which learns a unified metric capable of capturing relations across multiple data distributions.
arXiv Detail & Related papers (2023-09-16T10:34:01Z) - SUMMIT: Source-Free Adaptation of Uni-Modal Models to Multi-Modal
Targets [30.262094419776208]
Current approaches assume that the source data is available during adaptation and that the source consists of paired multi-modal data.
We propose a switching framework which automatically chooses between two complementary methods of cross-modal pseudo-label fusion.
Our method achieves an improvement in mIoU of up to 12% over competing baselines.
arXiv Detail & Related papers (2023-08-23T02:57:58Z) - Learning the joint distribution of two sequences using little or no
paired data [16.189575655434844]
We present a noisy channel generative model of two sequences, for example text and speech.
We show that even tiny amount of paired data is sufficient to learn to relate the two modalities when a massive amount of unpaired data is available.
arXiv Detail & Related papers (2022-12-06T18:56:15Z) - Learning to Bound Counterfactual Inference in Structural Causal Models
from Observational and Randomised Data [64.96984404868411]
We derive a likelihood characterisation for the overall data that leads us to extend a previous EM-based algorithm.
The new algorithm learns to approximate the (unidentifiability) region of model parameters from such mixed data sources.
It delivers interval approximations to counterfactual results, which collapse to points in the identifiable case.
arXiv Detail & Related papers (2022-12-06T12:42:11Z) - MIRACLE: Causally-Aware Imputation via Learning Missing Data Mechanisms [82.90843777097606]
We propose a causally-aware imputation algorithm (MIRACLE) for missing data.
MIRACLE iteratively refines the imputation of a baseline by simultaneously modeling the missingness generating mechanism.
We conduct extensive experiments on synthetic and a variety of publicly available datasets to show that MIRACLE is able to consistently improve imputation.
arXiv Detail & Related papers (2021-11-04T22:38:18Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - Lightweight Data Fusion with Conjugate Mappings [11.760099863897835]
We present an approach to data fusion that combines the interpretability of structured probabilistic graphical models with the flexibility of neural networks.
The proposed method, lightweight data fusion (LDF), emphasizes posterior analysis over latent variables using two types of information.
arXiv Detail & Related papers (2020-11-20T19:47:13Z) - Autoregressive Score Matching [113.4502004812927]
We propose autoregressive conditional score models (AR-CSM) where we parameterize the joint distribution in terms of the derivatives of univariable log-conditionals (scores)
For AR-CSM models, this divergence between data and model distributions can be computed and optimized efficiently, requiring no expensive sampling or adversarial training.
We show with extensive experimental results that it can be applied to density estimation on synthetic data, image generation, image denoising, and training latent variable models with implicit encoders.
arXiv Detail & Related papers (2020-10-24T07:01:24Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.