A New Perspective on Learning Context-Specific Independence
- URL: http://arxiv.org/abs/2006.06896v1
- Date: Fri, 12 Jun 2020 01:11:02 GMT
- Title: A New Perspective on Learning Context-Specific Independence
- Authors: Yujia Shen, Arthur Choi, Adnan Darwiche
- Abstract summary: Local structure such as context-specific independence (CSI) has received much attention in the probabilistic graphical model (PGM) literature.
In this paper, we provide a new perspective on how to learn CSIs from data.
- Score: 18.273290530700567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Local structure such as context-specific independence (CSI) has received much
attention in the probabilistic graphical model (PGM) literature, as it
facilitates the modeling of large complex systems, as well as for reasoning
with them. In this paper, we provide a new perspective on how to learn CSIs
from data. We propose to first learn a functional and parameterized
representation of a conditional probability table (CPT), such as a neural
network. Next, we quantize this continuous function, into an arithmetic circuit
representation that facilitates efficient inference. In the first step, we can
leverage the many powerful tools that have been developed in the machine
learning literature. In the second step, we exploit more recently-developed
analytic tools from explainable AI, for the purposes of learning CSIs. Finally,
we contrast our approach, empirically and conceptually, with more traditional
variable-splitting approaches, that search for CSIs more explicitly.
Related papers
- Using Machine Learning to Discover Parsimonious and Physically-Interpretable Representations of Catchment-Scale Rainfall-Runoff Dynamics [1.1510009152620668]
An underexplored aspect of machine learning is how to develop minimally-optimal representations.
Our own view is that ML-based modeling should be based in use of computational units that are fundamentally interpretable by design.
We show, in the context of lumped catchment modeling, that physical interpretability and predictive performance can both be achieved using a relatively parsimonious distributed-state network.
arXiv Detail & Related papers (2024-12-06T08:30:01Z) - Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting & Beyond [61.18736646013446]
In pursuit of a deeper understanding of its surprising behaviors, we investigate the utility of a simple yet accurate model of a trained neural network.
Across three case studies, we illustrate how it can be applied to derive new empirical insights on a diverse range of prominent phenomena.
arXiv Detail & Related papers (2024-10-31T22:54:34Z) - Coding for Intelligence from the Perspective of Category [66.14012258680992]
Coding targets compressing and reconstructing data, and intelligence.
Recent trends demonstrate the potential homogeneity of these two fields.
We propose a novel problem of Coding for Intelligence from the category theory view.
arXiv Detail & Related papers (2024-07-01T07:05:44Z) - Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment [10.814585613336778]
Causal representation learning aims to combine the core strengths of machine learning and causality.
This thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations.
arXiv Detail & Related papers (2024-06-19T09:14:40Z) - A Survey to Recent Progress Towards Understanding In-Context Learning [37.933016939520684]
In-Context Learning (ICL) empowers Large Language Models (LLMs) with the ability to learn from a few examples provided in the prompt.
Despite encouragingly empirical success, the underlying mechanism of ICL remains unclear.
arXiv Detail & Related papers (2024-02-03T17:13:03Z) - A Novel Neural-symbolic System under Statistical Relational Learning [50.747658038910565]
We propose a general bi-level probabilistic graphical reasoning framework called GBPGR.
In GBPGR, the results of symbolic reasoning are utilized to refine and correct the predictions made by the deep learning models.
Our approach achieves high performance and exhibits effective generalization in both transductive and inductive tasks.
arXiv Detail & Related papers (2023-09-16T09:15:37Z) - ShadowNet for Data-Centric Quantum System Learning [188.683909185536]
We propose a data-centric learning paradigm combining the strength of neural-network protocols and classical shadows.
Capitalizing on the generalization power of neural networks, this paradigm can be trained offline and excel at predicting previously unseen systems.
We present the instantiation of our paradigm in quantum state tomography and direct fidelity estimation tasks and conduct numerical analysis up to 60 qubits.
arXiv Detail & Related papers (2023-08-22T09:11:53Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - AI Centered on Scene Fitting and Dynamic Cognitive Network [4.228224431041357]
This paper briefly analyzes the advantages and problems of AI mainstream technology and puts forward: To achieve stronger Artificial Intelligence, the end-to-end function calculation must be changed.
It also discusses the concrete scheme named Dynamic Cognitive Network model (DC Net)
arXiv Detail & Related papers (2020-10-02T06:13:41Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.