Ising Models with Hidden Markov Structure: Applications to Probabilistic Inference in Machine Learning
- URL: http://arxiv.org/abs/2504.13927v2
- Date: Sat, 14 Jun 2025 11:57:17 GMT
- Title: Ising Models with Hidden Markov Structure: Applications to Probabilistic Inference in Machine Learning
- Authors: F. Herrera, U. A. Rozikov, M. V. Velasco,
- Abstract summary: We investigate tree-indexed Markov chains (Gibbs measures) defined by a Hamiltonian that couples two Ising layers.<n>Specifically, we explore translation-invariant Gibbs measures (TIGM) of this Hamiltonian on Cayley trees.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we investigate tree-indexed Markov chains (Gibbs measures) defined by a Hamiltonian that couples two Ising layers: hidden spins \(s(x) \in \{\pm 1\}\) and observed spins \(\sigma(x) \in \{\pm 1\}\) on a Cayley tree. The Hamiltonian incorporates Ising interactions within each layer and site-wise emission couplings between layers, extending hidden Markov models to a bilayer Markov random field. Specifically, we explore translation-invariant Gibbs measures (TIGM) of this Hamiltonian on Cayley trees. Under certain explicit conditions on the model's parameters, we demonstrate that there can be up to three distinct TIGMs. Each of these measures represents an equilibrium state of the spin system. These measures provide a structured approach to inference on hierarchical data in machine learning. They have practical applications in tasks such as denoising, weakly supervised learning, and anomaly detection. The Cayley tree structure is particularly advantageous for exact inference due to its tractability.
Related papers
- Variational phylogenetic inference with products over bipartitions [48.2982114295171]
We present a novel variational family based on coalescent times of a single-linkage clustering and derive a closed-form density of the resulting distribution over trees.<n>Our method performs inference over all of tree space, it does not require any Markov chain Monte Carlo subroutines, and our variational family is differentiable.
arXiv Detail & Related papers (2025-02-21T00:06:57Z) - Inferring Kernel $ε$-Machines: Discovering Structure in Complex Systems [49.1574468325115]
We introduce causal diffusion components that encode the kernel causal-state estimates as a set of coordinates in a reduced dimension space.
We show how each component extracts predictive features from data and demonstrate their application on four examples.
arXiv Detail & Related papers (2024-10-01T21:14:06Z) - Disentangled Representation Learning with the Gromov-Monge Gap [65.73194652234848]
Learning disentangled representations from unlabelled data is a fundamental challenge in machine learning.
We introduce a novel approach to disentangled representation learning based on quadratic optimal transport.
We demonstrate the effectiveness of our approach for quantifying disentanglement across four standard benchmarks.
arXiv Detail & Related papers (2024-07-10T16:51:32Z) - A Critical Assessment of Interpretable and Explainable Machine Learning for Intrusion Detection [0.0]
We study the use of overly complex and opaque ML models, unaccounted data imbalances and correlated features, inconsistent influential features across different explanation methods, and the implausible utility of explanations.
Specifically, we advise avoiding complex opaque models such as Deep Neural Networks and instead using interpretable ML models such as Decision Trees.
We find that feature-based model explanations are most often inconsistent across different settings.
arXiv Detail & Related papers (2024-07-04T15:35:42Z) - Bayesian Circular Regression with von Mises Quasi-Processes [57.88921637944379]
In this work we explore a family of expressive and interpretable distributions over circle-valued random functions.<n>For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Gibbs sampling.<n>We present experiments applying this model to the prediction of wind directions and the percentage of the running gait cycle as a function of joint angles.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - CWF: Consolidating Weak Features in High-quality Mesh Simplification [50.634070540791555]
We propose a smooth functional that simultaneously considers all of these requirements.
The functional comprises a normal anisotropy term and a Centroidal Voronoi Tessellation (CVT) energy term.
arXiv Detail & Related papers (2024-04-24T05:37:17Z) - Gaussian Mixture Models for Affordance Learning using Bayesian Networks [50.18477618198277]
Affordances are fundamental descriptors of relationships between actions, objects and effects.
This paper approaches the problem of an embodied agent exploring the world and learning these affordances autonomously from its sensory experiences.
arXiv Detail & Related papers (2024-02-08T22:05:45Z) - Community Detection in the Multi-View Stochastic Block Model [47.43048484980115]
This paper considers the problem of community detection on multiple potentially correlated graphs from antheoretical perspective.
We first put forth a random graph model, called the multi-view information block model (MVSBM), designed to generate correlated graphs on the same set of nodes.
arXiv Detail & Related papers (2024-01-17T13:39:38Z) - Human-in-the-loop: Towards Label Embeddings for Measuring Classification Difficulty [14.452983136429967]
In supervised learning, uncertainty can already occur in the first stage of the training process, the annotation phase.
The main idea of this work is to drop the assumption of a ground truth label and instead embed the annotations into a multidimensional space.
The methods developed in this paper readily extend to various situations where multiple annotators independently label instances.
arXiv Detail & Related papers (2023-11-15T11:23:15Z) - Heisenberg-limited Hamiltonian learning for interacting bosons [6.352264764099532]
We develop a protocol for learning a class of interacting bosonic Hamiltonians from dynamics with Heisenberg-limited scaling.
In the protocol, we only use bosonic coherent states, beam splitters, phase shifters, and homodyne measurements.
arXiv Detail & Related papers (2023-07-10T16:44:23Z) - On the Joint Interaction of Models, Data, and Features [82.60073661644435]
We introduce a new tool, the interaction tensor, for empirically analyzing the interaction between data and model through features.
Based on these observations, we propose a conceptual framework for feature learning.
Under this framework, the expected accuracy for a single hypothesis and agreement for a pair of hypotheses can both be derived in closed-form.
arXiv Detail & Related papers (2023-06-07T21:35:26Z) - Quantum Scar States in Coupled Random Graph Models [0.0]
We analyze the Hilbert space connectivity of the $L$ site PXP-model by constructing the Hamiltonian matrices via a Gray code numbering of basis states.
We study the entanglement structure of their energy eigenstates and find two classes of weakly-entangled mid-spectrum states.
arXiv Detail & Related papers (2023-05-14T11:02:10Z) - Emergent Order in Classical Data Representations on Ising Spin Models [0.0]
classical data on quantum spin Hamiltonians yields ordered spin ground states which are used to discriminate data types for binary classification.
We assess the ground states of the Ising Hamiltonian encoded with three separate data sets containing two classes of data.
A new methodology is proposed to predict a certain data class using the ground state of the encoded Ising Hamiltonian.
arXiv Detail & Related papers (2023-03-02T18:17:35Z) - Machine Learned Calabi-Yau Metrics and Curvature [0.0]
Finding Ricci-flat (Calabi-Yau) metrics is a long standing problem in geometry with deep implications for string theory and phenomenology.
A new attack on this problem uses neural networks to engineer approximations to the Calabi-Yau metric within a given K"ahler class.
arXiv Detail & Related papers (2022-11-17T18:59:03Z) - Unsupervised Learning of Equivariant Structure from Sequences [30.974508897223124]
We present an unsupervised framework to learn the symmetry from the time sequence of length at least three.
We will demonstrate that, with our framework, the hidden disentangled structure of the dataset naturally emerges as a by-product.
arXiv Detail & Related papers (2022-10-12T07:29:18Z) - Quantum annealing of Cayley-tree Ising spins at small scales [0.7046417074932257]
Ising Hamiltonians programmed with up to $N=22$ spins mapped on various Cayley tree graphs.
This demonstrates the possibilities of high-dimensional qubit connection in quantum simulators.
arXiv Detail & Related papers (2020-11-03T12:05:51Z) - Deducing neighborhoods of classes from a fitted model [68.8204255655161]
In this article a new kind of interpretable machine learning method is presented.
It can help to understand the partitioning of the feature space into predicted classes in a classification model using quantile shifts.
Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed.
arXiv Detail & Related papers (2020-09-11T16:35:53Z) - Tractable Inference in Credal Sentential Decision Diagrams [116.6516175350871]
Probabilistic sentential decision diagrams are logic circuits where the inputs of disjunctive gates are annotated by probability values.
We develop the credal sentential decision diagrams, a generalisation of their probabilistic counterpart that allows for replacing the local probabilities with credal sets of mass functions.
For a first empirical validation, we consider a simple application based on noisy seven-segment display images.
arXiv Detail & Related papers (2020-08-19T16:04:34Z) - Structural Causal Models Are (Solvable by) Credal Networks [70.45873402967297]
Causal inferences can be obtained by standard algorithms for the updating of credal nets.
This contribution should be regarded as a systematic approach to represent structural causal models by credal networks.
Experiments show that approximate algorithms for credal networks can immediately be used to do causal inference in real-size problems.
arXiv Detail & Related papers (2020-08-02T11:19:36Z) - Learning from Aggregate Observations [82.44304647051243]
We study the problem of learning from aggregate observations where supervision signals are given to sets of instances.
We present a general probabilistic framework that accommodates a variety of aggregate observations.
Simple maximum likelihood solutions can be applied to various differentiable models.
arXiv Detail & Related papers (2020-04-14T06:18:50Z) - Bayesian inference of chaotic dynamics by merging data assimilation,
machine learning and expectation-maximization [0.0]
We show how to combine data assimilation and machine learning in several ways to reconstruct high-dimensional chaotic dynamics.
We numerically and successfully test the approach on two relevant low-order chaotic models with distinct identifiability.
arXiv Detail & Related papers (2020-01-17T12:46:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.