Interaction Screening and Pseudolikelihood Approaches for Tensor
Learning in Ising Models
- URL: http://arxiv.org/abs/2310.13232v1
- Date: Fri, 20 Oct 2023 02:42:32 GMT
- Title: Interaction Screening and Pseudolikelihood Approaches for Tensor
Learning in Ising Models
- Authors: Tianyu Liu and Somabha Mukherjee
- Abstract summary: We study two well known methods of Ising structure learning, namely the pseudolikelihood approach and the interaction screening approach.
We show that both approaches retrieve the underlying hypernetwork structure using a sample size logarithmic in the number of network nodes.
- Score: 8.622642118842624
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we study two well known methods of Ising structure learning,
namely the pseudolikelihood approach and the interaction screening approach, in
the context of tensor recovery in $k$-spin Ising models. We show that both
these approaches, with proper regularization, retrieve the underlying
hypernetwork structure using a sample size logarithmic in the number of network
nodes, and exponential in the maximum interaction strength and maximum
node-degree. We also track down the exact dependence of the rate of tensor
recovery on the interaction order $k$, that is allowed to grow with the number
of samples and nodes, for both the approaches. Finally, we provide a
comparative discussion of the performance of the two approaches based on
simulation studies, which also demonstrate the exponential dependence of the
tensor recovery rate on the maximum coupling strength.
Related papers
- Compositional Curvature Bounds for Deep Neural Networks [7.373617024876726]
A key challenge that threatens the widespread use of neural networks in safety-critical applications is their vulnerability to adversarial attacks.
We study the second-order behavior of continuously differentiable deep neural networks, focusing on robustness against adversarial perturbations.
We introduce a novel algorithm to analytically compute provable upper bounds on the second derivative of neural networks.
arXiv Detail & Related papers (2024-06-07T17:50:15Z) - Performance Gaps in Multi-view Clustering under the Nested Matrix-Tensor
Model [7.4968526280735945]
We study the estimation of a planted signal hidden in a recently introduced nested matrix-tensor model.
We quantify here the performance gap between a tensor-based approach and a tractable alternative approach.
arXiv Detail & Related papers (2024-02-16T13:31:43Z) - Alteration Detection of Tensor Dependence Structure via
Sparsity-Exploited Reranking Algorithm [3.7363073304294336]
We formulate the problem under the popularly adopted tensor-normal distributions and aim at two-sample correlation/partial correlation comparisons.
We propose a novel Sparsity-Exploited Reranking Algorithm (SERA) to further improve the multiple testing efficiency.
The properties of the proposed test are derived and the algorithm is shown to control the false discovery at the pre-specified level.
arXiv Detail & Related papers (2023-10-13T01:04:22Z) - A Nested Matrix-Tensor Model for Noisy Multi-view Clustering [5.132856740094742]
We propose a nested matrix-tensor model which extends the spiked rank-one tensor model of order three.
We show that our theoretical results allow us to anticipate the exact accuracy of the proposed clustering approach.
Our analysis unveils unexpected and non-trivial phase transition phenomena depending on the model parameters.
arXiv Detail & Related papers (2023-05-31T16:13:46Z) - Interpolation-based Correlation Reduction Network for Semi-Supervised
Graph Learning [49.94816548023729]
We propose a novel graph contrastive learning method, termed Interpolation-based Correlation Reduction Network (ICRN)
In our method, we improve the discriminative capability of the latent feature by enlarging the margin of decision boundaries.
By combining the two settings, we extract rich supervision information from both the abundant unlabeled nodes and the rare yet valuable labeled nodes for discnative representation learning.
arXiv Detail & Related papers (2022-06-06T14:26:34Z) - Convex Analysis of the Mean Field Langevin Dynamics [49.66486092259375]
convergence rate analysis of the mean field Langevin dynamics is presented.
$p_q$ associated with the dynamics allows us to develop a convergence theory parallel to classical results in convex optimization.
arXiv Detail & Related papers (2022-01-25T17:13:56Z) - Counterfactual Maximum Likelihood Estimation for Training Deep Networks [83.44219640437657]
Deep learning models are prone to learning spurious correlations that should not be learned as predictive clues.
We propose a causality-based training framework to reduce the spurious correlations caused by observable confounders.
We conduct experiments on two real-world tasks: Natural Language Inference (NLI) and Image Captioning.
arXiv Detail & Related papers (2021-06-07T17:47:16Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning [92.05556163518999]
MARL exacerbates matters by imposing various constraints on communication and observability.
For value-based methods, it poses challenges in accurately representing the optimal value function.
For policy gradient methods, it makes training the critic difficult and exacerbates the problem of the lagging critic.
We show that from a learning theory perspective, both problems can be addressed by accurately representing the associated action-value function.
arXiv Detail & Related papers (2021-05-31T23:08:05Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.