Machine learning for structure-property relationships: Scalability and
limitations
- URL: http://arxiv.org/abs/2304.05502v1
- Date: Tue, 11 Apr 2023 21:17:28 GMT
- Title: Machine learning for structure-property relationships: Scalability and
limitations
- Authors: Zhongzheng Tian, Sheng Zhang, Gia-Wei Chern
- Abstract summary: We present a scalable machine learning (ML) framework for predicting intensive properties and particularly classifying phases of many-body systems.
Based on the locality assumption, ML model is developed for the prediction of intensive properties of a finite-size block.
We show that the applicability of this approach depends on whether the block-size of the ML model is greater than the characteristic length scale of the system.
- Score: 3.664479980617018
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a scalable machine learning (ML) framework for predicting
intensive properties and particularly classifying phases of many-body systems.
Scalability and transferability are central to the unprecedented computational
efficiency of ML methods. In general, linear-scaling computation can be
achieved through the divide and conquer approach, and the locality of physical
properties is key to partitioning the system into sub-domains that can be
solved separately. Based on the locality assumption, ML model is developed for
the prediction of intensive properties of a finite-size block. Predictions of
large-scale systems can then be obtained by averaging results of the ML model
from randomly sampled blocks of the system. We show that the applicability of
this approach depends on whether the block-size of the ML model is greater than
the characteristic length scale of the system. In particular, in the case of
phase identification across a critical point, the accuracy of the ML prediction
is limited by the diverging correlation length. The two-dimensional Ising model
is used to demonstrate the proposed framework. We obtain an intriguing scaling
relation between the prediction accuracy and the ratio of ML block size over
the spin-spin correlation length. Implications for practical applications are
also discussed.
Related papers
- Uncertainty Quantification in Large Language Models Through Convex Hull Analysis [0.36832029288386137]
This study proposes a novel geometric approach to uncertainty quantification using convex hull analysis.
The proposed method leverages the spatial properties of response embeddings to measure the dispersion and variability of model outputs.
arXiv Detail & Related papers (2024-06-28T07:47:34Z) - Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Data-freeWeight Compress and Denoise for Large Language Models [101.53420111286952]
We propose a novel approach termed Data-free Joint Rank-k Approximation for compressing the parameter matrices.
We achieve a model pruning of 80% parameters while retaining 93.43% of the original performance without any calibration data.
arXiv Detail & Related papers (2024-02-26T05:51:47Z) - Probabilistic ML Verification via Weighted Model Integration [11.812078181471634]
Probability formal verification (PFV) of machine learning models is in its infancy.
We propose a unifying framework for the PFV of ML systems based on Weighted Model Integration (WMI)
We show how successful scaling techniques in the ML verification literature can be generalized beyond their original scope.
arXiv Detail & Related papers (2024-02-07T14:24:04Z) - Variable Importance Matching for Causal Inference [73.25504313552516]
We describe a general framework called Model-to-Match that achieves these goals.
Model-to-Match uses variable importance measurements to construct a distance metric.
We operationalize the Model-to-Match framework with LASSO.
arXiv Detail & Related papers (2023-02-23T00:43:03Z) - A Data-driven feature selection and machine-learning model benchmark for
the prediction of longitudinal dispersion coefficient [29.58577229101903]
An accurate prediction on Longitudinal Dispersion(LD) coefficient can produce a performance leap in related simulation.
In this study, a global optimal feature set was proposed through numerical comparison of the distilled local optimums in performance with representative ML models.
Results show that the support vector machine has significantly better performance than other models.
arXiv Detail & Related papers (2021-07-16T09:50:38Z) - Generalized Matrix Factorization: efficient algorithms for fitting
generalized linear latent variable models to large data arrays [62.997667081978825]
Generalized Linear Latent Variable models (GLLVMs) generalize such factor models to non-Gaussian responses.
Current algorithms for estimating model parameters in GLLVMs require intensive computation and do not scale to large datasets.
We propose a new approach for fitting GLLVMs to high-dimensional datasets, based on approximating the model using penalized quasi-likelihood.
arXiv Detail & Related papers (2020-10-06T04:28:19Z) - Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled
Markov Chains [34.77971292478243]
The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture.
We develop a training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient.
We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.
arXiv Detail & Related papers (2020-10-05T08:11:55Z) - Surrogate Locally-Interpretable Models with Supervised Machine Learning
Algorithms [8.949704905866888]
Supervised Machine Learning algorithms have become popular in recent years due to their superior predictive performance over traditional statistical methods.
The main focus is on interpretability, the resulting surrogate model also has reasonably good predictive performance.
arXiv Detail & Related papers (2020-07-28T23:46:16Z) - Slice Sampling for General Completely Random Measures [74.24975039689893]
We present a novel Markov chain Monte Carlo algorithm for posterior inference that adaptively sets the truncation level using auxiliary slice variables.
The efficacy of the proposed algorithm is evaluated on several popular nonparametric models.
arXiv Detail & Related papers (2020-06-24T17:53:53Z) - Localized Debiased Machine Learning: Efficient Inference on Quantile
Treatment Effects and Beyond [69.83813153444115]
We consider an efficient estimating equation for the (local) quantile treatment effect ((L)QTE) in causal inference.
Debiased machine learning (DML) is a data-splitting approach to estimating high-dimensional nuisances.
We propose localized debiased machine learning (LDML), which avoids this burdensome step.
arXiv Detail & Related papers (2019-12-30T14:42:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.