Graph convolutional network for predicting abnormal grain growth in Monte Carlo simulations of microstructural evolution
- URL: http://arxiv.org/abs/2110.09326v2
- Date: Thu, 11 Jul 2024 03:45:01 GMT
- Title: Graph convolutional network for predicting abnormal grain growth in Monte Carlo simulations of microstructural evolution
- Authors: Ryan Cohn, Elizabeth Holm,
- Abstract summary: We generate a large dataset of Monte Carlo simulations of abnormal grain growth.
We train simple graph convolution networks to predict which initial microstructures will exhibit abnormal grain growth.
The graph neural network outperformed the computer vision method and achieved 73% prediction accuracy and fewer false positives.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent developments in graph neural networks show promise for predicting the occurrence of abnormal grain growth, which has been a particularly challenging area of research due to its apparent stochastic nature. In this study, we generate a large dataset of Monte Carlo simulations of abnormal grain growth. We train simple graph convolution networks to predict which initial microstructures will exhibit abnormal grain growth, and compare the results to a standard computer vision approach for the same task. The graph neural network outperformed the computer vision method and achieved 73% prediction accuracy and fewer false positives. It also provided some physical insight into feature importance and the relevant length scale required to maximize predictive performance. Analysis of the uncertainty in the Monte Carlo simulations provides additional insights for ongoing work in this area.
Related papers
- Predicting Grain Growth in Polycrystalline Materials Using Deep Learning Time Series Models [0.9558392439655014]
Grain Growth strongly influences the mechanical behavior of materials, making its prediction a key objective in microstructural engineering.<n>In this study, several deep learning approaches were evaluated, including recurrent neural networks (RNN), long short-term memory (LSTM), temporal convolutional networks (TCN), and transformers.<n>The LSTM network achieved the highest accuracy (above 90%) and the most stable performance, maintaining physically consistent predictions over extended horizons.
arXiv Detail & Related papers (2025-11-07T18:29:42Z) - Modeling Gene Expression Distributional Shifts for Unseen Genetic Perturbations [44.619690829431214]
We train a neural network to predict distributional responses in gene expression following genetic perturbations.<n>Our model predicts gene-level histograms conditioned on perturbations and outperforms baselines in capturing higher-order statistics.
arXiv Detail & Related papers (2025-07-01T06:04:28Z) - High-fidelity Grain Growth Modeling: Leveraging Deep Learning for Fast Computations [0.0]
We introduce a machine learning framework that combines a Convolutional Long Short-Term Memory networks with an Autoencoder to efficiently predict grain growth evolution.<n>Results demonstrated that our machine learning approach accelerates grain growth prediction by up to SI89times faster.
arXiv Detail & Related papers (2025-05-08T15:43:40Z) - Learning Augmentation Policies from A Model Zoo for Time Series Forecasting [58.66211334969299]
We introduce AutoTSAug, a learnable data augmentation method based on reinforcement learning.
By augmenting the marginal samples with a learnable policy, AutoTSAug substantially improves forecasting performance.
arXiv Detail & Related papers (2024-09-10T07:34:19Z) - CogDPM: Diffusion Probabilistic Models via Cognitive Predictive Coding [62.075029712357]
This work introduces the Cognitive Diffusion Probabilistic Models (CogDPM)
CogDPM features a precision estimation method based on the hierarchical sampling capabilities of diffusion models and weight the guidance with precision weights estimated by the inherent property of diffusion models.
We apply CogDPM to real-world prediction tasks using the United Kindom precipitation and surface wind datasets.
arXiv Detail & Related papers (2024-05-03T15:54:50Z) - Asymptotic generalization error of a single-layer graph convolutional network [0.0]
We predict the performances of a single-layer graph convolutional network trained on data produced by attributed block models.
We study the high signal-to-noise ratio limit, detail the convergence rates of the GCN and show that, while consistent, it does not reach the Bayes-optimal rate for any of the considered cases.
arXiv Detail & Related papers (2024-02-06T09:07:26Z) - FourierGNN: Rethinking Multivariate Time Series Forecasting from a Pure
Graph Perspective [48.00240550685946]
Current state-of-the-art graph neural network (GNN)-based forecasting methods usually require both graph networks (e.g., GCN) and temporal networks (e.g., LSTM) to capture inter-series (spatial) dynamics and intra-series (temporal) dependencies, respectively.
We propose a novel Fourier Graph Neural Network (FourierGNN) by stacking our proposed Fourier Graph Operator (FGO) to perform matrix multiplications in Fourier space.
Our experiments on seven datasets have demonstrated superior performance with higher efficiency and fewer parameters compared with state-of-the-
arXiv Detail & Related papers (2023-11-10T17:13:26Z) - Robust Graph Representation Learning via Predictive Coding [46.22695915912123]
Predictive coding is a message-passing framework initially developed to model information processing in the brain.
In this work, we build models that rely on the message-passing rule of predictive coding.
We show that the proposed models are comparable to standard ones in terms of performance in both inductive and transductive tasks.
arXiv Detail & Related papers (2022-12-09T03:58:22Z) - Dense Hebbian neural networks: a replica symmetric picture of
unsupervised learning [4.133728123207142]
We consider dense, associative neural-networks trained with no supervision.
We investigate their computational capabilities analytically, via a statistical-mechanics approach, and numerically, via Monte Carlo simulations.
arXiv Detail & Related papers (2022-11-25T12:40:06Z) - Learning Feynman Diagrams using Graph Neural Networks [70.540936204654]
This research uses the graph attention layer which makes matrix element predictions to 1 significant figure accuracy above 90% of the time.
Peak performance was achieved in making predictions to 3 significant figure accuracy over 10% of the time with less than 200 epochs of training.
arXiv Detail & Related papers (2022-11-25T05:53:28Z) - Predicting Biomedical Interactions with Probabilistic Model Selection
for Graph Neural Networks [5.156812030122437]
Current biological networks are noisy, sparse, and incomplete. Experimental identification of such interactions is both time-consuming and expensive.
Deep graph neural networks have shown their effectiveness in modeling graph-structured data and achieved good performance in biomedical interaction prediction.
Our proposed method enables the graph convolutional networks to dynamically adapt their depths to accommodate an increasing number of interactions.
arXiv Detail & Related papers (2022-11-22T20:44:28Z) - ProGReST: Prototypical Graph Regression Soft Trees for Molecular
Property Prediction [1.6114012813668934]
Prototypical Graph Regression Self-explainable Trees (ProGReST) model combines prototype learning, soft decision trees, and Graph Neural Networks.
In ProGReST, the rationale is obtained along with prediction due to the model's built-in interpretability.
arXiv Detail & Related papers (2022-10-07T10:21:24Z) - Towards the Explanation of Graph Neural Networks in Digital Pathology
with Information Flows [67.23405590815602]
Graph Neural Networks (GNNs) are widely adopted in digital pathology.
Existing explainers discover an explanatory subgraph relevant to the prediction.
An explanatory subgraph should be not only necessary for prediction, but also sufficient to uncover the most predictive regions.
We propose IFEXPLAINER, which generates a necessary and sufficient explanation for GNNs.
arXiv Detail & Related papers (2021-12-18T10:19:01Z) - Edge-variational Graph Convolutional Networks for Uncertainty-aware
Disease Prediction [7.6146285961466]
We propose a generalizable framework that can automatically integrate imaging data with non-imaging data in populations for uncertainty-aware disease prediction.
Experimental results on four databases show that our method can consistently and significantly improve the diagnostic accuracy for Autism spectrum disorder, Alzheimer's disease, and ocular diseases.
arXiv Detail & Related papers (2020-09-06T15:53:17Z) - Towards Deeper Graph Neural Networks [63.46470695525957]
Graph convolutions perform neighborhood aggregation and represent one of the most important graph operations.
Several recent studies attribute this performance deterioration to the over-smoothing issue.
We propose Deep Adaptive Graph Neural Network (DAGNN) to adaptively incorporate information from large receptive fields.
arXiv Detail & Related papers (2020-07-18T01:11:14Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.