Combining feature-based approaches with graph neural networks and symbolic regression for synergistic performance and interpretability
- URL: http://arxiv.org/abs/2509.03547v2
- Date: Fri, 05 Sep 2025 16:42:17 GMT
- Title: Combining feature-based approaches with graph neural networks and symbolic regression for synergistic performance and interpretability
- Authors: Rogério Almeida Gouvêa, Pierre-Paul De Breuck, Tatiane Pretto, Gian-Marco Rignanese, Marcos José Leite Santos,
- Abstract summary: MatterVial is an innovative hybrid framework for feature-based machine learning in materials science.<n>Our approach combines the chemical transparency of traditional feature-based models with the predictive power of deep learning architectures.<n>An integrated interpretability module, employing surrogate models and symbolic regression, decodes the latent GNN-derived descriptors into explicit, physically meaningful formulas.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This study introduces MatterVial, an innovative hybrid framework for feature-based machine learning in materials science. MatterVial expands the feature space by integrating latent representations from a diverse suite of pretrained graph neural network (GNN) models including: structure-based (MEGNet), composition-based (ROOST), and equivariant (ORB) graph networks, with computationally efficient, GNN-approximated descriptors and novel features from symbolic regression. Our approach combines the chemical transparency of traditional feature-based models with the predictive power of deep learning architectures. When augmenting the feature-based model MODNet on Matbench tasks, this method yields significant error reductions and elevates its performance to be competitive with, and in several cases superior to, state-of-the-art end-to-end GNNs, with accuracy increases exceeding 40% for multiple tasks. An integrated interpretability module, employing surrogate models and symbolic regression, decodes the latent GNN-derived descriptors into explicit, physically meaningful formulas. This unified framework advances materials informatics by providing a high-performance, transparent tool that aligns with the principles of explainable AI, paving the way for more targeted and autonomous materials discovery.
Related papers
- A Roadmap for Applying Graph Neural Networks to Numerical Data: Insights from Cementitious Materials [5.565428903960444]
This work is among the first few studies to implement Graph neural network (GNN) models to design concrete.<n>GNN is capable of learning from data structured as graphs, capturing relationships through irregular or topology-dependent connections.<n>The proposed framework establishes a strong foundation for future multi-modal and physics-informed GNN models.
arXiv Detail & Related papers (2025-12-16T19:17:05Z) - Recurrent U-Net-Based Graph Neural Network (RUGNN) for Accurate Deformation Predictions in Sheet Material Forming [13.180335574191432]
This study developed a new graph neural network surrogate model named Recurrent U Net-based Graph Neural Network (RUGNN)<n>The RUGNN model can achieve accurate predictions of sheet material deformation fields across multiple forming timesteps.
arXiv Detail & Related papers (2025-07-10T08:14:18Z) - Do We Really Need GNNs with Explicit Structural Modeling? MLPs Suffice for Language Model Representations [50.45261187796993]
Graph Neural Networks (GNNs) fail to fully utilize structural information, whereas Multi-Layer Perceptrons (MLPs) exhibit a surprising ability in structure-aware tasks.<n>This paper introduces a comprehensive probing framework from an information-theoretic perspective.
arXiv Detail & Related papers (2025-06-26T18:10:28Z) - Language Embedding Meets Dynamic Graph: A New Exploration for Neural Architecture Representation Learning [38.323486764309]
We introduce LeDG-Former, an innovative framework that addresses limitations through the synergistic integration of language-based semantic embedding and dynamic graph representation learning.<n>Specifically, we propose a language embedding framework where both neural architectures and hardware platform specifications are projected into a unified semantic space.<n>Our framework achieves superior performance on the cell-structured NAS-Bench-101 and NAS-Bench-201 datasets.
arXiv Detail & Related papers (2025-06-09T13:20:02Z) - SA-GAT-SR: Self-Adaptable Graph Attention Networks with Symbolic Regression for high-fidelity material property prediction [1.4403769872061323]
We introduce a novel computational paradigm, Self-Adaptable Graph Attention Networks integrated with Symbolic Regression (SA-GAT-SR)<n>Our framework employs a self-adaptable encoding algorithm that automatically identifies and adjust attention weights so as to screen critical features from an expansive 180-dimensional feature space.<n>The integrated SR module subsequently distills these features into compact analytical expressions that explicitly reveal quantum-mechanically meaningful relationships.
arXiv Detail & Related papers (2025-05-01T16:05:10Z) - Instruction-Guided Autoregressive Neural Network Parameter Generation [49.800239140036496]
We propose IGPG, an autoregressive framework that unifies parameter synthesis across diverse tasks and architectures.<n>By autoregressively generating neural network weights' tokens, IGPG ensures inter-layer coherence and enables efficient adaptation across models and datasets.<n>Experiments on multiple datasets demonstrate that IGPG consolidates diverse pretrained models into a single, flexible generative framework.
arXiv Detail & Related papers (2025-04-02T05:50:19Z) - Generalized Factor Neural Network Model for High-dimensional Regression [50.554377879576066]
We tackle the challenges of modeling high-dimensional data sets with latent low-dimensional structures hidden within complex, non-linear, and noisy relationships.<n>Our approach enables a seamless integration of concepts from non-parametric regression, factor models, and neural networks for high-dimensional regression.
arXiv Detail & Related papers (2025-02-16T23:13:55Z) - Spatiotemporal Graph Learning with Direct Volumetric Information Passing and Feature Enhancement [62.91536661584656]
We propose a dual-module framework, Cell-embedded and Feature-enhanced Graph Neural Network (aka, CeFeGNN) for learning.<n>We embed learnable cell attributions to the common node-edge message passing process, which better captures the spatial dependency of regional features.<n>Experiments on various PDE systems and one real-world dataset demonstrate that CeFeGNN achieves superior performance compared with other baselines.
arXiv Detail & Related papers (2024-09-26T16:22:08Z) - Mechanistic Neural Networks for Scientific Machine Learning [58.99592521721158]
We present Mechanistic Neural Networks, a neural network design for machine learning applications in the sciences.
It incorporates a new Mechanistic Block in standard architectures to explicitly learn governing differential equations as representations.
Central to our approach is a novel Relaxed Linear Programming solver (NeuRLP) inspired by a technique that reduces solving linear ODEs to solving linear programs.
arXiv Detail & Related papers (2024-02-20T15:23:24Z) - The Role of Foundation Models in Neuro-Symbolic Learning and Reasoning [54.56905063752427]
Neuro-Symbolic AI (NeSy) holds promise to ensure the safe deployment of AI systems.
Existing pipelines that train the neural and symbolic components sequentially require extensive labelling.
New architecture, NeSyGPT, fine-tunes a vision-language foundation model to extract symbolic features from raw data.
arXiv Detail & Related papers (2024-02-02T20:33:14Z) - MAgNET: A Graph U-Net Architecture for Mesh-Based Simulations [0.5185522256407782]
We present MAgNET, which extends the well-known convolutional neural networks to accommodate arbitrary graph-structured data.
We demonstrate the predictive capabilities of MAgNET in surrogate modeling for non-linear finite element simulations in the mechanics of solids.
arXiv Detail & Related papers (2022-11-01T19:23:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.