Machine learning magnetism from simple global descriptors
- URL: http://arxiv.org/abs/2509.05909v1
- Date: Sun, 07 Sep 2025 03:53:45 GMT
- Title: Machine learning magnetism from simple global descriptors
- Authors: Ahmed E. Fahmy,
- Abstract summary: We train machine learning classifiers on experimentally validated magnetic materials using a limited number of simple compositional, structural, and electronic descriptors.<n>Our propagation vector classifiers achieve accuracies above 92%, outperforming recent studies in reliably distinguishing zero from nonzero propagation vector structures.<n>In parallel, LightGBM and XGBoost models trained directly on the Materials Project labels achieve accuracies of 84-86%, which proves useful for large-scale screening for magnetic classes.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The reliable identification of magnetic ground states remains a major challenge in high-throughput materials databases, where density functional theory (DFT) workflows often converge to ferromagnetic (FM) solutions. Here, we partially address this challenge by developing machine learning classifiers trained on experimentally validated MAGNDATA magnetic materials leveraging a limited number of simple compositional, structural, and electronic descriptors sourced from the Materials Project database. Our propagation vector classifiers achieve accuracies above 92%, outperforming recent studies in reliably distinguishing zero from nonzero propagation vector structures, and exposing a systematic ferromagnetic bias inherent to the Materials Project database for more than 7,843 materials. In parallel, LightGBM and XGBoost models trained directly on the Materials Project labels achieve accuracies of 84-86% (with macro F1 average scores of 63-66%), which proves useful for large-scale screening for magnetic classes, if refined by MAGNDATA-trained classifiers. These results underscore the role of machine learning techniques as corrective and exploratory tools, enabling more trustworthy databases and accelerating progress toward the identification of materials with various properties.
Related papers
- A New Workflow for Materials Discovery Bridging the Gap Between Experimental Databases and Graph Neural Networks [10.116093920635583]
We propose an alignment process between experimental databases and Crystallographic Information Files (CIF) from the Inorganic Crystal Structure Database (ICSD)<n>Our approach enables the creation of a database that can fully leverage state-of-the-art model architectures for material property prediction.<n>We demonstrate significant improvements in both Mean Absolute Error (MAE) and Correct Classification Rate ( CCR) in predicting the ordering temperatures and magnetic ground states of magnetic materials.
arXiv Detail & Related papers (2026-01-31T14:44:02Z) - Adapting HFMCA to Graph Data: Self-Supervised Learning for Generalizable fMRI Representations [57.054499278843856]
Functional magnetic resonance imaging (fMRI) analysis faces significant challenges due to limited dataset sizes and domain variability between studies.<n>Traditional self-supervised learning methods inspired by computer vision often rely on positive and negative sample pairs.<n>We propose adapting a recently developed Hierarchical Functional Maximal Correlation Algorithm (HFMCA) to graph-structured fMRI data.
arXiv Detail & Related papers (2025-10-05T12:35:01Z) - Advancing Magnetic Materials Discovery -- A structure-based machine learning approach for magnetic ordering and magnetic moment prediction [0.0]
Accurately predicting magnetic behavior across diverse materials systems remains a longstanding challenge.<n>In this work, a refined descriptor is proposed that significantly improves the prediction of two critical magnetic properties.<n>The model achieves an accuracy of 82.4% for magnetic ordering classification and balanced recall across FM and FiM classes.
arXiv Detail & Related papers (2025-07-02T17:24:50Z) - Learning to Rank Chain-of-Thought: Using a Small Model [77.75522308463667]
This paper introduces the Energy Outcome Reward Model (EORM), a highly efficient, lightweight verifier designed to address this challenge.<n>EORM uses an energy-based framework to rank Chain-of-Thought (CoT) solutions, learning to distinguish correct from incorrect reasoning using only simple outcome labels.<n>With only 55M parameters, over 127 times smaller than typical reward models, EORM boosts the accuracy of Llama 3 8B to 90.7% on GSM8k and 63.7% on MATH.
arXiv Detail & Related papers (2025-05-21T01:06:29Z) - DARWIN 1.5: Large Language Models as Materials Science Adapted Learners [46.7259033847682]
We propose DARWIN 1.5, the largest open-source large language model tailored for materials science.<n> DARWIN eliminates the need for task-specific descriptors and enables a flexible, unified approach to material property prediction and discovery.<n>Our approach integrates 6M material domain papers and 21 experimental datasets from 49,256 materials across modalities while enabling cross-task knowledge transfer.
arXiv Detail & Related papers (2024-12-16T16:51:27Z) - MAmmoTH-VL: Eliciting Multimodal Reasoning with Instruction Tuning at Scale [66.73529246309033]
multimodal large language models (MLLMs) have shown significant potential in a broad range of multimodal tasks.<n>Existing instruction-tuning datasets only provide phrase-level answers without any intermediate rationales.<n>We introduce a scalable and cost-effective method to construct a large-scale multimodal instruction-tuning dataset with rich intermediate rationales.
arXiv Detail & Related papers (2024-12-06T18:14:24Z) - Northeast Materials Database (NEMAD): Enabling Discovery of High Transition Temperature Magnetic Compounds [1.1856958240619673]
This study uses Large Language Models (LLMs) to create a comprehensive, experiment-based, magnetic materials database.
The database incorporates chemical composition, magnetic phase transition temperatures, structural details, and magnetic properties.
Machine learning models were developed to classify materials and predict transition temperatures.
arXiv Detail & Related papers (2024-09-24T02:27:10Z) - AI-accelerated Discovery of Altermagnetic Materials [48.261668305411845]
Altermagnetism, a new magnetic phase, has been theoretically proposed and experimentally verified to be distinct from ferromagnetism and antiferromagnetism.<n>We propose an automated discovery approach empowered by an AI search engine.<n>We successfully discovered 50 new altermagnetic materials that cover metals, semiconductors, and insulators.
arXiv Detail & Related papers (2023-11-08T01:06:48Z) - Large Language Models as Master Key: Unlocking the Secrets of Materials
Science with GPT [9.33544942080883]
This article presents a new natural language processing (NLP) task called structured information inference (SII) to address the complexities of information extraction at the device level in materials science.
We accomplished this task by tuning GPT-3 on an existing perovskite solar cell FAIR dataset with 91.8% F1-score and extended the dataset with data published since its release.
We also designed experiments to predict the electrical performance of solar cells and design materials or devices with targeted parameters using large language models (LLMs)
arXiv Detail & Related papers (2023-04-05T04:01:52Z) - Prediction of Large Magnetic Moment Materials With Graph Neural Networks
and Random Forests [0.0]
We use state-of-the-art machine learning methods to scan the Inorganic Crystal Structure Database for ferromagnetic materials.
For random forests, we use a method to select nearly one hundred relevant descriptors based on chemical composition and crystal structure.
We find 15 materials that are likely to have large magnetic moments and have not been yet studied experimentally.
arXiv Detail & Related papers (2021-11-29T17:09:37Z) - Multilinear Compressive Learning with Prior Knowledge [106.12874293597754]
Multilinear Compressive Learning (MCL) framework combines Multilinear Compressive Sensing and Machine Learning into an end-to-end system.
Key idea behind MCL is the assumption of the existence of a tensor subspace which can capture the essential features from the signal for the downstream learning task.
In this paper, we propose a novel solution to address both of the aforementioned requirements, i.e., How to find those tensor subspaces in which the signals of interest are highly separable?
arXiv Detail & Related papers (2020-02-17T19:06:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.