FusionMAE: large-scale pretrained model to optimize and simplify diagnostic and control of fusion plasma
- URL: http://arxiv.org/abs/2509.12945v1
- Date: Tue, 16 Sep 2025 10:50:29 GMT
- Title: FusionMAE: large-scale pretrained model to optimize and simplify diagnostic and control of fusion plasma
- Authors: Zongyu Yang, Zhenghao Yang, Wenjing Tian, Jiyuan Li, Xiang Sun, Guohui Zheng, Songfen Liu, Niannian Wu, Rongpeng Li, Zhaohe Xu, Bo Li, Zhongbing Shi, Zhe Gao, Wei Chen, Xiaoquan Ji, Min Xu, Wulyu Zhong,
- Abstract summary: Fusion masked auto-encoder (FusionMAE) is pre-trained to compress the information from 88 diagnostic signals into a concrete embedding.<n>The model acquires the capability for 'virtual backup diagnosis', enabling the inference of missing diagnostic data with 96.7% reliability.<n>This work pioneers large-scale AI model integration in fusion energy, demonstrating how pre-trained embeddings can simplify the system interface.
- Score: 12.903472110983586
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In magnetically confined fusion device, the complex, multiscale, and nonlinear dynamics of plasmas necessitate the integration of extensive diagnostic systems to effectively monitor and control plasma behaviour. The complexity and uncertainty arising from these extensive systems and their tangled interrelations has long posed a significant obstacle to the acceleration of fusion energy development. In this work, a large-scale model, fusion masked auto-encoder (FusionMAE) is pre-trained to compress the information from 88 diagnostic signals into a concrete embedding, to provide a unified interface between diagnostic systems and control actuators. Two mechanisms are proposed to ensure a meaningful embedding: compression-reduction and missing-signal reconstruction. Upon completion of pre-training, the model acquires the capability for 'virtual backup diagnosis', enabling the inference of missing diagnostic data with 96.7% reliability. Furthermore, the model demonstrates three emergent capabilities: automatic data analysis, universal control-diagnosis interface, and enhancement of control performance on multiple tasks. This work pioneers large-scale AI model integration in fusion energy, demonstrating how pre-trained embeddings can simplify the system interface, reducing necessary diagnostic systems and optimize operation performance for future fusion reactors.
Related papers
- TokaMark: A Comprehensive Benchmark for MAST Tokamak Plasma Models [56.94569090844015]
TokaMark is a structured benchmark to evaluate AI models on real experimental data collected from the Mega Ampere Spherical Tokamak (MAST)<n>TokaMark aims to accelerate progress in data-driven AI-based plasma modeling, contributing to the broader goal of achieving sustainable and stable fusion energy.
arXiv Detail & Related papers (2026-02-05T16:49:44Z) - Hypergraph Contrastive Sensor Fusion for Multimodal Fault Diagnosis in Induction Motors [9.439278188304502]
This paper proposes the Multimodal Hypergraph Contrastive Attention Network (MM-HCAN), a unified framework for robust fault diagnosis.<n> MM-HCAN is the first to integrate contrastive learning within a hypergraph topology specifically designed for multimodal sensor fusion.<n>It achieves up to 99.82% accuracy with strong cross-domain generalisation and resilience to noise, demonstrating its suitability for real-world deployment.
arXiv Detail & Related papers (2025-10-17T11:28:26Z) - USAD: End-to-End Human Activity Recognition via Diffusion Model with Spatiotemporal Attention [8.061018589897277]
Human activity recognition is a task that finds broad applications in health monitoring, safety protection, and sports analysis.<n>Despite proliferating research, human activity recognition still faces key challenges, including the scarcity of labeled samples for rare activities.<n>This paper proposes a comprehensive optimization approach centered on multi-attention interaction mechanisms.
arXiv Detail & Related papers (2025-07-03T17:38:44Z) - RadFabric: Agentic AI System with Reasoning Capability for Radiology [61.25593938175618]
RadFabric is a multi agent, multimodal reasoning framework that unifies visual and textual analysis for comprehensive CXR interpretation.<n>System employs specialized CXR agents for pathology detection, an Anatomical Interpretation Agent to map visual findings to precise anatomical structures, and a Reasoning Agent powered by large multimodal reasoning models to synthesize visual, anatomical, and clinical data into transparent and evidence based diagnoses.
arXiv Detail & Related papers (2025-06-17T03:10:33Z) - Data Driven Diagnosis for Large Cyber-Physical-Systems with Minimal Prior Information [40.827236517095514]
We present a new diagnostic approach that operates with minimal prior knowledge.<n>Our method combines a neural network-based symptom generator, which employs subsystem-level anomaly detection, with a new graph diagnosis algorithm.<n>Our results thus highlight our approach's potential for practical applications with large and complex cyber-physical systems.
arXiv Detail & Related papers (2025-06-12T11:56:58Z) - Learning to Hear Broken Motors: Signature-Guided Data Augmentation for Induction-Motor Diagnostics [40.044978986425676]
The application of machine learning (ML) algorithms in the intelligent diagnosis of three-phase engines has the potential to significantly enhance diagnostic performance and accuracy.<n>We propose a novel unsupervised anomaly generation methodology that takes into account the engine physics model.<n>We propose Signature-Guided Data Augmentation (SGDA), an unsupervised framework that synthesizes physically plausible faults directly in the frequency domain of healthy current signals.
arXiv Detail & Related papers (2025-06-10T03:36:16Z) - Attention-Based Multiscale Temporal Fusion Network for Uncertain-Mode Fault Diagnosis in Multimode Processes [1.9389881806157316]
Fault diagnosis in multimode processes plays a critical role in ensuring the safe operation of industrial systems.<n>It faces a great challenge yet to be addressed - that is, the significant distributional differences among monitoring data from multiple modes.<n>This paper introduces a novel method called attention-based multiscale temporal fusion network.<n>The proposed model achieves superior diagnostic performance and maintains a small model size.
arXiv Detail & Related papers (2025-04-07T15:16:22Z) - GS-TransUNet: Integrated 2D Gaussian Splatting and Transformer UNet for Accurate Skin Lesion Analysis [44.99833362998488]
We present a novel approach that combines 2D Gaussian splatting with the Transformer UNet architecture for automated skin cancer diagnosis.<n>Our findings illustrate significant advancements in the precision of segmentation and classification.<n>This integration sets new benchmarks in the field and highlights the potential for further research into multi-task medical image analysis methodologies.
arXiv Detail & Related papers (2025-02-23T23:28:47Z) - Multimodal Super-Resolution: Discovering hidden physics and its application to fusion plasmas [4.606409469837236]
A non-linear system governed by multi-spatial and multi-temporal physics scales cannot be fully understood with a single diagnostic.
We introduce a machine learning methodology to address this issue.
We use other diagnostics to produce super-resolution data, capturing detailed structural evolution and responses to perturbations previously unobservable.
arXiv Detail & Related papers (2024-05-09T17:06:51Z) - Integrating LLMs for Explainable Fault Diagnosis in Complex Systems [0.0]
This paper introduces an integrated system designed to enhance the explainability of fault diagnostics in complex systems, such as nuclear power plants.
By combining a physics-based diagnostic tool with a Large Language Model, we offer a novel solution that not only identifies faults but also provides clear, understandable explanations of their causes and implications.
arXiv Detail & Related papers (2024-02-08T22:11:21Z) - Brain Imaging-to-Graph Generation using Adversarial Hierarchical Diffusion Models for MCI Causality Analysis [44.45598796591008]
Brain imaging-to-graph generation (BIGG) framework is proposed to map functional magnetic resonance imaging (fMRI) into effective connectivity for mild cognitive impairment analysis.
The hierarchical transformers in the generator are designed to estimate the noise at multiple scales.
Evaluations of the ADNI dataset demonstrate the feasibility and efficacy of the proposed model.
arXiv Detail & Related papers (2023-05-18T06:54:56Z) - System Resilience through Health Monitoring and Reconfiguration [56.448036299746285]
We demonstrate an end-to-end framework to improve the resilience of man-made systems to unforeseen events.
The framework is based on a physics-based digital twin model and three modules tasked with real-time fault diagnosis, prognostics and reconfiguration.
arXiv Detail & Related papers (2022-08-30T20:16:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.