Reducing Simulation Dependence in Neutrino Telescopes with Masked Point Transformers
- URL: http://arxiv.org/abs/2510.01733v1
- Date: Thu, 02 Oct 2025 07:18:19 GMT
- Title: Reducing Simulation Dependence in Neutrino Telescopes with Masked Point Transformers
- Authors: Felix J. Yu, Nicholas Kamp, Carlos A. Argüelles,
- Abstract summary: We present the first self-supervised training pipeline for neutrino telescopes.<n>By shifting the majority of training to real data, this approach minimizes reliance on simulations.<n>This represents a fundamental departure from previous machine learning applications in neutrino telescopes.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning techniques in neutrino physics have traditionally relied on simulated data, which provides access to ground-truth labels. However, the accuracy of these simulations and the discrepancies between simulated and real data remain significant concerns, particularly for large-scale neutrino telescopes that operate in complex natural media. In recent years, self-supervised learning has emerged as a powerful paradigm for reducing dependence on labeled datasets. Here, we present the first self-supervised training pipeline for neutrino telescopes, leveraging point cloud transformers and masked autoencoders. By shifting the majority of training to real data, this approach minimizes reliance on simulations, thereby mitigating associated systematic uncertainties. This represents a fundamental departure from previous machine learning applications in neutrino telescopes, paving the way for substantial improvements in event reconstruction and classification.
Related papers
- Learning Hamiltonians for solid-state quantum simulators [0.0]
We introduce a generalizable framework for learning to identify effective Hamiltonians directly from experimental data in solid-state quantum systems.<n>Our approach is based on a physics-informed neural network architecture that embeds physical constraints directly into the model structure.
arXiv Detail & Related papers (2026-03-03T11:37:43Z) - Simulation-Based Pretraining and Domain Adaptation for Astronomical Time Series with Minimal Labeled Data [0.12744523252873352]
We present a pre-training approach that leverages simulations, significantly reducing the need for labeled examples from real observations.<n>Our models, trained on simulated data from multiple astronomical surveys (ZTF and LSST), learn generalizable representations that transfer effectively to downstream tasks.<n>Remarkably, our models exhibit effective zero-shot transfer capabilities, achieving comparable performance on future telescope (LSST) simulations when trained solely on existing telescope (ZTF) data.
arXiv Detail & Related papers (2025-10-14T20:07:14Z) - Fusing CFD and measurement data using transfer learning [49.1574468325115]
We introduce a non-linear method based on neural networks combining simulation and measurement data via transfer learning.<n>In a first step, the neural network is trained on simulation data to learn spatial features of the distributed quantities.<n>The second step involves transfer learning on the measurement data to correct for systematic errors between simulation and measurement by only re-training a small subset of the entire neural network model.
arXiv Detail & Related papers (2025-07-28T07:21:46Z) - Learning Efficient Representations of Neutrino Telescope Events [0.0]
Neutrino telescopes detect rare interactions of particles produced in some of the most extreme environments in the Universe.<n>Given their size and the high frequency of background interactions, these telescopes amass an enormous quantity of large variance, high-dimensional data.<n>We present a novel approach, called om2vec, that employs transformer-based variational autoencoders to efficiently represent neutrino telescope events by learning compact and latent representations.
arXiv Detail & Related papers (2024-10-17T02:07:54Z) - Transformer-Powered Surrogates Close the ICF Simulation-Experiment Gap with Extremely Limited Data [24.24053233941972]
This paper presents a novel transformer-powered approach for enhancing prediction accuracy in multi-modal output scenarios.
The proposed approach integrates transformer-based architecture with a novel graph-based hyper- parameter optimization technique.
We demonstrate the efficacy of our approach on inertial confinement fusion experiments, where only 10 shots of real-world data are available.
arXiv Detail & Related papers (2023-12-06T17:53:06Z) - Physics-Enhanced Multi-fidelity Learning for Optical Surface Imprint [1.0878040851638]
We propose a novel method to use multi-fidelity neural networks (MFNN) to solve this inverse problem.
We build up the NN model via pure simulation data, and then bridge the sim-to-real gap via transfer learning.
Considering the difficulty of collecting real experimental data, we use NN to dig out the unknown physics and also implant the known physics into the transfer learning framework.
arXiv Detail & Related papers (2023-11-17T01:55:15Z) - Physics-Driven Turbulence Image Restoration with Stochastic Refinement [80.79900297089176]
Image distortion by atmospheric turbulence is a critical problem in long-range optical imaging systems.
Fast and physics-grounded simulation tools have been introduced to help the deep-learning models adapt to real-world turbulence conditions.
This paper proposes the Physics-integrated Restoration Network (PiRN) to help the network to disentangle theity from the degradation and the underlying image.
arXiv Detail & Related papers (2023-07-20T05:49:21Z) - Implicit Geometry and Interaction Embeddings Improve Few-Shot Molecular
Property Prediction [53.06671763877109]
We develop molecular embeddings that encode complex molecular characteristics to improve the performance of few-shot molecular property prediction.
Our approach leverages large amounts of synthetic data, namely the results of molecular docking calculations.
On multiple molecular property prediction benchmarks, training from the embedding space substantially improves Multi-Task, MAML, and Prototypical Network few-shot learning performance.
arXiv Detail & Related papers (2023-02-04T01:32:40Z) - Continual learning autoencoder training for a particle-in-cell
simulation via streaming [52.77024349608834]
upcoming exascale era will provide a new generation of physics simulations with high resolution.
These simulations will have a high resolution, which will impact the training of machine learning models since storing a high amount of simulation data on disk is nearly impossible.
This work presents an approach that trains a neural network concurrently to a running simulation without data on a disk.
arXiv Detail & Related papers (2022-11-09T09:55:14Z) - Simulation-Based Parallel Training [55.41644538483948]
We present our ongoing work to design a training framework that alleviates those bottlenecks.
It generates data in parallel with the training process.
We present a strategy to mitigate this bias with a memory buffer.
arXiv Detail & Related papers (2022-11-08T09:31:25Z) - Quantum-tailored machine-learning characterization of a superconducting
qubit [50.591267188664666]
We develop an approach to characterize the dynamics of a quantum device and learn device parameters.
This approach outperforms physics-agnostic recurrent neural networks trained on numerically generated and experimental data.
This demonstration shows how leveraging domain knowledge improves the accuracy and efficiency of this characterization task.
arXiv Detail & Related papers (2021-06-24T15:58:57Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.