Machine learning automorphic forms for black holes
- URL: http://arxiv.org/abs/2505.05549v1
- Date: Thu, 08 May 2025 18:00:00 GMT
- Title: Machine learning automorphic forms for black holes
- Authors: Vishnu Jejjala, Suresh Nampuri, Dumisani Nxumalo, Pratik Roy, Abinash Swain,
- Abstract summary: We show that machine learning can accurately predict modular weights from truncated expansions.<n>This study establishes a proof of concept for using machine learning to identify how data is organized in terms of modular symmetries in gravitational systems.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modular, Jacobi, and mock-modular forms serve as generating functions for BPS black hole degeneracies. By training feed-forward neural networks on Fourier coefficients of automorphic forms derived from the Dedekind eta function, Eisenstein series, and Jacobi theta functions, we demonstrate that machine learning techniques can accurately predict modular weights from truncated expansions. Our results reveal strong performance for negative weight modular and quasi-modular forms, particularly those arising in exact black hole counting formulae, with lower accuracy for positive weights and more complicated combinations of Jacobi theta functions. This study establishes a proof of concept for using machine learning to identify how data is organized in terms of modular symmetries in gravitational systems and suggests a pathway toward automated detection and verification of symmetries in quantum gravity.
Related papers
- Machine learning modularity [6.316250090403085]
This work introduces a machine learning framework for automatically simplifying complex expressions involving multiple elliptic Gamma functions.<n>The model learns to apply algebraic identities, particularly the SL$(2,mathbbZ)$ and SL$(2,mathbbZ)$ modular transformations, to reduce heavily scrambled expressions to their canonical forms.
arXiv Detail & Related papers (2026-01-05T04:17:55Z) - Disordered Dynamics in High Dimensions: Connections to Random Matrices and Machine Learning [52.26396748560348]
We provide an overview of high dimensional dynamical systems driven by random matrices.<n>We focus on applications to simple models of learning and generalization in machine learning theory.
arXiv Detail & Related papers (2026-01-03T00:12:32Z) - MOFGPT: Generative Design of Metal-Organic Frameworks using Language Models [5.417632175667162]
Metal-Organic Frameworks (MOFs) with application-specific properties remain a central challenge in materials chemistry.<n>We present a reinforcement learning-enhanced, transformer-based framework for the de novo design of MOFs.<n>By integrating property feedback into sequence generation, our method drives the model toward synthesizable, topologically valid MOFs.
arXiv Detail & Related papers (2025-05-30T20:09:11Z) - GausSim: Foreseeing Reality by Gaussian Simulator for Elastic Objects [55.02281855589641]
GausSim is a novel neural network-based simulator designed to capture the dynamic behaviors of real-world elastic objects represented through Gaussian kernels.<n>We leverage continuum mechanics and treat each kernel as a Center of Mass System (CMS) that represents continuous piece of matter.<n>In addition, GausSim incorporates explicit physics constraints, such as mass and momentum conservation, ensuring interpretable results and robust, physically plausible simulations.
arXiv Detail & Related papers (2024-12-23T18:58:17Z) - Fourier Neural Operators for Learning Dynamics in Quantum Spin Systems [77.88054335119074]
We use FNOs to model the evolution of random quantum spin systems.
We apply FNOs to a compact set of Hamiltonian observables instead of the entire $2n$ quantum wavefunction.
arXiv Detail & Related papers (2024-09-05T07:18:09Z) - On the Sample Complexity of Quantum Boltzmann Machine Learning [0.0]
We give an operational definition of QBM learning in terms of the difference in expectation values between the model and target.<n>We prove that a solution can be obtained with gradient descent using at most a number of Gibbs states.<n>In particular, we give pre-training strategies based on mean-field, Gaussian Fermionic, and geometrically local Hamiltonians.
arXiv Detail & Related papers (2023-06-26T18:00:50Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - On the Integration of Physics-Based Machine Learning with Hierarchical
Bayesian Modeling Techniques [0.0]
This paper proposes to embed mechanics-based models into the mean function of a Gaussian Process (GP) model and characterize potential discrepancies through kernel machines.
The stationarity of the kernel function is a difficult hurdle in the sequential processing of long data sets, resolved through hierarchical Bayesian techniques.
Using numerical and experimental examples, potential applications of the proposed method to structural dynamics inverse problems are demonstrated.
arXiv Detail & Related papers (2023-03-01T02:29:41Z) - Modular Deep Learning [120.36599591042908]
Transfer learning has recently become the dominant paradigm of machine learning.
It remains unclear how to develop models that specialise towards multiple tasks without incurring negative interference.
Modular deep learning has emerged as a promising solution to these challenges.
arXiv Detail & Related papers (2023-02-22T18:11:25Z) - Lorentz group equivariant autoencoders [6.858459233149096]
Lorentz group autoencoder (LGAE)
We develop an autoencoder model equivariant with respect to the proper, orthochronous Lorentz group $mathrmSO+(2,1)$, with a latent space living in the representations of the group.
We present our architecture and several experimental results on jets at the LHC and find it outperforms graph and convolutional neural network baseline models on several compression, reconstruction, and anomaly detection metrics.
arXiv Detail & Related papers (2022-12-14T17:19:46Z) - Learning the nonlinear dynamics of soft mechanical metamaterials with
graph networks [3.609538870261841]
We propose a machine learning approach to study the dynamics of soft mechanical metamaterials.
The proposed approach can significantly reduce the computational cost when compared to direct numerical simulation.
arXiv Detail & Related papers (2022-02-24T00:20:28Z) - Equivariant vector field network for many-body system modeling [65.22203086172019]
Equivariant Vector Field Network (EVFN) is built on a novel equivariant basis and the associated scalarization and vectorization layers.
We evaluate our method on predicting trajectories of simulated Newton mechanics systems with both full and partially observed data.
arXiv Detail & Related papers (2021-10-26T14:26:25Z) - Physics-Integrated Variational Autoencoders for Robust and Interpretable
Generative Modeling [86.9726984929758]
We focus on the integration of incomplete physics models into deep generative models.
We propose a VAE architecture in which a part of the latent space is grounded by physics.
We demonstrate generative performance improvements over a set of synthetic and real-world datasets.
arXiv Detail & Related papers (2021-02-25T20:28:52Z) - Learning with Density Matrices and Random Features [44.98964870180375]
A density matrix describes the statistical state of a quantum system.
It is a powerful formalism to represent both the quantum and classical uncertainty of quantum systems.
This paper explores how density matrices can be used as a building block for machine learning models.
arXiv Detail & Related papers (2021-02-08T17:54:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.