Machine Learning on generalized Complete Intersection Calabi-Yau
Manifolds
- URL: http://arxiv.org/abs/2209.10157v1
- Date: Wed, 21 Sep 2022 07:30:07 GMT
- Title: Machine Learning on generalized Complete Intersection Calabi-Yau
Manifolds
- Authors: Wei Cui, Xin Gao and Juntao Wang
- Abstract summary: Generalized Complete Intersection Calabi-Yau Manifold (gCICY) is a new construction of Calabi-Yau manifold.
In this paper, we try to make some progress in this direction using neural network.
- Score: 16.923362862181445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generalized Complete Intersection Calabi-Yau Manifold (gCICY) is a new
construction of Calabi-Yau manifolds established recently. However, the
generation of new gCICYs using standard algebraic method is very laborious. Due
to this complexity, the number of gCICYs and their classification still remain
unknown. In this paper, we try to make some progress in this direction using
neural network. The results showed that our trained models can have a high
precision on the existing type $(1,1)$ and type $(2,1)$ gCICYs in the
literature. Moreover, They can achieve a $97\%$ precision in predicting new
gCICY which is generated differently from those used for training and testing.
This shows that machine learning could be an effective method to classify and
generate new gCICY.
Related papers
- Robustly Learning Monotone Generalized Linear Models via Data Augmentation [37.42736399673992]
We give the first-time algorithm that achieves a constant-factor approximation for textitany monotone Lipschitz activation.
Our work resolves a well-known open problem, by developing a robust counterpart to the classical GLMtron algorithm.
arXiv Detail & Related papers (2025-02-12T17:59:21Z) - Gradients of Functions of Large Matrices [18.361820028457718]
We show how to differentiate workhorses of numerical linear algebra efficiently.
We derive previously unknown adjoint systems for Lanczos and Arnoldi iterations, implement them in JAX, and show that the resulting code can compete with Diffrax.
All this is achieved without any problem-specific code optimisation.
arXiv Detail & Related papers (2024-05-27T15:39:45Z) - Continual Zero-Shot Learning through Semantically Guided Generative
Random Walks [56.65465792750822]
We address the challenge of continual zero-shot learning where unseen information is not provided during training, by leveraging generative modeling.
We propose our learning algorithm that employs a novel semantically guided Generative Random Walk (GRW) loss.
Our algorithm achieves state-of-the-art performance on AWA1, AWA2, CUB, and SUN datasets, surpassing existing CZSL methods by 3-7%.
arXiv Detail & Related papers (2023-08-23T18:10:12Z) - Structural Credit Assignment with Coordinated Exploration [0.0]
Methods aimed at improving structural credit assignment can generally be classified into two categories.
We propose the use of Boltzmann machines or a recurrent network for coordinated exploration.
Experimental results demonstrate that coordinated exploration significantly exceeds independent exploration in training speed.
arXiv Detail & Related papers (2023-07-25T04:55:45Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Pretraining Without Attention [114.99187017618408]
This work explores pretraining without attention by using recent advances in sequence routing based on state-space models (SSMs)
BiGS is able to match BERT pretraining accuracy on GLUE and can be extended to long-form pretraining of 4096 tokens without approximation.
arXiv Detail & Related papers (2022-12-20T18:50:08Z) - Fast variable selection makes scalable Gaussian process BSS-ANOVA a
speedy and accurate choice for tabular and time series regression [0.0]
Gaussian processes (GPs) are non-parametric regression engines with a long history.
One of a number of scalable GP approaches is the Karhunen-Lo'eve (KL) decomposed kernel BSS-ANOVA, developed in 2009.
A new method of forward variable selection, quickly and effectively limits the number of terms, yielding a method with competitive accuracies.
arXiv Detail & Related papers (2022-05-26T23:41:43Z) - Deep metric learning improves lab of origin prediction of genetically
engineered plasmids [63.05016513788047]
Genetic engineering attribution (GEA) is the ability to make sequence-lab associations.
We propose a method, based on metric learning, that ranks the most likely labs-of-origin.
We are able to extract key signatures in plasmid sequences for particular labs, allowing for an interpretable examination of the model's outputs.
arXiv Detail & Related papers (2021-11-24T16:29:03Z) - Machine learning for complete intersection Calabi-Yau manifolds: a
methodological study [0.0]
We revisit the question of predicting Hodge numbers $h1,1$ and $h2,1$ of complete Calabi-Yau intersections using machine learning (ML)
We obtain 97% (resp. 99%) accuracy for $h1,1$ using a neural network inspired by the Inception model for the old dataset, using only 30% (resp. 70%) of the data for training.
For the new one, a simple linear regression leads to almost 100% accuracy with 30% of the data for training.
arXiv Detail & Related papers (2020-07-30T19:43:49Z) - Explore and Exploit with Heterotic Line Bundle Models [0.0]
We use deep reinforcement learning to explore a class of heterotic $SU(5)$ GUT models constructed from line bundle sums.
We perform several experiments where A3C agents are trained to search for such models.
arXiv Detail & Related papers (2020-03-10T15:49:33Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z) - Lipreading using Temporal Convolutional Networks [57.41253104365274]
Current model for recognition of isolated words in-the-wild consists of a residual network and Bi-directional Gated Recurrent Unit layers.
We address the limitations of this model and we propose changes which further improve its performance.
Our proposed model results in an absolute improvement of 1.2% and 3.2%, respectively, in these datasets.
arXiv Detail & Related papers (2020-01-23T17:49:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.