Heterotic String Model Building with Monad Bundles and Reinforcement
Learning
- URL: http://arxiv.org/abs/2108.07316v1
- Date: Mon, 16 Aug 2021 19:04:19 GMT
- Title: Heterotic String Model Building with Monad Bundles and Reinforcement
Learning
- Authors: Andrei Constantin, Thomas R. Harvey, Andre Lukas
- Abstract summary: We study heterotic SO GUT models on Calabi-Yau three-folds with monad bundles.
We show that reinforcement learning can be used successfully to explore monad bundles.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We use reinforcement learning as a means of constructing string
compactifications with prescribed properties. Specifically, we study heterotic
SO(10) GUT models on Calabi-Yau three-folds with monad bundles, in search of
phenomenologically promising examples. Due to the vast number of bundles and
the sparseness of viable choices, methods based on systematic scanning are not
suitable for this class of models. By focusing on two specific manifolds with
Picard numbers two and three, we show that reinforcement learning can be used
successfully to explore monad bundles. Training can be accomplished with
minimal computing resources and leads to highly efficient policy networks. They
produce phenomenologically promising states for nearly 100% of episodes and
within a small number of steps. In this way, hundreds of new candidate standard
models are found.
Related papers
- DiffuMatch: Category-Agnostic Spectral Diffusion Priors for Robust Non-rigid Shape Matching [53.39693288324375]
We show that both in-network regularization and functional map training can be replaced with data-driven methods.<n>We first train a generative model of functional maps in the spectral domain using score-based generative modeling.<n>We then exploit the resulting model to promote the structural properties of ground truth functional maps on new shape collections.
arXiv Detail & Related papers (2025-07-31T16:44:54Z) - SPaR: Self-Play with Tree-Search Refinement to Improve Instruction-Following in Large Language Models [88.29990536278167]
We introduce SPaR, a self-play framework integrating tree-search self-refinement to yield valid and comparable preference pairs.
Our experiments show that a LLaMA3-8B model, trained over three iterations guided by SPaR, surpasses GPT-4-Turbo on the IFEval benchmark without losing general capabilities.
arXiv Detail & Related papers (2024-12-16T09:47:43Z) - HM3: Heterogeneous Multi-Class Model Merging [0.0]
We explore training-free model merging techniques to consolidate auxiliary guard-rail models into a single, multi-functional model.
We propose Heterogeneous Multi-Class Model Merging (HM3) as a simple technique for merging multi-class classifiers with heterogeneous label spaces.
We report promising results for merging BERT-based guard models, some of which attain an average F1-score higher than the source models while reducing the inference time by up to 44%.
arXiv Detail & Related papers (2024-09-27T22:42:45Z) - Dynamic Online Ensembles of Basis Expansions [0.0]
We show how to use random feature approximations to achieve scalable, online ensembling of dynamic models.
We propose a novel method to ensemble static and dynamic models together.
arXiv Detail & Related papers (2024-05-02T15:09:59Z) - CodeGen2: Lessons for Training LLMs on Programming and Natural Languages [116.74407069443895]
We unify encoder and decoder-based models into a single prefix-LM.
For learning methods, we explore the claim of a "free lunch" hypothesis.
For data distributions, the effect of a mixture distribution and multi-epoch training of programming and natural languages on model performance is explored.
arXiv Detail & Related papers (2023-05-03T17:55:25Z) - Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - Class-Incremental Learning with Strong Pre-trained Models [97.84755144148535]
Class-incremental learning (CIL) has been widely studied under the setting of starting from a small number of classes (base classes)
We explore an understudied real-world setting of CIL that starts with a strong model pre-trained on a large number of base classes.
Our proposed method is robust and generalizes to all analyzed CIL settings.
arXiv Detail & Related papers (2022-04-07T17:58:07Z) - Low-Rank Constraints for Fast Inference in Structured Models [110.38427965904266]
This work demonstrates a simple approach to reduce the computational and memory complexity of a large class of structured models.
Experiments with neural parameterized structured models for language modeling, polyphonic music modeling, unsupervised grammar induction, and video modeling show that our approach matches the accuracy of standard models at large state spaces.
arXiv Detail & Related papers (2022-01-08T00:47:50Z) - Deep Learning Hamiltonian Monte Carlo [0.6554326244334867]
We generalize the Hamiltonian Monte Carlo algorithm with a stack of neural network layers.
We demonstrate that our model is able to successfully mix between modes of different topologies.
arXiv Detail & Related papers (2021-05-07T17:50:18Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - A Multi-Semantic Metapath Model for Large Scale Heterogeneous Network
Representation Learning [52.83948119677194]
We propose a multi-semantic metapath (MSM) model for large scale heterogeneous representation learning.
Specifically, we generate multi-semantic metapath-based random walks to construct the heterogeneous neighborhood to handle the unbalanced distributions.
We conduct systematical evaluations for the proposed framework on two challenging datasets: Amazon and Alibaba.
arXiv Detail & Related papers (2020-07-19T22:50:20Z) - Machine Learning String Standard Models [0.0]
unsupervised and supervised learning are considered.
We find that, for a fixed compactification manifold, relatively small neural networks are capable of distinguishing consistent line bundle models with the correct gauge group.
Learning non-topological properties, specifically the number of Higgs multiplets, turns out to be more difficult, but is possible using sizeable networks and feature-enhanced data sets.
arXiv Detail & Related papers (2020-03-30T11:14:14Z) - Explore and Exploit with Heterotic Line Bundle Models [0.0]
We use deep reinforcement learning to explore a class of heterotic $SU(5)$ GUT models constructed from line bundle sums.
We perform several experiments where A3C agents are trained to search for such models.
arXiv Detail & Related papers (2020-03-10T15:49:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.