Molecular Energy Learning Using Alternative Blackbox Matrix-Matrix
Multiplication Algorithm for Exact Gaussian Process
- URL: http://arxiv.org/abs/2109.09817v1
- Date: Mon, 20 Sep 2021 19:59:06 GMT
- Title: Molecular Energy Learning Using Alternative Blackbox Matrix-Matrix
Multiplication Algorithm for Exact Gaussian Process
- Authors: Jiace Sun, Lixue Cheng, Thomas F. Miller III
- Abstract summary: We present an application of the blackbox matrix-matrix multiplication (BBMM) algorithm to scale up the Gaussian Process (GP) training of molecular energies.
An alternative implementation of BBMM (AltBBMM) is also proposed to train more efficiently with the same accuracy and transferability.
The accuracy and transferability of both algorithms are examined on the benchmark of organic molecules with 7 and 13 heavy atoms.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present an application of the blackbox matrix-matrix multiplication (BBMM)
algorithm to scale up the Gaussian Process (GP) training of molecular energies
in the molecular-orbital based machine learning (MOB-ML) framework. An
alternative implementation of BBMM (AltBBMM) is also proposed to train more
efficiently (over four-fold speedup) with the same accuracy and transferability
as the original BBMM implementation. The training of MOB-ML was limited to 220
molecules, and BBMM and AltBBMM scale the training of MOB-ML up by over 30
times to 6500 molecules (more than a million pair energies). The accuracy and
transferability of both algorithms are examined on the benchmark datasets of
organic molecules with 7 and 13 heavy atoms. These lower-scaling
implementations of the GP preserve the state-of-the-art learning efficiency in
the low-data regime while extending it to the large-data regime with better
accuracy than other available machine learning works on molecular energies.
Related papers
- EPS-MoE: Expert Pipeline Scheduler for Cost-Efficient MoE Inference [49.94169109038806]
This paper introduces EPS-MoE, a novel expert pipeline scheduler for MoE.
Our results demonstrate an average 21% improvement in prefill throughput over existing parallel inference methods.
arXiv Detail & Related papers (2024-10-16T05:17:49Z) - DB-LLM: Accurate Dual-Binarization for Efficient LLMs [83.70686728471547]
Large language models (LLMs) have significantly advanced the field of natural language processing.
Existing ultra-low-bit quantization always causes severe accuracy drops.
We propose a novel Dual-Binarization method for LLMs, namely DB-LLM.
arXiv Detail & Related papers (2024-02-19T09:04:30Z) - A Multi-Grained Symmetric Differential Equation Model for Learning
Protein-Ligand Binding Dynamics [74.93549765488103]
In drug discovery, molecular dynamics simulation provides a powerful tool for predicting binding affinities, estimating transport properties, and exploring pocket sites.
We propose NeuralMD, the first machine learning surrogate that can facilitate numerical MD and provide accurate simulations in protein-ligand binding.
We show the efficiency and effectiveness of NeuralMD, with a 2000$times$ speedup over standard numerical MD simulation and outperforming all other ML approaches by up to 80% under the stability metric.
arXiv Detail & Related papers (2024-01-26T09:35:17Z) - RedMule: A Mixed-Precision Matrix-Matrix Operation Engine for Flexible
and Energy-Efficient On-Chip Linear Algebra and TinyML Training Acceleration [15.869673535117032]
Current training algorithms rely on floating-point matrix operations to meet the precision and dynamic range requirements.
RedMulE is a low-power specialized accelerator conceived for multi-precision floating-point General Matrix-Matrix Operations (GEMM-Ops) acceleration.
RedMulE achieves up to 58.5 GFLOPS and 117 GFLOPS for FP16 and FP8, respectively, with 99.4% utilization of the array of Computing Elements.
arXiv Detail & Related papers (2023-01-10T11:07:16Z) - Multi-Agent Automated Machine Learning [54.14038920246645]
We propose multi-agent automated machine learning (MA2ML) to handle joint optimization of modules in automated machine learning (AutoML)
MA2ML explicitly assigns credit to each agent according to its marginal contribution to enhance cooperation among modules, and incorporates off-policy learning to improve search efficiency.
Experiments show that MA2ML yields the state-of-the-art top-1 accuracy on ImageNet under constraints of computational cost.
arXiv Detail & Related papers (2022-10-17T13:32:59Z) - Molecular-orbital-based Machine Learning for Open-shell and
Multi-reference Systems with Kernel Addition Gaussian Process Regression [0.0]
We introduce a novel machine learning strategy, kernel addition Gaussian process regression (KA-GPR), in molecular-orbital-based machine learning (MOB-ML)
We learn the total correlation energies of general electronic structure theories for closed- and open-shell systems by introducing a machine learning strategy.
The learning efficiency of MOB-ML (KA-GPR) is the same as the original MOB-ML method for the smallest criegee molecule, which is a closed-shell molecule with multi-reference characters.
arXiv Detail & Related papers (2022-07-17T23:20:19Z) - Molecular Dipole Moment Learning via Rotationally Equivariant Gaussian
Process Regression with Derivatives in Molecular-orbital-based Machine
Learning [0.0]
This study extends the accurate and transferable molecular-orbital-based machine learning (MOB-ML) approach.
A molecular-orbital-based (MOB) pairwise decomposition of the correlation part of the dipole moment is applied.
The proposed problem setup, feature design, and ML algorithm are shown to provide highly-accurate models.
arXiv Detail & Related papers (2022-05-31T02:42:50Z) - Accurate Machine Learned Quantum-Mechanical Force Fields for
Biomolecular Simulations [51.68332623405432]
Molecular dynamics (MD) simulations allow atomistic insights into chemical and biological processes.
Recently, machine learned force fields (MLFFs) emerged as an alternative means to execute MD simulations.
This work proposes a general approach to constructing accurate MLFFs for large-scale molecular simulations.
arXiv Detail & Related papers (2022-05-17T13:08:28Z) - Accurate Molecular-Orbital-Based Machine Learning Energies via
Unsupervised Clustering of Chemical Space [0.0]
We introduce an unsupervised clustering algorithm to improve training efficiency and accuracy in predicting energies using molecular-orbital-based machine learning (MOB-ML)
This work determines clusters via the Gaussian mixture model (GMM) in an entirely automatic manner.
arXiv Detail & Related papers (2022-04-21T00:56:16Z) - NNP/MM: Accelerating molecular dynamics simulations with machine
learning potentials and molecular mechanic [38.50309739333058]
We introduce an optimized implementation of the hybrid method (NNP/MM), which combines neural network potentials (NNP) and molecular mechanics (MM)
This approach models a portion of the system, such as a small molecule, using NNP while employing MM for the remaining system to boost efficiency.
It has enabled us to increase the simulation speed by 5 times and achieve a combined sampling of one microsecond for each complex, marking the longest simulations ever reported for this class of simulation.
arXiv Detail & Related papers (2022-01-20T10:57:20Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.