Particle Swarm Based Hyper-Parameter Optimization for Machine Learned
Interatomic Potentials
- URL: http://arxiv.org/abs/2101.00049v1
- Date: Thu, 31 Dec 2020 19:27:17 GMT
- Title: Particle Swarm Based Hyper-Parameter Optimization for Machine Learned
Interatomic Potentials
- Authors: Suresh Kondati Natarajan and Miguel A. Caro
- Abstract summary: Training non-empirical interatomic potential energy surfaces (PES) using machine learning (ML) approaches is becoming popular in molecular and materials research.
We propose a two-step optimization strategy in which the HPs related to the feature extraction stage are optimized first, followed by the optimization of the HPs in the training stage.
This strategy is computationally more efficient than optimizing all HPs at the same time by means of significantly reducing the number of models needed to be trained.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modeling non-empirical and highly flexible interatomic potential energy
surfaces (PES) using machine learning (ML) approaches is becoming popular in
molecular and materials research. Training an ML-PES is typically performed in
two stages: feature extraction and structure-property relationship modeling.
The feature extraction stage transforms atomic positions into a
symmetry-invariant mathematical representation. This representation can be
fine-tuned by adjusting on a set of so-called "hyper-parameters" (HPs).
Subsequently, an ML algorithm such as neural networks or Gaussian process
regression (GPR) is used to model the structure-PES relationship based on
another set of HPs. Choosing optimal values for the two sets of HPs is critical
to ensure the high quality of the resulting ML-PES model.
In this paper, we explore HP optimization strategies tailored for ML-PES
generation using a custom-coded parallel particle swarm optimizer (available
freely at https://github.com/suresh0807/PPSO.git). We employ the smooth overlap
of atomic positions (SOAP) descriptor in combination with GPR-based Gaussian
approximation potentials (GAP) and optimize HPs for four distinct systems: a
toy C dimer, amorphous carbon, $\alpha$-Fe, and small organic molecules (QM9
dataset). We propose a two-step optimization strategy in which the HPs related
to the feature extraction stage are optimized first, followed by the
optimization of the HPs in the training stage. This strategy is computationally
more efficient than optimizing all HPs at the same time by means of
significantly reducing the number of ML models needed to be trained to obtain
the optimal HPs. This approach can be trivially extended to other combinations
of descriptor and ML algorithm and brings us another step closer to fully
automated ML-PES generation.
Related papers
- AffineQuant: Affine Transformation Quantization for Large Language Models [58.45460102764]
Post-Training Quantization (PTQ) has emerged as a subject of considerable interest due to its compression efficiency and cost-effectiveness in the context of training.
Existing PTQ methods for Large-scale Language Models (LLMs) limit the optimization scope to scaling transformations between pre- and post-quantization weights.
In this paper, we advocate for the direct optimization using equivalent Affine transformations in PTQ (AffineQuant)
arXiv Detail & Related papers (2024-03-19T08:40:21Z) - Federated Conditional Stochastic Optimization [110.513884892319]
Conditional optimization has found in a wide range of machine learning tasks, such as in-variant learning tasks, AUPRC, andAML.
This paper proposes algorithms for distributed federated learning.
arXiv Detail & Related papers (2023-10-04T01:47:37Z) - Two-step hyperparameter optimization method: Accelerating hyperparameter
search by using a fraction of a training dataset [0.15420205433587747]
We present a two-step HPO method as a strategic solution to curbing computational demands and wait times.
We present our recent application of the two-step HPO method to the development of neural network emulators for aerosol activation.
arXiv Detail & Related papers (2023-02-08T02:38:26Z) - HOAX: A Hyperparameter Optimization Algorithm Explorer for Neural
Networks [0.0]
The bottleneck for trajectory-based methods to study photoinduced processes is still the huge number of electronic structure calculations.
We present an innovative solution, in which the amount of electronic structure calculations is drastically reduced, by employing machine learning algorithms and methods borrowed from the realm of artificial intelligence.
arXiv Detail & Related papers (2023-02-01T11:12:35Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Multi-objective hyperparameter optimization with performance uncertainty [62.997667081978825]
This paper presents results on multi-objective hyperparameter optimization with uncertainty on the evaluation of Machine Learning algorithms.
We combine the sampling strategy of Tree-structured Parzen Estimators (TPE) with the metamodel obtained after training a Gaussian Process Regression (GPR) with heterogeneous noise.
Experimental results on three analytical test functions and three ML problems show the improvement over multi-objective TPE and GPR.
arXiv Detail & Related papers (2022-09-09T14:58:43Z) - Towards Learning Universal Hyperparameter Optimizers with Transformers [57.35920571605559]
We introduce the OptFormer, the first text-based Transformer HPO framework that provides a universal end-to-end interface for jointly learning policy and function prediction.
Our experiments demonstrate that the OptFormer can imitate at least 7 different HPO algorithms, which can be further improved via its function uncertainty estimates.
arXiv Detail & Related papers (2022-05-26T12:51:32Z) - Towards Robust and Automatic Hyper-Parameter Tunning [39.04604349338802]
We introduce a new class of HPO method and explore how the low-rank factorization of intermediate layers of a convolutional network can be used to define an analytical response surface.
We quantify how this surface behaves as a surrogate to model performance and can be solved using a trust-region search algorithm, which we call autoHyper.
arXiv Detail & Related papers (2021-11-28T05:27:34Z) - HYPPO: A Surrogate-Based Multi-Level Parallelism Tool for Hyperparameter
Optimization [0.2844198651668139]
HYPPO uses adaptive surrogate models and accounts for uncertainty in model predictions to find accurate and reliable models that make robust predictions.
We demonstrate various software features on time-series prediction and image classification problems as well as a scientific application in computed tomography image reconstruction.
arXiv Detail & Related papers (2021-10-04T20:14:22Z) - Genealogical Population-Based Training for Hyperparameter Optimization [1.0514231683620516]
We experimentally demonstrate that our method cuts down by 2 to 3 times the computational cost required.
Our method is search-algorithm so that the inner search routine can be any search algorithm like TPE, GP, CMA or random search.
arXiv Detail & Related papers (2021-09-30T08:49:41Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.