On Computing the Hyperparameter of Extreme Learning Machines: Algorithm
and Application to Computational PDEs, and Comparison with Classical and
High-Order Finite Elements
- URL: http://arxiv.org/abs/2110.14121v1
- Date: Wed, 27 Oct 2021 02:05:26 GMT
- Title: On Computing the Hyperparameter of Extreme Learning Machines: Algorithm
and Application to Computational PDEs, and Comparison with Classical and
High-Order Finite Elements
- Authors: Suchuan Dong, Jielin Yang
- Abstract summary: We consider the use of extreme learning machines (ELM) for computational partial differential equations (PDE)
In ELM the hidden-layer coefficients in the neural network are assigned to random values generated on $[-R_m,R_m]$ and fixed.
We present a method for computing the optimal value of $R_m$ based on the differential evolution algorithm.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the use of extreme learning machines (ELM) for computational
partial differential equations (PDE). In ELM the hidden-layer coefficients in
the neural network are assigned to random values generated on $[-R_m,R_m]$ and
fixed, where $R_m$ is a user-provided constant, and the output-layer
coefficients are trained by a linear or nonlinear least squares computation. We
present a method for computing the optimal value of $R_m$ based on the
differential evolution algorithm. The presented method enables us to illuminate
the characteristics of the optimal $R_m$ for two types of ELM configurations:
(i) Single-Rm-ELM, in which a single $R_m$ is used for generating the random
coefficients in all the hidden layers, and (ii) Multi-Rm-ELM, in which multiple
$R_m$ constants are involved with each used for generating the random
coefficients of a different hidden layer. We adopt the optimal $R_m$ from this
method and also incorporate other improvements into the ELM implementation. In
particular, here we compute all the differential operators involving the output
fields of the last hidden layer by a forward-mode auto-differentiation, as
opposed to the reverse-mode auto-differentiation in a previous work. These
improvements significantly reduce the network training time and enhance the ELM
performance. We systematically compare the computational performance of the
current improved ELM with that of the finite element method (FEM), both the
classical second-order FEM and the high-order FEM with Lagrange elements of
higher degrees, for solving a number of linear and nonlinear PDEs. It is shown
that the current improved ELM far outperforms the classical FEM. Its
computational performance is comparable to that of the high-order FEM for
smaller problem sizes, and for larger problem sizes the ELM markedly
outperforms the high-order FEM.
Related papers
- Sample-efficient Learning of Infinite-horizon Average-reward MDPs with General Function Approximation [53.17668583030862]
We study infinite-horizon average-reward Markov decision processes (AMDPs) in the context of general function approximation.
We propose a novel algorithmic framework named Local-fitted Optimization with OPtimism (LOOP)
We show that LOOP achieves a sublinear $tildemathcalO(mathrmpoly(d, mathrmsp(V*)) sqrtTbeta )$ regret, where $d$ and $beta$ correspond to AGEC and log-covering number of the hypothesis class respectively
arXiv Detail & Related papers (2024-04-19T06:24:22Z) - Global optimization of MPS in quantum-inspired numerical analysis [0.0]
The study focuses on the search for the lowest eigenstates of a Hamiltonian equation.
Five algorithms are introduced: imaginary-time evolution, steepest gradient descent, an improved descent, an implicitly restarted Arnoldi method, and density matrix renormalization group (DMRG) optimization.
arXiv Detail & Related papers (2023-03-16T16:03:51Z) - A distribution-free mixed-integer optimization approach to hierarchical modelling of clustered and longitudinal data [0.0]
We introduce an innovative algorithm that evaluates cluster effects for new data points, thereby increasing the robustness and precision of this model.
The inferential and predictive efficacy of this approach is further illustrated through its application in student scoring and protein expression.
arXiv Detail & Related papers (2023-02-06T23:34:51Z) - Fast Computation of Optimal Transport via Entropy-Regularized Extragradient Methods [75.34939761152587]
Efficient computation of the optimal transport distance between two distributions serves as an algorithm that empowers various applications.
This paper develops a scalable first-order optimization-based method that computes optimal transport to within $varepsilon$ additive accuracy.
arXiv Detail & Related papers (2023-01-30T15:46:39Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Log-based Sparse Nonnegative Matrix Factorization for Data
Representation [55.72494900138061]
Nonnegative matrix factorization (NMF) has been widely studied in recent years due to its effectiveness in representing nonnegative data with parts-based representations.
We propose a new NMF method with log-norm imposed on the factor matrices to enhance the sparseness.
A novel column-wisely sparse norm, named $ell_2,log$-(pseudo) norm, is proposed to enhance the robustness of the proposed method.
arXiv Detail & Related papers (2022-04-22T11:38:10Z) - Deep Learning-Based Power Control for Uplink Cell-Free Massive MIMO
Systems [31.06830781747216]
Instead of using supervised learning, the proposed method relies on unsupervised learning, in which optimal power allocations are not required to be known.
A deep neural network (DNN) is trained to learn the map between fading coefficients and power coefficients within short time.
It is interesting to note that the spectral efficiency of mMIMO systems with the proposed method outperforms previous optimization methods for max-min optimization.
arXiv Detail & Related papers (2021-10-18T03:48:54Z) - Efficient Matrix-Free Approximations of Second-Order Information, with
Applications to Pruning and Optimization [16.96639526117016]
We investigate matrix-free, linear-time approaches for estimating Inverse-Hessian Vector Products (IHVPs)
These algorithms yield state-of-the-art results for network pruning and optimization with lower computational overhead relative to existing second-order methods.
arXiv Detail & Related papers (2021-07-07T17:01:34Z) - Self-supervised Symmetric Nonnegative Matrix Factorization [82.59905231819685]
Symmetric nonnegative factor matrix (SNMF) has demonstrated to be a powerful method for data clustering.
Inspired by ensemble clustering that aims to seek better clustering results, we propose self-supervised SNMF (S$3$NMF)
We take advantage of the sensitivity to code characteristic of SNMF, without relying on any additional information.
arXiv Detail & Related papers (2021-03-02T12:47:40Z) - Efficient Learning of Generative Models via Finite-Difference Score
Matching [111.55998083406134]
We present a generic strategy to efficiently approximate any-order directional derivative with finite difference.
Our approximation only involves function evaluations, which can be executed in parallel, and no gradient computations.
arXiv Detail & Related papers (2020-07-07T10:05:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.