ChebNet: Efficient and Stable Constructions of Deep Neural Networks with
Rectified Power Units via Chebyshev Approximations
- URL: http://arxiv.org/abs/1911.05467v3
- Date: Fri, 1 Dec 2023 08:37:27 GMT
- Title: ChebNet: Efficient and Stable Constructions of Deep Neural Networks with
Rectified Power Units via Chebyshev Approximations
- Authors: Shanshan Tang and Bo Li and Haijun Yu
- Abstract summary: We propose a new and more stable way to construct RePU deep neural networks based on Chebyshev approximations.
The approximation of smooth functions by ChebNets is no worse than the approximation by deep RePU nets using power series approach.
- Score: 6.0889567811100385
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In a previous study [B. Li, S. Tang and H. Yu, Commun. Comput. Phy.
27(2):379-411, 2020], it is shown that deep neural networks built with
rectified power units (RePU) as activation functions can give better
approximation for sufficient smooth functions than those built with rectified
linear units, by converting polynomial approximations using power series into
deep neural networks with optimal complexity and no approximation error.
However, in practice, power series approximations are not easy to obtain due to
the associated stability issue. In this paper, we propose a new and more stable
way to construct RePU deep neural networks based on Chebyshev polynomial
approximations. By using a hierarchical structure of Chebyshev polynomial
approximation in frequency domain, we obtain efficient and stable deep neural
network construction, which we call ChebNet. The approximation of smooth
functions by ChebNets is no worse than the approximation by deep RePU nets
using power series. On the same time, ChebNets are much more stable. Numerical
results show that the constructed ChebNets can be further fine-tuned to obtain
much better results than those obtained by tuning deep RePU nets constructed by
power series approach. As spectral accuracy is hard to obtain by direct
training of deep neural networks, ChebNets provide a practical way to obtain
spectral accuracy, it is expected to be useful in real applications that
require efficient approximations of smooth functions.
Related papers
- Approximation with Random Shallow ReLU Networks with Applications to Model Reference Adaptive Control [0.0]
We show that ReLU networks with randomly generated weights and biases achieve $L_infty$ error of $O(m-1/2)$ with high probability.
We show how the result can be used to get approximations of required accuracy in a model reference adaptive control application.
arXiv Detail & Related papers (2024-03-25T19:39:17Z) - Bayesian Inference Accelerator for Spiking Neural Networks [3.145754107337963]
spiking neural networks (SNNs) have the potential to reduce computational area and power.
In this work, we demonstrate an optimization framework for developing and implementing efficient Bayesian SNNs in hardware.
We demonstrate accuracies comparable to Bayesian binary networks with full-precision Bernoulli parameters, while requiring up to $25times$ less spikes.
arXiv Detail & Related papers (2024-01-27T16:27:19Z) - A Generalization of Continuous Relaxation in Structured Pruning [0.3277163122167434]
Trends indicate that deeper and larger neural networks with an increasing number of parameters achieve higher accuracy than smaller neural networks.
We generalize structured pruning with algorithms for network augmentation, pruning, sub-network collapse and removal.
The resulting CNN executes efficiently on GPU hardware without computationally expensive sparse matrix operations.
arXiv Detail & Related papers (2023-08-28T14:19:13Z) - Provable Data Subset Selection For Efficient Neural Network Training [73.34254513162898]
We introduce the first algorithm to construct coresets for emphRBFNNs, i.e., small weighted subsets that approximate the loss of the input data on any radial basis function network.
We then perform empirical evaluations on function approximation and dataset subset selection on popular network architectures and data sets.
arXiv Detail & Related papers (2023-03-09T10:08:34Z) - Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach [64.23331120621118]
This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
arXiv Detail & Related papers (2022-08-08T03:13:24Z) - Benefits of Overparameterized Convolutional Residual Networks: Function
Approximation under Smoothness Constraint [48.25573695787407]
We prove that large ConvResNets can not only approximate a target function in terms of function value, but also exhibit sufficient first-order smoothness.
Our theory partially justifies the benefits of using deep and wide networks in practice.
arXiv Detail & Related papers (2022-06-09T15:35:22Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Lipschitz constant estimation of Neural Networks via sparse polynomial
optimization [47.596834444042685]
LiPopt is a framework for computing increasingly tighter upper bounds on the Lipschitz constant of neural networks.
We show how to use the sparse connectivity of a network, to significantly reduce the complexity.
We conduct experiments on networks with random weights as well as networks trained on MNIST.
arXiv Detail & Related papers (2020-04-18T18:55:02Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.