TorchNTK: A Library for Calculation of Neural Tangent Kernels of PyTorch
Models
- URL: http://arxiv.org/abs/2205.12372v1
- Date: Tue, 24 May 2022 21:27:58 GMT
- Title: TorchNTK: A Library for Calculation of Neural Tangent Kernels of PyTorch
Models
- Authors: Andrew Engel, Zhichao Wang, Anand D. Sarwate, Sutanay Choudhury, Tony
Chiang
- Abstract summary: We introduce torchNTK, a python library to calculate the empirical neural tangent kernel (NTK) of neural network models in the PyTorch framework.
A feature of the library is that we expose the user to layerwise NTK components, and show that in some regimes a layerwise calculation is more memory efficient.
- Score: 16.30276204466139
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce torchNTK, a python library to calculate the empirical neural
tangent kernel (NTK) of neural network models in the PyTorch framework. We
provide an efficient method to calculate the NTK of multilayer perceptrons. We
compare the explicit differentiation implementation against autodifferentiation
implementations, which have the benefit of extending the utility of the library
to any architecture supported by PyTorch, such as convolutional networks. A
feature of the library is that we expose the user to layerwise NTK components,
and show that in some regimes a layerwise calculation is more memory efficient.
We conduct preliminary experiments to demonstrate use cases for the software
and probe the NTK.
Related papers
- Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Efficient Dataset Distillation Using Random Feature Approximation [109.07737733329019]
We propose a novel algorithm that uses a random feature approximation (RFA) of the Neural Network Gaussian Process (NNGP) kernel.
Our algorithm provides at least a 100-fold speedup over KIP and can run on a single GPU.
Our new method, termed an RFA Distillation (RFAD), performs competitively with KIP and other dataset condensation algorithms in accuracy over a range of large-scale datasets.
arXiv Detail & Related papers (2022-10-21T15:56:13Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - TSFEDL: A Python Library for Time Series Spatio-Temporal Feature
Extraction and Prediction using Deep Learning (with Appendices on Detailed
Network Architectures and Experimental Cases of Study) [9.445070013080601]
The TSFE library is built upon a set offlow+Keras and PyTorch modules under the AGPLv3 license.
The performance validation of the architectures included in this proposal confirms the usefulness of this Python package.
arXiv Detail & Related papers (2022-06-07T10:58:33Z) - Scaling Neural Tangent Kernels via Sketching and Random Features [53.57615759435126]
Recent works report that NTK regression can outperform finitely-wide neural networks trained on small-scale datasets.
We design a near input-sparsity time approximation algorithm for NTK, by sketching the expansions of arc-cosine kernels.
We show that a linear regressor trained on our CNTK features matches the accuracy of exact CNTK on CIFAR-10 dataset while achieving 150x speedup.
arXiv Detail & Related papers (2021-06-15T04:44:52Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z) - TorchRadon: Fast Differentiable Routines for Computed Tomography [0.0]
The TorchRadon library is designed to help researchers working on CT problems to combine deep learning and model-based approaches.
Compared to the existing Astra Toolbox, TorchRadon is up to 125 faster.
Because of its speed and GPU support, TorchRadon can also be effectively used as a fast backend for the implementation of iterative algorithms.
arXiv Detail & Related papers (2020-09-29T09:20:22Z) - Captum: A unified and generic model interpretability library for PyTorch [49.72749684393332]
We introduce a novel, unified, open-source model interpretability library for PyTorch.
The library contains generic implementations of a number of gradient and perturbation-based attribution algorithms.
It can be used for both classification and non-classification models.
arXiv Detail & Related papers (2020-09-16T18:57:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.