Streaming Encoding Algorithms for Scalable Hyperdimensional Computing
- URL: http://arxiv.org/abs/2209.09868v2
- Date: Wed, 21 Sep 2022 00:45:56 GMT
- Title: Streaming Encoding Algorithms for Scalable Hyperdimensional Computing
- Authors: Anthony Thomas, Behnam Khaleghi, Gopi Krishna Jha, Sanjoy Dasgupta,
Nageen Himayat, Ravi Iyer, Nilesh Jain, and Tajana Rosing
- Abstract summary: Hyperdimensional computing (HDC) is a paradigm for data representation and learning originating in computational neuroscience.
In this work, we explore a family of streaming encoding techniques based on hashing.
We show formally that these methods enjoy comparable guarantees on performance for learning applications while being substantially more efficient than existing alternatives.
- Score: 12.829102171258882
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Hyperdimensional computing (HDC) is a paradigm for data representation and
learning originating in computational neuroscience. HDC represents data as
high-dimensional, low-precision vectors which can be used for a variety of
information processing tasks like learning or recall. The mapping to
high-dimensional space is a fundamental problem in HDC, and existing methods
encounter scalability issues when the input data itself is high-dimensional. In
this work, we explore a family of streaming encoding techniques based on
hashing. We show formally that these methods enjoy comparable guarantees on
performance for learning applications while being substantially more efficient
than existing alternatives. We validate these results experimentally on a
popular high-dimensional classification problem and show that our approach
easily scales to very large data sets.
Related papers
- A Weighted K-Center Algorithm for Data Subset Selection [70.49696246526199]
Subset selection is a fundamental problem that can play a key role in identifying smaller portions of the training data.
We develop a novel factor 3-approximation algorithm to compute subsets based on the weighted sum of both k-center and uncertainty sampling objective functions.
arXiv Detail & Related papers (2023-12-17T04:41:07Z) - An Encoding Framework for Binarized Images using HyperDimensional
Computing [0.0]
This article proposes a novel light-weight approach to encode binarized images that preserves similarity of patterns at nearby locations.
The method reaches an accuracy of 97.35% on the test set for the MNIST data set and 84.12% for the Fashion-MNIST data set.
arXiv Detail & Related papers (2023-12-01T09:34:28Z) - In search of the most efficient and memory-saving visualization of high
dimensional data [0.0]
We argue that the visualization of multidimensional data is well approximated of two-directed embedding of undimensional nearest-neighbor graphs.
Existing reduction methods are too slow and do not allow interactive manipulation.
We show that high-quality embeddings are produced with minimal time and memory complexity.
arXiv Detail & Related papers (2023-02-27T20:56:13Z) - HDTorch: Accelerating Hyperdimensional Computing with GP-GPUs for Design
Space Exploration [4.783565770657063]
We introduce HDTorch, an open-source, PyTorch-based HDC library with extensions for hypervector operations.
We analyze four HDC benchmark datasets in terms of accuracy, runtime, and memory consumption.
We perform the first-ever HD training and inference analysis of the entirety of the CHB-MIT EEG epilepsy database.
arXiv Detail & Related papers (2022-06-09T19:46:08Z) - An Extension to Basis-Hypervectors for Learning from Circular Data in
Hyperdimensional Computing [62.997667081978825]
Hyperdimensional Computing (HDC) is a computation framework based on properties of high-dimensional random spaces.
We present a study on basis-hypervector sets, which leads to practical contributions to HDC in general.
We introduce a method to learn from circular data, an important type of information never before addressed in machine learning with HDC.
arXiv Detail & Related papers (2022-05-16T18:04:55Z) - CvS: Classification via Segmentation For Small Datasets [52.821178654631254]
This paper presents CvS, a cost-effective classifier for small datasets that derives the classification labels from predicting the segmentation maps.
We evaluate the effectiveness of our framework on diverse problems showing that CvS is able to achieve much higher classification results compared to previous methods when given only a handful of examples.
arXiv Detail & Related papers (2021-10-29T18:41:15Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Classification using Hyperdimensional Computing: A Review [16.329917143918028]
This paper introduces the background of HD computing, and reviews the data representation, data transformation, and similarity measurement.
Evaluations indicate that HD computing shows great potential in addressing problems using data in the form of letters, signals and images.
arXiv Detail & Related papers (2020-04-19T23:51:44Z) - Learnable Subspace Clustering [76.2352740039615]
We develop a learnable subspace clustering paradigm to efficiently solve the large-scale subspace clustering problem.
The key idea is to learn a parametric function to partition the high-dimensional subspaces into their underlying low-dimensional subspaces.
To the best of our knowledge, this paper is the first work to efficiently cluster millions of data points among the subspace clustering methods.
arXiv Detail & Related papers (2020-04-09T12:53:28Z) - Auto-Encoding Twin-Bottleneck Hashing [141.5378966676885]
This paper proposes an efficient and adaptive code-driven graph.
It is updated by decoding in the context of an auto-encoder.
Experiments on benchmarked datasets clearly show the superiority of our framework over the state-of-the-art hashing methods.
arXiv Detail & Related papers (2020-02-27T05:58:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.