Guidelines for enhancing data locality in selected machine learning
algorithms
- URL: http://arxiv.org/abs/2001.03000v1
- Date: Thu, 9 Jan 2020 14:16:40 GMT
- Title: Guidelines for enhancing data locality in selected machine learning
algorithms
- Authors: Imen Chakroun and Tom Vander Aa and Thomas J. Ashby
- Abstract summary: We analyze one of the means to increase the performances of machine learning algorithms which is exploiting data locality.
Repeated data access can be seen as redundancy in data movement.
This work also identifies some of the opportunities for avoiding these redundancies by directly reusing results.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To deal with the complexity of the new bigger and more complex generation of
data, machine learning (ML) techniques are probably the first and foremost
used. For ML algorithms to produce results in a reasonable amount of time, they
need to be implemented efficiently. In this paper, we analyze one of the means
to increase the performances of machine learning algorithms which is exploiting
data locality. Data locality and access patterns are often at the heart of
performance issues in computing systems due to the use of certain hardware
techniques to improve performance. Altering the access patterns to increase
locality can dramatically increase performance of a given algorithm. Besides,
repeated data access can be seen as redundancy in data movement. Similarly,
there can also be redundancy in the repetition of calculations. This work also
identifies some of the opportunities for avoiding these redundancies by
directly reusing computation results. We start by motivating why and how a more
efficient implementation can be achieved by exploiting reuse in the memory
hierarchy of modern instruction set processors. Next we document the
possibilities of such reuse in some selected machine learning algorithms.
Related papers
- From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models [63.188607839223046]
This survey focuses on the benefits of scaling compute during inference.
We explore three areas under a unified mathematical formalism: token-level generation algorithms, meta-generation algorithms, and efficient generation.
arXiv Detail & Related papers (2024-06-24T17:45:59Z) - Accelerating Machine Learning Algorithms with Adaptive Sampling [1.539582851341637]
It is often sufficient to substitute computationally intensive subroutines with a special kind of randomized counterparts that results in almost no degradation in quality.
This thesis demonstrates that it is often sufficient, instead, to substitute computationally intensive subroutines with a special kind of randomized counterparts that results in almost no degradation in quality.
arXiv Detail & Related papers (2023-09-25T15:25:59Z) - A Survey From Distributed Machine Learning to Distributed Deep Learning [0.356008609689971]
Distributed machine learning has been proposed, which involves distributing the data and algorithm across several machines.
We divide these algorithms in classification and clustering (traditional machine learning), deep learning and deep reinforcement learning groups.
Based on the investigation of the mentioned algorithms, we highlighted the limitations that should be addressed in future research.
arXiv Detail & Related papers (2023-07-11T13:06:42Z) - Performance and Energy Consumption of Parallel Machine Learning
Algorithms [0.0]
Machine learning models have achieved remarkable success in various real-world applications.
Model training in machine learning requires large-scale data sets and multiple iterations before it can work properly.
Parallelization of training algorithms is a common strategy to speed up the process of training.
arXiv Detail & Related papers (2023-05-01T13:04:39Z) - Benchmarking Processor Performance by Multi-Threaded Machine Learning
Algorithms [0.0]
In this paper, I will make a performance comparison of multi-threaded machine learning clustering algorithms.
I will be working on Linear Regression, Random Forest, and K-Nearest Neighbors to determine the performance characteristics of the algorithms.
arXiv Detail & Related papers (2021-09-11T13:26:58Z) - Providing Meaningful Data Summarizations Using Examplar-based Clustering
in Industry 4.0 [67.80123919697971]
We show, that our GPU implementation provides speedups of up to 72x using single-precision and up to 452x using half-precision compared to conventional CPU algorithms.
We apply our algorithm to real-world data from injection molding manufacturing processes and discuss how found summaries help with steering this specific process to cut costs and reduce the manufacturing of bad parts.
arXiv Detail & Related papers (2021-05-25T15:55:14Z) - Evolving Reinforcement Learning Algorithms [186.62294652057062]
We propose a method for meta-learning reinforcement learning algorithms.
The learned algorithms are domain-agnostic and can generalize to new environments not seen during training.
We highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games.
arXiv Detail & Related papers (2021-01-08T18:55:07Z) - Hard-ODT: Hardware-Friendly Online Decision Tree Learning Algorithm and
System [17.55491405857204]
In the era of big data, traditional decision tree induction algorithms are not suitable for learning large-scale datasets.
We introduce a new quantile-based algorithm to improve the induction of the Hoeffding tree, one of the state-of-the-art online learning models.
We present Hard-ODT, a high-performance, hardware-efficient and scalable online decision tree learning system on a field-programmable gate array (FPGA) with system-level optimization techniques.
arXiv Detail & Related papers (2020-12-11T12:06:44Z) - A Survey on Large-scale Machine Learning [67.6997613600942]
Machine learning can provide deep insights into data, allowing machines to make high-quality predictions.
Most sophisticated machine learning approaches suffer from huge time costs when operating on large-scale data.
Large-scale Machine Learning aims to learn patterns from big data with comparable performance efficiently.
arXiv Detail & Related papers (2020-08-10T06:07:52Z) - Spiking Neural Networks Hardware Implementations and Challenges: a
Survey [53.429871539789445]
Spiking Neural Networks are cognitive algorithms mimicking neuron and synapse operational principles.
We present the state of the art of hardware implementations of spiking neural networks.
We discuss the strategies employed to leverage the characteristics of these event-driven algorithms at the hardware level.
arXiv Detail & Related papers (2020-05-04T13:24:00Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.