Trends in Energy Estimates for Computing in AI/Machine Learning
Accelerators, Supercomputers, and Compute-Intensive Applications
- URL: http://arxiv.org/abs/2210.17331v1
- Date: Wed, 12 Oct 2022 16:14:33 GMT
- Title: Trends in Energy Estimates for Computing in AI/Machine Learning
Accelerators, Supercomputers, and Compute-Intensive Applications
- Authors: Sadasivan Shankar, Albert Reuther
- Abstract summary: We examine the computational energy requirements of different systems driven by the geometrical scaling law.
We show that energy efficiency due to geometrical scaling is slowing down.
At the application level, general-purpose AI-ML methods can be computationally energy intensive.
- Score: 3.2634122554914
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We examine the computational energy requirements of different systems driven
by the geometrical scaling law, and increasing use of Artificial Intelligence
or Machine Learning (AI-ML) over the last decade. With more scientific and
technology applications based on data-driven discovery, machine learning
methods, especially deep neural networks, have become widely used. In order to
enable such applications, both hardware accelerators and advanced AI-ML methods
have led to the introduction of new architectures, system designs, algorithms,
and software. Our analysis of energy trends indicates three important
observations: 1) Energy efficiency due to geometrical scaling is slowing down;
2) The energy efficiency at the bit-level does not translate into efficiency at
the instruction-level, or at the system-level for a variety of systems,
especially for large-scale AI-ML accelerators or supercomputers; 3) At the
application level, general-purpose AI-ML methods can be computationally energy
intensive, off-setting the gains in energy from geometrical scaling and special
purpose accelerators. Further, our analysis provides specific pointers for
integrating energy efficiency with performance analysis for enabling
high-performance and sustainable computing in the future.
Related papers
- Machine Learning Insides OptVerse AI Solver: Design Principles and
Applications [74.67495900436728]
We present a comprehensive study on the integration of machine learning (ML) techniques into Huawei Cloud's OptVerse AI solver.
We showcase our methods for generating complex SAT and MILP instances utilizing generative models that mirror multifaceted structures of real-world problem.
We detail the incorporation of state-of-the-art parameter tuning algorithms which markedly elevate solver performance.
arXiv Detail & Related papers (2024-01-11T15:02:15Z) - Energy Estimates Across Layers of Computing: From Devices to Large-Scale
Applications in Machine Learning for Natural Language Processing, Scientific
Computing, and Cryptocurrency Mining [0.0]
Estimates of energy usage in layers of computing from devices to algorithms have been determined and analyzed.
Three large-scale computing applications such as Artificial Intelligence (AI)/Machine Learning for Natural Language Processing, Scientific Simulations, and Cryptocurrency Mining have been estimated.
arXiv Detail & Related papers (2023-10-11T14:14:05Z) - A Comprehensive Performance Study of Large Language Models on Novel AI
Accelerators [2.88634411143577]
Large language models (LLMs) are being considered as a promising approach to address some of the challenging problems.
Specialized AI accelerator hardware systems have recently become available for accelerating AI applications.
arXiv Detail & Related papers (2023-10-06T21:55:57Z) - Computation-efficient Deep Learning for Computer Vision: A Survey [121.84121397440337]
Deep learning models have reached or even exceeded human-level performance in a range of visual perception tasks.
Deep learning models usually demand significant computational resources, leading to impractical power consumption, latency, or carbon emissions in real-world scenarios.
New research focus is computationally efficient deep learning, which strives to achieve satisfactory performance while minimizing the computational cost during inference.
arXiv Detail & Related papers (2023-08-27T03:55:28Z) - Towards Implementing Energy-aware Data-driven Intelligence for Smart
Health Applications on Mobile Platforms [4.648824029505978]
On-device deep learning frameworks are proficient in utilizing computing resources in mobile platforms seamlessly.
However, energy resources in a mobile device are typically limited.
We introduce a new framework through an energy-aware, adaptive model comprehension and realization.
arXiv Detail & Related papers (2023-02-01T15:34:24Z) - Deep learning applied to computational mechanics: A comprehensive
review, state of the art, and the classics [77.34726150561087]
Recent developments in artificial neural networks, particularly deep learning (DL), are reviewed in detail.
Both hybrid and pure machine learning (ML) methods are discussed.
History and limitations of AI are recounted and discussed, with particular attention at pointing out misstatements or misconceptions of the classics.
arXiv Detail & Related papers (2022-12-18T02:03:00Z) - Precise Energy Consumption Measurements of Heterogeneous Artificial
Intelligence Workloads [0.534434568021034]
We present measurements of the energy consumption of two typical applications of deep learning models on different types of compute nodes.
One advantage of our approach is that the information on energy consumption is available to all users of the supercomputer.
arXiv Detail & Related papers (2022-12-03T21:40:55Z) - Reconfigurable Intelligent Surface Assisted Mobile Edge Computing with
Heterogeneous Learning Tasks [53.1636151439562]
Mobile edge computing (MEC) provides a natural platform for AI applications.
We present an infrastructure to perform machine learning tasks at an MEC with the assistance of a reconfigurable intelligent surface (RIS)
Specifically, we minimize the learning error of all participating users by jointly optimizing transmit power of mobile users, beamforming vectors of the base station, and the phase-shift matrix of the RIS.
arXiv Detail & Related papers (2020-12-25T07:08:50Z) - A Survey on Large-scale Machine Learning [67.6997613600942]
Machine learning can provide deep insights into data, allowing machines to make high-quality predictions.
Most sophisticated machine learning approaches suffer from huge time costs when operating on large-scale data.
Large-scale Machine Learning aims to learn patterns from big data with comparable performance efficiently.
arXiv Detail & Related papers (2020-08-10T06:07:52Z) - Spiking Neural Networks Hardware Implementations and Challenges: a
Survey [53.429871539789445]
Spiking Neural Networks are cognitive algorithms mimicking neuron and synapse operational principles.
We present the state of the art of hardware implementations of spiking neural networks.
We discuss the strategies employed to leverage the characteristics of these event-driven algorithms at the hardware level.
arXiv Detail & Related papers (2020-05-04T13:24:00Z) - Resource-Efficient Neural Networks for Embedded Systems [23.532396005466627]
We provide an overview of the current state of the art of machine learning techniques.
We focus on resource-efficient inference based on deep neural networks (DNNs), the predominant machine learning models of the past decade.
We substantiate our discussion with experiments on well-known benchmark data sets using compression techniques.
arXiv Detail & Related papers (2020-01-07T14:17:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.