The MIT Supercloud Workload Classification Challenge
- URL: http://arxiv.org/abs/2204.05839v2
- Date: Wed, 13 Apr 2022 18:31:04 GMT
- Title: The MIT Supercloud Workload Classification Challenge
- Authors: Benny J. Tang, Qiqi Chen, Matthew L. Weiss, Nathan Frey, Joseph
McDonald, David Bestor, Charles Yee, William Arcand, Chansup Byun, Daniel
Edelman, Matthew Hubbell, Michael Jones, Jeremy Kepner, Anna Klein, Adam
Michaleas, Peter Michaleas, Lauren Milechin, Julia Mullen, Andrew Prout,
Albert Reuther, Antonio Rosa, Andrew Bowne, Lindsey McEvoy, Baolin Li, Devesh
Tiwari, Vijay Gadepally, Siddharth Samsi
- Abstract summary: In this paper, we present a workload classification challenge based on the MIT Supercloud dataset.
The goal of this challenge is to foster algorithmic innovations in the analysis of compute workloads.
- Score: 10.458111248130944
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-Performance Computing (HPC) centers and cloud providers support an
increasingly diverse set of applications on heterogenous hardware. As
Artificial Intelligence (AI) and Machine Learning (ML) workloads have become an
increasingly larger share of the compute workloads, new approaches to optimized
resource usage, allocation, and deployment of new AI frameworks are needed. By
identifying compute workloads and their utilization characteristics, HPC
systems may be able to better match available resources with the application
demand. By leveraging datacenter instrumentation, it may be possible to develop
AI-based approaches that can identify workloads and provide feedback to
researchers and datacenter operators for improving operational efficiency. To
enable this research, we released the MIT Supercloud Dataset, which provides
detailed monitoring logs from the MIT Supercloud cluster. This dataset includes
CPU and GPU usage by jobs, memory usage, and file system logs. In this paper,
we present a workload classification challenge based on this dataset. We
introduce a labelled dataset that can be used to develop new approaches to
workload classification and present initial results based on existing
approaches. The goal of this challenge is to foster algorithmic innovations in
the analysis of compute workloads that can achieve higher accuracy than
existing methods. Data and code will be made publicly available via the
Datacenter Challenge website : https://dcc.mit.edu.
Related papers
- Federated Fine-Tuning of LLMs on the Very Edge: The Good, the Bad, the Ugly [62.473245910234304]
This paper takes a hardware-centric approach to explore how Large Language Models can be brought to modern edge computing systems.
We provide a micro-level hardware benchmark, compare the model FLOP utilization to a state-of-the-art data center GPU, and study the network utilization in realistic conditions.
arXiv Detail & Related papers (2023-10-04T20:27:20Z) - Asynchronous Parallel Incremental Block-Coordinate Descent for
Decentralized Machine Learning [55.198301429316125]
Machine learning (ML) is a key technique for big-data-driven modelling and analysis of massive Internet of Things (IoT) based intelligent and ubiquitous computing.
For fast-increasing applications and data amounts, distributed learning is a promising emerging paradigm since it is often impractical or inefficient to share/aggregate data.
This paper studies the problem of training an ML model over decentralized systems, where data are distributed over many user devices.
arXiv Detail & Related papers (2022-02-07T15:04:15Z) - SOLIS -- The MLOps journey from data acquisition to actionable insights [62.997667081978825]
In this paper we present a unified deployment pipeline and freedom-to-operate approach that supports all requirements while using basic cross-platform tensor framework and script language engines.
This approach however does not supply the needed procedures and pipelines for the actual deployment of machine learning capabilities in real production grade systems.
arXiv Detail & Related papers (2021-12-22T14:45:37Z) - Privacy-Preserving Serverless Edge Learning with Decentralized Small
Data [13.254530176359182]
Distributed training strategies have recently become a promising approach to ensure data privacy when training deep models.
This paper extends conventional serverless platforms with serverless edge learning architectures and provides an efficient distributed training framework from the networking perspective.
arXiv Detail & Related papers (2021-11-29T21:04:49Z) - The MIT Supercloud Dataset [3.375826083518709]
We introduce the MIT Supercloud dataset which aims to foster innovative AI/ML approaches to the analysis of large scale HPC and datacenter/cloud operations.
We provide detailed monitoring logs from the MIT Supercloud system, which include CPU and GPU usage by jobs, memory usage, file system logs, and physical monitoring data.
This paper discusses the details of the dataset, collection methodology, data availability, and discusses potential challenge problems being developed using this data.
arXiv Detail & Related papers (2021-08-04T13:06:17Z) - Confluence of Artificial Intelligence and High Performance Computing for
Accelerated, Scalable and Reproducible Gravitational Wave Detection [4.081122815035999]
We demonstrate how connecting DOE and NSF-sponsored cyberinfrastructure allows for new ways to publish machine learning models.
We then use this workflow to search for binary black hole gravitational wave signals in open source advanced LIGO data.
We find that using this workflow, an ensemble of four openly available deep learning models can be run on HAL and process the entire month of August 2017 of advanced LIGO data in just seven minutes.
arXiv Detail & Related papers (2020-12-15T19:00:29Z) - Cost-effective Machine Learning Inference Offload for Edge Computing [0.3149883354098941]
This paper proposes a novel offloading mechanism by leveraging installed-base on-premises (edge) computational resources.
The proposed mechanism allows the edge devices to offload heavy and compute-intensive workloads to edge nodes instead of using remote cloud.
arXiv Detail & Related papers (2020-12-07T21:11:02Z) - Faster Secure Data Mining via Distributed Homomorphic Encryption [108.77460689459247]
Homomorphic Encryption (HE) is receiving more and more attention recently for its capability to do computations over the encrypted field.
We propose a novel general distributed HE-based data mining framework towards one step of solving the scaling problem.
We verify the efficiency and effectiveness of our new framework by testing over various data mining algorithms and benchmark data-sets.
arXiv Detail & Related papers (2020-06-17T18:14:30Z) - Artificial Intelligence (AI)-Centric Management of Resources in Modern
Distributed Computing Systems [22.550075095184514]
Cloud Data Centres (DCS) are large scale, complex, heterogeneous, and distributed across multiple networks and geographical boundaries.
The Internet of Things (IoT)-driven applications are producing a huge amount of data that requires real-time processing and fast response.
Existing Resource Management Systems (RMS) rely on either static or solutions inadequate for such composite and dynamic systems.
arXiv Detail & Related papers (2020-06-09T06:54:07Z) - A Privacy-Preserving Distributed Architecture for
Deep-Learning-as-a-Service [68.84245063902908]
This paper introduces a novel distributed architecture for deep-learning-as-a-service.
It is able to preserve the user sensitive data while providing Cloud-based machine and deep learning services.
arXiv Detail & Related papers (2020-03-30T15:12:03Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.