Bringing AI pipelines onto cloud-HPC: setting a baseline for accuracy of
COVID-19 AI diagnosis
- URL: http://arxiv.org/abs/2108.01033v1
- Date: Mon, 2 Aug 2021 16:45:00 GMT
- Title: Bringing AI pipelines onto cloud-HPC: setting a baseline for accuracy of
COVID-19 AI diagnosis
- Authors: Iacopo Colonnelli and Barbara Cantalupo and Concetto Spampinato and
Matteo Pennisi and Marco Aldinucci
- Abstract summary: We advocate the StreamFlow Management System as a crucial ingredient to define a parametric pipeline, called "CLAIRE COVID-19 Universal Pipeline"
This pipeline is able to explore the optimization methods to classify COVID-19 lung lesions from CT scans, compare them for accuracy, and set a performance baseline.
It requires a massive computing power, which is found in traditional HPC infrastructure thanks to the portability-by-design of pipelines designed with StreamFlow.
- Score: 5.313553229474047
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: HPC is an enabling platform for AI. The introduction of AI workloads in the
HPC applications basket has non-trivial consequences both on the way of
designing AI applications and on the way of providing HPC computing. This is
the leitmotif of the convergence between HPC and AI. The formalized definition
of AI pipelines is one of the milestones of HPC-AI convergence. If well
conducted, it allows, on the one hand, to obtain portable and scalable
applications. On the other hand, it is crucial for the reproducibility of
scientific pipelines. In this work, we advocate the StreamFlow Workflow
Management System as a crucial ingredient to define a parametric pipeline,
called "CLAIRE COVID-19 Universal Pipeline," which is able to explore the
optimization space of methods to classify COVID-19 lung lesions from CT scans,
compare them for accuracy, and therefore set a performance baseline. The
universal pipeline automatizes the training of many different Deep Neural
Networks (DNNs) and many different hyperparameters. It, therefore, requires a
massive computing power, which is found in traditional HPC infrastructure
thanks to the portability-by-design of pipelines designed with StreamFlow.
Using the universal pipeline, we identified a DNN reaching over 90% accuracy in
detecting COVID-19 lesions in CT scans.
Related papers
- Plug-and-Play Benchmarking of Reinforcement Learning Algorithms for Large-Scale Flow Control [61.155940786140455]
Reinforcement learning (RL) has shown promising results in active flow control (AFC)<n>Current AFC benchmarks rely on external computational fluid dynamics (CFD) solvers, are not fully differentiable, and provide limited 3D and multi-agent support.<n>We introduce FluidGym, the first standalone, fully differentiable benchmark suite for RL in AFC.
arXiv Detail & Related papers (2026-01-21T14:13:44Z) - Towards Interactive Lesion Segmentation in Whole-Body PET/CT with Promptable Models [9.254535645473311]
We present our submission to the autoPET/CT IV challenge.<n>We extend the framework with promptable capabilities by encoding user-provided foreground and background clicks as additional input channels.<n>Our ensemble of EDT-based models, trained with and without external data, achieves the strongest cross-validation performance.
arXiv Detail & Related papers (2025-08-29T14:49:58Z) - FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - GraphPipe: Improving Performance and Scalability of DNN Training with Graph Pipeline Parallelism [20.44114440511298]
Deep neural networks (DNNs) continue to grow rapidly in size, making them infeasible to train on a single device.
This paper presents a new pipeline-parallel scheme that partitions a DNN into pipeline stages whose dependencies are identified by a directed acyclic graph.
We also develop GraphPipe, a distributed system that exploits MME strategies to enable performant and scalable DNN training.
arXiv Detail & Related papers (2024-06-24T21:32:51Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - FPGA-QHAR: Throughput-Optimized for Quantized Human Action Recognition
on The Edge [0.6254873489691849]
This paper proposed an integrated end-to-end HAR scalable HW/SW accelerator co-design based on an enhanced 8-bit quantized Two-Stream SimpleNet-PyTorch CNN architecture.
Our development uses partially streaming dataflow architecture to achieve higher throughput versus network design and resource utilization trade-off.
Our proposed methodology achieved nearly 81% prediction accuracy with an approximately 24 FPS real-time inference throughput at 187MHz on ZCU104.
arXiv Detail & Related papers (2023-11-04T10:38:21Z) - Complexity-Driven CNN Compression for Resource-constrained Edge AI [1.6114012813668934]
We propose a novel and computationally efficient pruning pipeline by exploiting the inherent layer-level complexities of CNNs.
We define three modes of pruning, namely parameter-aware (PA), FLOPs-aware (FA), and memory-aware (MA), to introduce versatile compression of CNNs.
arXiv Detail & Related papers (2022-08-26T16:01:23Z) - Tiny-HR: Towards an interpretable machine learning pipeline for heart
rate estimation on edge devices [6.59529078336196]
This paper is a proof of concept, machine learning (ML) pipeline that extracts heart rate from pressure sensor data acquired on low-power edge devices.
The pipeline consists an upsampler neural network, a signal quality classifier, and a 1D-convolutional neural network optimized for efficient and accurate heart rate estimation.
The results indicate that the proposed ML and hybrid pipeline reduces energy and time per inference by 82% and 28% compared to traditional algorithms.
arXiv Detail & Related papers (2022-08-16T22:38:10Z) - Where Is My Training Bottleneck? Hidden Trade-Offs in Deep Learning
Preprocessing Pipelines [77.45213180689952]
Preprocessing pipelines in deep learning aim to provide sufficient data throughput to keep the training processes busy.
We introduce a new perspective on efficiently preparing datasets for end-to-end deep learning pipelines.
We obtain an increased throughput of 3x to 13x compared to an untuned system.
arXiv Detail & Related papers (2022-02-17T14:31:58Z) - An AI-based Domain-Decomposition Non-Intrusive Reduced-Order Model for
Extended Domains applied to Multiphase Flow in Pipes [0.0]
We present a new AI-based non-intrusive reduced-order model within a domain decomposition framework.
It is capable of making predictions for domains significantly larger than the domain used in training.
The framework is applied to multiphase slug flow in a horizontal pipe for which an AI-DDNIROM is trained on high-fidelity CFD simulations.
arXiv Detail & Related papers (2022-02-13T00:32:17Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge
Computing: A Contextual-Bandit Approach [81.5261621619557]
We propose an adaptive anomaly detection scheme with hierarchical edge computing (HEC)
We first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer.
Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network.
arXiv Detail & Related papers (2021-08-09T08:45:47Z) - Layer Pruning on Demand with Intermediate CTC [50.509073206630994]
We present a training and pruning method for ASR based on the connectionist temporal classification (CTC)
We show that a Transformer-CTC model can be pruned in various depth on demand, improving real-time factor from 0.005 to 0.002 on GPU.
arXiv Detail & Related papers (2021-06-17T02:40:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.