Operating critical machine learning models in resource constrained
regimes
- URL: http://arxiv.org/abs/2303.10181v2
- Date: Sun, 4 Feb 2024 09:11:08 GMT
- Title: Operating critical machine learning models in resource constrained
regimes
- Authors: Raghavendra Selvan, Julian Sch\"on, Erik B Dam
- Abstract summary: We investigate the trade-off between resource consumption and performance of deep learning models.
Deep learning models are used in critical settings such as in clinics.
- Score: 0.18416014644193066
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The accelerated development of machine learning methods, primarily deep
learning, are causal to the recent breakthroughs in medical image analysis and
computer aided intervention. The resource consumption of deep learning models
in terms of amount of training data, compute and energy costs are known to be
massive. These large resource costs can be barriers in deploying these models
in clinics, globally. To address this, there are cogent efforts within the
machine learning community to introduce notions of resource efficiency. For
instance, using quantisation to alleviate memory consumption. While most of
these methods are shown to reduce the resource utilisation, they could come at
a cost in performance. In this work, we probe into the trade-off between
resource consumption and performance, specifically, when dealing with models
that are used in critical settings such as in clinics.
Related papers
- ssProp: Energy-Efficient Training for Convolutional Neural Networks with Scheduled Sparse Back Propagation [4.77407121905745]
Back-propagation (BP) is a major source of computational expense during training deep learning models.
We propose a general, energy-efficient convolution module that can be seamlessly integrated into any deep learning architecture.
arXiv Detail & Related papers (2024-08-22T17:22:59Z) - Multimodal Deep Learning for Low-Resource Settings: A Vector Embedding Alignment Approach for Healthcare Applications [3.2549142515720044]
We advocate for leveraging vector embeddings to enable flexible and efficient computational methodologies.
Our paper investigates the efficiency of using vector embeddings from single-modal foundation models and multi-modal Vision-Language Models.
We propose a simple yet effective inference-time method to enhance performance by aligning image-text embeddings.
arXiv Detail & Related papers (2024-06-02T01:13:01Z) - Equity through Access: A Case for Small-scale Deep Learning [6.474195280156721]
Large-scale resources have been used to train progressively larger models which are resource intensive in terms of compute, data, energy, and carbon emissions.
These costs are becoming a new type of entry barrier to researchers and practitioners with limited access to resources at such scale.
In this work, we take a comprehensive look at the landscape of existing DL models for vision tasks and demonstrate their usefulness in settings where resources are limited.
arXiv Detail & Related papers (2024-03-19T09:17:18Z) - Computation-efficient Deep Learning for Computer Vision: A Survey [121.84121397440337]
Deep learning models have reached or even exceeded human-level performance in a range of visual perception tasks.
Deep learning models usually demand significant computational resources, leading to impractical power consumption, latency, or carbon emissions in real-world scenarios.
New research focus is computationally efficient deep learning, which strives to achieve satisfactory performance while minimizing the computational cost during inference.
arXiv Detail & Related papers (2023-08-27T03:55:28Z) - Learnability with Time-Sharing Computational Resource Concerns [65.268245109828]
We present a theoretical framework that takes into account the influence of computational resources in learning theory.
This framework can be naturally applied to stream learning where the incoming data streams can be potentially endless.
It may also provide a theoretical perspective for the design of intelligent supercomputing operating systems.
arXiv Detail & Related papers (2023-05-03T15:54:23Z) - On Efficient Training of Large-Scale Deep Learning Models: A Literature
Review [90.87691246153612]
The field of deep learning has witnessed significant progress, particularly in computer vision (CV), natural language processing (NLP), and speech.
The use of large-scale models trained on vast amounts of data holds immense promise for practical applications.
With the increasing demands on computational capacity, a comprehensive summarization on acceleration techniques of training deep learning models is still much anticipated.
arXiv Detail & Related papers (2023-04-07T11:13:23Z) - Learnware: Small Models Do Big [69.88234743773113]
The prevailing big model paradigm, which has achieved impressive results in natural language processing and computer vision applications, has not yet addressed those issues, whereas becoming a serious source of carbon emissions.
This article offers an overview of the learnware paradigm, which attempts to enable users not need to build machine learning models from scratch, with the hope of reusing small models to do things even beyond their original purposes.
arXiv Detail & Related papers (2022-10-07T15:55:52Z) - Advancing Reacting Flow Simulations with Data-Driven Models [50.9598607067535]
Key to effective use of machine learning tools in multi-physics problems is to couple them to physical and computer models.
The present chapter reviews some of the open opportunities for the application of data-driven reduced-order modeling of combustion systems.
arXiv Detail & Related papers (2022-09-05T16:48:34Z) - Benchmarking Resource Usage for Efficient Distributed Deep Learning [10.869092085691687]
We conduct over 3,400 experiments training an array of deep networks representing various domains/tasks.
We fit power law models that describe how training time scales with available compute resources and energy constraints.
arXiv Detail & Related papers (2022-01-28T21:24:15Z) - Bayesian active learning for production, a systematic study and a
reusable library [85.32971950095742]
In this paper, we analyse the main drawbacks of current active learning techniques.
We do a systematic study on the effects of the most common issues of real-world datasets on the deep active learning process.
We derive two techniques that can speed up the active learning loop such as partial uncertainty sampling and larger query size.
arXiv Detail & Related papers (2020-06-17T14:51:11Z) - Resource-Efficient Neural Networks for Embedded Systems [23.532396005466627]
We provide an overview of the current state of the art of machine learning techniques.
We focus on resource-efficient inference based on deep neural networks (DNNs), the predominant machine learning models of the past decade.
We substantiate our discussion with experiments on well-known benchmark data sets using compression techniques.
arXiv Detail & Related papers (2020-01-07T14:17:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.