Third ArchEdge Workshop: Exploring the Design Space of Efficient Deep
Neural Networks
- URL: http://arxiv.org/abs/2011.10912v1
- Date: Sun, 22 Nov 2020 01:56:46 GMT
- Title: Third ArchEdge Workshop: Exploring the Design Space of Efficient Deep
Neural Networks
- Authors: Fuxun Yu, Dimitrios Stamoulis, Di Wang, Dimitrios Lymberopoulos, Xiang
Chen
- Abstract summary: This paper gives an overview of our ongoing work on the design space exploration of efficient deep neural networks (DNNs)
We cover two aspects: (1) static architecture design efficiency and (2) dynamic model execution efficiency.
We highlight several open questions that are poised to draw research attention in the next few years.
- Score: 14.195694804273801
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper gives an overview of our ongoing work on the design space
exploration of efficient deep neural networks (DNNs). Specifically, we cover
two aspects: (1) static architecture design efficiency and (2) dynamic model
execution efficiency. For static architecture design, different from existing
end-to-end hardware modeling assumptions, we conduct full-stack profiling at
the GPU core level to identify better accuracy-latency trade-offs for DNN
designs. For dynamic model execution, different from prior work that tackles
model redundancy at the DNN-channels level, we explore a new dimension of DNN
feature map redundancy to be dynamically traversed at runtime. Last, we
highlight several open questions that are poised to draw research attention in
the next few years.
Related papers
- Improvement of Spiking Neural Network with Bit Planes and Color Models [0.0]
Spiking neural network (SNN) has emerged as a promising paradigm in computational neuroscience and artificial intelligence.
We present a novel approach to enhance the performance of SNN for images through a new coding method that exploits bit plane representation.
arXiv Detail & Related papers (2024-09-28T15:52:49Z) - From Alexnet to Transformers: Measuring the Non-linearity of Deep Neural Networks with Affine Optimal Transport [32.39176908225668]
We introduce the concept of the non-linearity signature of DNN, the first theoretically sound solution for measuring the non-linearity of deep neural networks.
We provide extensive experimental results that highlight the practical usefulness of the proposed non-linearity signature.
arXiv Detail & Related papers (2023-10-17T17:50:22Z) - Flexible Channel Dimensions for Differentiable Architecture Search [50.33956216274694]
We propose a novel differentiable neural architecture search method with an efficient dynamic channel allocation algorithm.
We show that the proposed framework is able to find DNN architectures that are equivalent to previous methods in task accuracy and inference latency.
arXiv Detail & Related papers (2023-06-13T15:21:38Z) - Towards a General Purpose CNN for Long Range Dependencies in
$\mathrm{N}$D [49.57261544331683]
We propose a single CNN architecture equipped with continuous convolutional kernels for tasks on arbitrary resolution, dimensionality and length without structural changes.
We show the generality of our approach by applying the same CCNN to a wide set of tasks on sequential (1$mathrmD$) and visual data (2$mathrmD$)
Our CCNN performs competitively and often outperforms the current state-of-the-art across all tasks considered.
arXiv Detail & Related papers (2022-06-07T15:48:02Z) - Comparison Analysis of Traditional Machine Learning and Deep Learning
Techniques for Data and Image Classification [62.997667081978825]
The purpose of the study is to analyse and compare the most common machine learning and deep learning techniques used for computer vision 2D object classification tasks.
Firstly, we will present the theoretical background of the Bag of Visual words model and Deep Convolutional Neural Networks (DCNN)
Secondly, we will implement a Bag of Visual Words model, the VGG16 CNN Architecture.
arXiv Detail & Related papers (2022-04-11T11:34:43Z) - QuadraLib: A Performant Quadratic Neural Network Library for
Architecture Optimization and Design Exploration [31.488940932186246]
Quadratic Deep Neuron Networks (QDNNs) show better non-linearity and learning capability than the first-order DNNs.
Our design has good performance regarding prediction accuracy and computation consumption on multiple learning tasks.
arXiv Detail & Related papers (2022-04-01T18:06:54Z) - PaSca: a Graph Neural Architecture Search System under the Scalable
Paradigm [24.294196319217907]
Graph neural networks (GNNs) have achieved state-of-the-art performance in various graph-based tasks.
However, GNNs do not scale well to data size and message passing steps.
This paper proposes PasCa, a new paradigm and system that offers a principled approach to systemically construct and explore the design space for scalable GNNs.
arXiv Detail & Related papers (2022-03-01T17:26:50Z) - Sparse Flows: Pruning Continuous-depth Models [107.98191032466544]
We show that pruning improves generalization for neural ODEs in generative modeling.
We also show that pruning finds minimal and efficient neural ODE representations with up to 98% less parameters compared to the original network, without loss of accuracy.
arXiv Detail & Related papers (2021-06-24T01:40:17Z) - A Design Space Study for LISTA and Beyond [79.76740811464597]
In recent years, great success has been witnessed in building problem-specific deep networks from unrolling iterative algorithms.
This paper revisits the role of unrolling as a design approach for deep networks, to what extent its resulting special architecture is superior, and can we find better?
Using LISTA for sparse recovery as a representative example, we conduct the first thorough design space study for the unrolled models.
arXiv Detail & Related papers (2021-04-08T23:01:52Z) - Design Space for Graph Neural Networks [81.88707703106232]
We study the architectural design space for Graph Neural Networks (GNNs) which consists of 315,000 different designs over 32 different predictive tasks.
Our key results include: (1) A comprehensive set of guidelines for designing well-performing GNNs; (2) while best GNN designs for different tasks vary significantly, the GNN task space allows for transferring the best designs across different tasks; (3) models discovered using our design space achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-11-17T18:59:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.