Virtuoso: Video-based Intelligence for real-time tuning on SOCs
- URL: http://arxiv.org/abs/2112.13076v1
- Date: Fri, 24 Dec 2021 14:47:41 GMT
- Title: Virtuoso: Video-based Intelligence for real-time tuning on SOCs
- Authors: Jayoung Lee, PengCheng Wang, Ran Xu, Venkat Dasari, Noah Weston, Yin
Li, Saurabh Bagchi, and Somali Chaterji
- Abstract summary: Underlying Virtuoso is a multi-branch execution kernel capable of running at different operating points in the accuracy-energy-latency axes.
We benchmark 15 state-of-the-art or widely used protocols, including Faster R-CNN (FRCNN), YOLO v3, SSD, EfficientDet, SELSA, MEGA, REPP, FastAdapt, and our in-house adaptive variants of FRCNN+, YOLO+, SSD+, and EfficientDet+.
- Score: 24.086595996055074
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Efficient and adaptive computer vision systems have been proposed to make
computer vision tasks, such as image classification and object detection,
optimized for embedded or mobile devices. These solutions, quite recent in
their origin, focus on optimizing the model (a deep neural network, DNN) or the
system by designing an adaptive system with approximation knobs. In spite of
several recent efforts, we show that existing solutions suffer from two major
drawbacks. First, the system does not consider energy consumption of the models
while making a decision on which model to run. Second, the evaluation does not
consider the practical scenario of contention on the device, due to other
co-resident workloads. In this work, we propose an efficient and adaptive video
object detection system, Virtuoso, which is jointly optimized for accuracy,
energy efficiency, and latency. Underlying Virtuoso is a multi-branch execution
kernel that is capable of running at different operating points in the
accuracy-energy-latency axes, and a lightweight runtime scheduler to select the
best fit execution branch to satisfy the user requirement. To fairly compare
with Virtuoso, we benchmark 15 state-of-the-art or widely used protocols,
including Faster R-CNN (FRCNN), YOLO v3, SSD, EfficientDet, SELSA, MEGA, REPP,
FastAdapt, and our in-house adaptive variants of FRCNN+, YOLO+, SSD+, and
EfficientDet+ (our variants have enhanced efficiency for mobiles). With this
comprehensive benchmark, Virtuoso has shown superiority to all the above
protocols, leading the accuracy frontier at every efficiency level on NVIDIA
Jetson mobile GPUs. Specifically, Virtuoso has achieved an accuracy of 63.9%,
which is more than 10% higher than some of the popular object detection models,
FRCNN at 51.1%, and YOLO at 49.5%.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - Efficient Modulation for Vision Networks [122.1051910402034]
We propose efficient modulation, a novel design for efficient vision networks.
We demonstrate that the modulation mechanism is particularly well suited for efficient networks.
Our network can accomplish better trade-offs between accuracy and efficiency.
arXiv Detail & Related papers (2024-03-29T03:48:35Z) - EdgeYOLO: An Edge-Real-Time Object Detector [69.41688769991482]
This paper proposes an efficient, low-complexity and anchor-free object detector based on the state-of-the-art YOLO framework.
We develop an enhanced data augmentation method to effectively suppress overfitting during training, and design a hybrid random loss function to improve the detection accuracy of small objects.
Our baseline model can reach the accuracy of 50.6% AP50:95 and 69.8% AP50 in MS 2017 dataset, 26.4% AP50:95 and 44.8% AP50 in VisDrone 2019-DET dataset, and it meets real-time requirements (FPS>=30) on edge-computing device Nvidia
arXiv Detail & Related papers (2023-02-15T06:05:14Z) - Neural Nets with a Newton Conjugate Gradient Method on Multiple GPUs [0.0]
Training deep neural networks consumes increasing computational resource shares in many compute centers.
We introduce a novel second-order optimization method that requires the effect of the Hessian on a vector only.
We compare the proposed second-order method with two state-of-the-arts on five representative neural network problems.
arXiv Detail & Related papers (2022-08-03T12:38:23Z) - EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for
Mobile Vision Applications [68.35683849098105]
We introduce split depth-wise transpose attention (SDTA) encoder that splits input tensors into multiple channel groups.
Our EdgeNeXt model with 1.3M parameters achieves 71.2% top-1 accuracy on ImageNet-1K.
Our EdgeNeXt model with 5.6M parameters achieves 79.4% top-1 accuracy on ImageNet-1K.
arXiv Detail & Related papers (2022-06-21T17:59:56Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - PP-PicoDet: A Better Real-Time Object Detector on Mobile Devices [13.62426382827205]
PP-PicoDet family of real-time object detectors achieves superior performance on object detection for mobile devices.
Models achieve better trade-offs between accuracy and latency compared to other popular models.
arXiv Detail & Related papers (2021-11-01T12:53:17Z) - FADNet++: Real-Time and Accurate Disparity Estimation with Configurable
Networks [19.29846600092521]
FADNet++ is an efficient deep network for disparity estimation.
It can boost its accuracy with a fast model inference speed for real-time applications.
It achieves a new state-of-the-art result for the SceneFlow dataset.
arXiv Detail & Related papers (2021-10-06T08:50:33Z) - Toward Accurate Platform-Aware Performance Modeling for Deep Neural
Networks [0.17499351967216337]
We provide a machine learning-based method, PerfNetV2, which improves the accuracy of our previous work for modeling the neural network performance on a variety of GPU accelerators.
Given an application, the proposed method can be used to predict the inference time and training time of the convolutional neural networks used in the application.
Our case studies show that PerfNetV2 yields a mean absolute percentage error within 13.1% on LeNet, AlexNet, and VGG16 on NVIDIA GTX-1080Ti, while the error rate on a previous work published in ICBD 2018 could be as large as 200%.
arXiv Detail & Related papers (2020-12-01T01:42:23Z) - Scaling Up Deep Neural Network Optimization for Edge Inference [20.9711130126031]
Deep neural networks (DNNs) have been increasingly deployed on and integrated with edge devices, such as mobile phones, drones, robots and wearables.
To run DNN inference directly on edge devices (a.k.a. edge inference) with a satisfactory performance, optimizing the DNN design is crucial.
We propose two approaches to scaling up DNN optimization. In the first approach, we reuse the performance predictors built on a proxy device, and leverage the performance monotonicity to scale up the DNN optimization without re-building performance predictors for each different device.
In the second approach, we build scalable performance predictors that can estimate
arXiv Detail & Related papers (2020-09-01T07:47:22Z) - PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with
Pattern-based Weight Pruning [57.20262984116752]
We introduce a new dimension, fine-grained pruning patterns inside the coarse-grained structures, revealing a previously unknown point in design space.
With the higher accuracy enabled by fine-grained pruning patterns, the unique insight is to use the compiler to re-gain and guarantee high hardware efficiency.
arXiv Detail & Related papers (2020-01-01T04:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.