LC-NAS: Latency Constrained Neural Architecture Search for Point Cloud
Networks
- URL: http://arxiv.org/abs/2008.10309v1
- Date: Mon, 24 Aug 2020 10:30:21 GMT
- Title: LC-NAS: Latency Constrained Neural Architecture Search for Point Cloud
Networks
- Authors: Guohao Li, Mengmeng Xu, Silvio Giancola, Ali Thabet, Bernard Ghanem
- Abstract summary: LC-NAS is able to find state-of-the-art architectures for point cloud classification with minimal computational cost.
We show how our searched architectures achieve any desired latency with a reasonably low drop in accuracy.
- Score: 73.78551758828294
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Point cloud architecture design has become a crucial problem for 3D deep
learning. Several efforts exist to manually design architectures with high
accuracy in point cloud tasks such as classification, segmentation, and
detection. Recent progress in automatic Neural Architecture Search (NAS)
minimizes the human effort in network design and optimizes high performing
architectures. However, these efforts fail to consider important factors such
as latency during inference. Latency is of high importance in time critical
applications like self-driving cars, robot navigation, and mobile applications,
that are generally bound by the available hardware. In this paper, we introduce
a new NAS framework, dubbed LC-NAS, where we search for point cloud
architectures that are constrained to a target latency. We implement a novel
latency constraint formulation to trade-off between accuracy and latency in our
architecture search. Contrary to previous works, our latency loss guarantees
that the final network achieves latency under a specified target value. This is
crucial when the end task is to be deployed in a limited hardware setting.
Extensive experiments show that LC-NAS is able to find state-of-the-art
architectures for point cloud classification in ModelNet40 with minimal
computational cost. We also show how our searched architectures achieve any
desired latency with a reasonably low drop in accuracy. Finally, we show how
our searched architectures easily transfer to a different task, part
segmentation on PartNet, where we achieve state-of-the-art results while
lowering latency by a factor of 10.
Related papers
- PNAS-MOT: Multi-Modal Object Tracking with Pareto Neural Architecture Search [64.28335667655129]
Multiple object tracking is a critical task in autonomous driving.
As tracking accuracy improves, neural networks become increasingly complex, posing challenges for their practical application in real driving scenarios due to the high level of latency.
In this paper, we explore the use of the neural architecture search (NAS) methods to search for efficient architectures for tracking, aiming for low real-time latency while maintaining relatively high accuracy.
arXiv Detail & Related papers (2024-03-23T04:18:49Z) - Search-time Efficient Device Constraints-Aware Neural Architecture
Search [6.527454079441765]
Deep learning techniques like computer vision and natural language processing can be computationally expensive and memory-intensive.
We automate the construction of task-specific deep learning architectures optimized for device constraints through Neural Architecture Search (NAS)
We present DCA-NAS, a principled method of fast neural network architecture search that incorporates edge-device constraints.
arXiv Detail & Related papers (2023-07-10T09:52:28Z) - Inference Latency Prediction at the Edge [0.3974789827371669]
State-of-the-art neural architectures (NAs) are typically designed through Neural Architecture Search (NAS) to identify NAs with good tradeoffs between accuracy and efficiency.
Since measuring the latency of a huge set of candidate architectures during NAS is not scalable, approaches are needed for predicting end-to-end inference latency on mobile devices.
We propose a latency prediction framework which addresses these challenges by developing operation-wise latency predictors.
arXiv Detail & Related papers (2022-10-06T00:46:06Z) - MAPLE-X: Latency Prediction with Explicit Microprocessor Prior Knowledge [87.41163540910854]
Deep neural network (DNN) latency characterization is a time-consuming process.
We propose MAPLE-X which extends MAPLE by incorporating explicit prior knowledge of hardware devices and DNN architecture latency.
arXiv Detail & Related papers (2022-05-25T11:08:20Z) - NAS-FCOS: Efficient Search for Object Detection Architectures [113.47766862146389]
We propose an efficient method to obtain better object detectors by searching for the feature pyramid network (FPN) and the prediction head of a simple anchor-free object detector.
With carefully designed search space, search algorithms, and strategies for evaluating network quality, we are able to find top-performing detection architectures within 4 days using 8 V100 GPUs.
arXiv Detail & Related papers (2021-10-24T12:20:04Z) - ISyNet: Convolutional Neural Networks design for AI accelerator [0.0]
Current state-of-the-art architectures are found with neural architecture search (NAS) taking model complexity into account.
We propose a measure of hardware efficiency of neural architecture search space - matrix efficiency measure (MEM); a search space comprising of hardware-efficient operations; a latency-aware scaling method.
We show the advantage of the designed architectures for the NPU devices on ImageNet and the generalization ability for the downstream classification and detection tasks.
arXiv Detail & Related papers (2021-09-04T20:57:05Z) - FLASH: Fast Neural Architecture Search with Hardware Optimization [7.263481020106725]
Neural architecture search (NAS) is a promising technique to design efficient and high-performance deep neural networks (DNNs)
This paper proposes FLASH, a very fast NAS methodology that co-optimizes the DNN accuracy and performance on a real hardware platform.
arXiv Detail & Related papers (2021-08-01T23:46:48Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - S3NAS: Fast NPU-aware Neural Architecture Search Methodology [2.607400740040335]
We present a fast NPU-aware NAS methodology, called S3NAS, to find a CNN architecture with higher accuracy than the existing ones.
We are able to find a network in 3 hours using TPUv3, which shows 82.72% top-1 accuracy on ImageNet with 11.66 ms latency.
arXiv Detail & Related papers (2020-09-04T04:45:50Z) - TF-NAS: Rethinking Three Search Freedoms of Latency-Constrained
Differentiable Neural Architecture Search [85.96350089047398]
We propose Three-Freedom NAS (TF-NAS) to achieve both good classification accuracy and precise latency constraint.
Experiments on ImageNet demonstrate the effectiveness of TF-NAS. Particularly, our searched TF-NAS-A obtains 76.9% top-1 accuracy, achieving state-of-the-art results with less latency.
arXiv Detail & Related papers (2020-08-12T13:44:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.