Accelerating Neural Architecture Exploration Across Modalities Using
Genetic Algorithms
- URL: http://arxiv.org/abs/2202.12934v1
- Date: Fri, 25 Feb 2022 20:01:36 GMT
- Title: Accelerating Neural Architecture Exploration Across Modalities Using
Genetic Algorithms
- Authors: Daniel Cummings, Sharath Nittur Sridhar, Anthony Sarah, Maciej Szankin
- Abstract summary: We show how genetic algorithms can be paired with lightly trained objective predictors in an iterative cycle to accelerate multi-objective architectural exploration.
NAS research efforts have centered around computer vision tasks and only recently have other modalities, such as the rapidly growing field of natural language processing, been investigated in depth.
- Score: 5.620334754517149
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural architecture search (NAS), the study of automating the discovery of
optimal deep neural network architectures for tasks in domains such as computer
vision and natural language processing, has seen rapid growth in the machine
learning research community. While there have been many recent advancements in
NAS, there is still a significant focus on reducing the computational cost
incurred when validating discovered architectures by making search more
efficient. Evolutionary algorithms, specifically genetic algorithms, have a
history of usage in NAS and continue to gain popularity versus other
optimization approaches as a highly efficient way to explore the architecture
objective space. Most NAS research efforts have centered around computer vision
tasks and only recently have other modalities, such as the rapidly growing
field of natural language processing, been investigated in depth. In this work,
we show how genetic algorithms can be paired with lightly trained objective
predictors in an iterative cycle to accelerate multi-objective architectural
exploration in a way that works in the modalities of both machine translation
and image classification.
Related papers
- A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - Evolution and Efficiency in Neural Architecture Search: Bridging the Gap Between Expert Design and Automated Optimization [1.7385545432331702]
The paper provides a comprehensive overview of Neural Architecture Search.
It emphasizes its evolution from manual design to automated, computationally-driven approaches.
It highlights its application across various domains, including medical imaging and natural language processing.
arXiv Detail & Related papers (2024-02-11T18:27:29Z) - An Approach for Efficient Neural Architecture Search Space Definition [0.0]
We propose a novel cell-based hierarchical search space, easy to comprehend and manipulate.
The objectives of the proposed approach are to optimize the search-time and to be general enough to handle most of state of the art CNN architectures.
arXiv Detail & Related papers (2023-10-25T08:07:29Z) - Neural Architecture Search: Insights from 1000 Papers [50.27255667347091]
We provide an organized and comprehensive guide to neural architecture search.
We give a taxonomy of search spaces, algorithms, and speedup techniques.
We discuss resources such as benchmarks, best practices, other surveys, and open-source libraries.
arXiv Detail & Related papers (2023-01-20T18:47:24Z) - HiveNAS: Neural Architecture Search using Artificial Bee Colony
Optimization [0.0]
In this study, we evaluate the viability of Artificial Bee Colony optimization for Neural Architecture Search.
Our proposed framework, HiveNAS, outperforms existing state-of-the-art Swarm Intelligence-based NAS frameworks in a fraction of the time.
arXiv Detail & Related papers (2022-11-18T14:11:47Z) - Neural Architecture Search for Dense Prediction Tasks in Computer Vision [74.9839082859151]
Deep learning has led to a rising demand for neural network architecture engineering.
neural architecture search (NAS) aims at automatically designing neural network architectures in a data-driven manner rather than manually.
NAS has become applicable to a much wider range of problems in computer vision.
arXiv Detail & Related papers (2022-02-15T08:06:50Z) - Learning Interpretable Models Through Multi-Objective Neural
Architecture Search [0.9990687944474739]
We propose a framework to optimize for both task performance and "introspectability," a surrogate metric for aspects of interpretability.
We demonstrate that jointly optimizing for task error and introspectability leads to more disentangled and debuggable architectures that perform within error.
arXiv Detail & Related papers (2021-12-16T05:50:55Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language
Processing [12.02718579660613]
We step outside the computer vision domain by leveraging the language modeling task, which is the core of natural language processing (NLP)
We have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it.
We have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation.
arXiv Detail & Related papers (2020-06-12T12:19:06Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.