Evaluating the Effectiveness of Efficient Neural Architecture Search for
Sentence-Pair Tasks
- URL: http://arxiv.org/abs/2010.04249v1
- Date: Thu, 8 Oct 2020 20:26:34 GMT
- Title: Evaluating the Effectiveness of Efficient Neural Architecture Search for
Sentence-Pair Tasks
- Authors: Ansel MacLaughlin, Jwala Dhamala, Anoop Kumar, Sriram Venkatapathy,
Ragav Venkatesan, Rahul Gupta
- Abstract summary: Neural Architecture Search (NAS) methods have recently achieved competitive or state-of-the-art (SOTA) performance on variety of natural language processing and computer vision tasks.
In this work, we explore the applicability of a SOTA NAS algorithm, Efficient Neural Architecture Search (ENAS) to two sentence pair tasks.
- Score: 14.963150544536203
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural Architecture Search (NAS) methods, which automatically learn entire
neural model or individual neural cell architectures, have recently achieved
competitive or state-of-the-art (SOTA) performance on variety of natural
language processing and computer vision tasks, including language modeling,
natural language inference, and image classification. In this work, we explore
the applicability of a SOTA NAS algorithm, Efficient Neural Architecture Search
(ENAS) (Pham et al., 2018) to two sentence pair tasks, paraphrase detection and
semantic textual similarity. We use ENAS to perform a micro-level search and
learn a task-optimized RNN cell architecture as a drop-in replacement for an
LSTM. We explore the effectiveness of ENAS through experiments on three
datasets (MRPC, SICK, STS-B), with two different models (ESIM, BiLSTM-Max), and
two sets of embeddings (Glove, BERT). In contrast to prior work applying ENAS
to NLP tasks, our results are mixed -- we find that ENAS architectures
sometimes, but not always, outperform LSTMs and perform similarly to random
architecture search.
Related papers
- A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - L$^{2}$NAS: Learning to Optimize Neural Architectures via
Continuous-Action Reinforcement Learning [23.25155249879658]
Differentiable architecture search (NAS) achieved remarkable results in deep neural network design.
We show that L$2$ achieves state-of-theart results on DART201 benchmark as well as NASS and Once-for-All search policies.
arXiv Detail & Related papers (2021-09-25T19:26:30Z) - Going Beyond Neural Architecture Search with Sampling-based Neural
Ensemble Search [31.059040393415003]
We present two novel sampling algorithms under our Neural Ensemble Search via Sampling (NESS) framework.
Our NESS algorithms are shown to be able to achieve improved performance in both classification and adversarial defense tasks.
arXiv Detail & Related papers (2021-09-06T15:18:37Z) - Generic Neural Architecture Search via Regression [27.78105839644199]
We propose a novel and generic neural architecture search (NAS) framework, termed Generic NAS (GenNAS)
GenNAS does not use task-specific labels but instead adopts textitregression on a set of manually designed synthetic signal bases for architecture evaluation.
We then propose an automatic task search to optimize the combination of synthetic signals using limited downstream-task-specific labels.
arXiv Detail & Related papers (2021-08-04T08:21:12Z) - Elastic Architecture Search for Diverse Tasks with Different Resources [87.23061200971912]
We study a new challenging problem of efficient deployment for diverse tasks with different resources, where the resource constraint and task of interest corresponding to a group of classes are dynamically specified at testing time.
Previous NAS approaches seek to design architectures for all classes simultaneously, which may not be optimal for some individual tasks.
We present a novel and general framework, called Elastic Architecture Search (EAS), permitting instant specializations at runtime for diverse tasks with various resource constraints.
arXiv Detail & Related papers (2021-08-03T00:54:27Z) - BiX-NAS: Searching Efficient Bi-directional Architecture for Medical
Image Segmentation [85.0444711725392]
We study a multi-scale upgrade of a bi-directional skip connected network and then automatically discover an efficient architecture by a novel two-phase Neural Architecture Search (NAS) algorithm, namely BiX-NAS.
Our proposed method reduces the network computational cost by sifting out ineffective multi-scale features at different levels and iterations.
We evaluate BiX-NAS on two segmentation tasks using three different medical image datasets, and the experimental results show that our BiX-NAS searched architecture achieves the state-of-the-art performance with significantly lower computational cost.
arXiv Detail & Related papers (2021-06-26T14:33:04Z) - A journey in ESN and LSTM visualisations on a language task [77.34726150561087]
We trained ESNs and LSTMs on a Cross-Situationnal Learning (CSL) task.
The results are of three kinds: performance comparison, internal dynamics analyses and visualization of latent space.
arXiv Detail & Related papers (2020-12-03T08:32:01Z) - Neural Architecture Search of SPD Manifold Networks [79.45110063435617]
We propose a new neural architecture search (NAS) problem of Symmetric Positive Definite (SPD) manifold networks.
We first introduce a geometrically rich and diverse SPD neural architecture search space for an efficient SPD cell design.
We exploit a differentiable NAS algorithm on our relaxed continuous search space for SPD neural architecture search.
arXiv Detail & Related papers (2020-10-27T18:08:57Z) - NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language
Processing [12.02718579660613]
We step outside the computer vision domain by leveraging the language modeling task, which is the core of natural language processing (NLP)
We have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it.
We have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation.
arXiv Detail & Related papers (2020-06-12T12:19:06Z) - MTL-NAS: Task-Agnostic Neural Architecture Search towards
General-Purpose Multi-Task Learning [71.90902837008278]
We propose to incorporate neural architecture search (NAS) into general-purpose multi-task learning (GP-MTL)
In order to adapt to different task combinations, we disentangle the GP-MTL networks into single-task backbones.
We also propose a novel single-shot gradient-based search algorithm that closes the performance gap between the searched architectures.
arXiv Detail & Related papers (2020-03-31T09:49:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.