FENAS: Flexible and Expressive Neural Architecture Search

Ramakanth Pasunuru, Mohit Bansal


Abstract
Architecture search is the automatic process of designing the model or cell structure that is optimal for the given dataset or task. Recently, this approach has shown good improvements in terms of performance (tested on language modeling and image classification) with reasonable training speed using a weight sharing-based approach called Efficient Neural Architecture Search (ENAS). In this work, we propose a novel architecture search algorithm called Flexible and Expressible Neural Architecture Search (FENAS), with more flexible and expressible search space than ENAS, in terms of more activation functions, input edges, and atomic operations. Also, our FENAS approach is able to reproduce the well-known LSTM and GRU architectures (unlike ENAS), and is also able to initialize with them for finding architectures more efficiently. We explore this extended search space via evolutionary search and show that FENAS performs significantly better on several popular text classification tasks and performs similar to ENAS on standard language model benchmark. Further, we present ablations and analyses on our FENAS approach.
Anthology ID:
2020.findings-emnlp.258
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2020
Month:
November
Year:
2020
Address:
Online
Editors:
Trevor Cohn, Yulan He, Yang Liu
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2869–2876
Language:
URL:
https://aclanthology.org/2020.findings-emnlp.258
DOI:
10.18653/v1/2020.findings-emnlp.258
Bibkey:
Cite (ACL):
Ramakanth Pasunuru and Mohit Bansal. 2020. FENAS: Flexible and Expressive Neural Architecture Search. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 2869–2876, Online. Association for Computational Linguistics.
Cite (Informal):
FENAS: Flexible and Expressive Neural Architecture Search (Pasunuru & Bansal, Findings 2020)
Copy Citation:
PDF:
https://aclanthology.org/2020.findings-emnlp.258.pdf
Data
CoLAGLUEMRPCMultiNLIQNLISSTSST-2