2019
DOI: 10.48550/arxiv.1912.10729
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

TextNAS: A Neural Architecture Search Space tailored for Text Representation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(7 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…Pham et al [19] propose an extension of NAS, namely ENAS, to speed up training processing by forcing all child networks to share weights. Apart from algorithms, NAS also has many valuable applications such as image classification [20,21], video segmentation [22], text representation [23] and etc. Hence, NAS is a demonstrated powerful tool and it is especially useful in continual learning scenarios when one needs to determine what is a good architecture for the new task.…”
Section: Related Workmentioning
confidence: 99%
“…Pham et al [19] propose an extension of NAS, namely ENAS, to speed up training processing by forcing all child networks to share weights. Apart from algorithms, NAS also has many valuable applications such as image classification [20,21], video segmentation [22], text representation [23] and etc. Hence, NAS is a demonstrated powerful tool and it is especially useful in continual learning scenarios when one needs to determine what is a good architecture for the new task.…”
Section: Related Workmentioning
confidence: 99%
“…Once the supernet is trained, each sampled structure can be directly and quickly evaluated without training from scratch. In our work, we also leverage the weight sharing strategy and build a one-shot model [8] with TextNAS [22] search space, while random search is adopted to sample architectures. Besides, we incorporate knowledge distillation and efficiency constraints as search hints to obtain effective and light student models.…”
Section: Related Workmentioning
confidence: 99%
“…Neural Architecture Search: The task is to find an effective and efficient architecture for AutoADR sub-model through teacherstudent framework. We adopt the search space in TextNAS [22] and build a corresponding network subsuming all possible architectures, which is called supernet [8] for the one-shot search algorithm. To apply knowledge distillation, we train the supernet with uniform sampling under the guidance of the soft predictions from teacher model.…”
Section: Autoadr 41 Overviewmentioning
confidence: 99%
See 2 more Smart Citations