Proceedings of ACL 2018, System Demonstrations 2018
DOI: 10.18653/v1/p18-4018
|View full text |Cite
|
Sign up to set email alerts
|

CRUISE: Cold-Start New Skill Development via Iterative Utterance Generation

Abstract: We present a system, CRUISE, that guides ordinary software developers to build a high quality natural language understanding (NLU) engine from scratch. This is the fundamental step of building a new skill for personal assistants. Unlike existing solutions that require either developers or crowdsourcing to manually generate and annotate a large number of utterances, we design a hybrid rulebased and data-driven approach with the capability to iteratively generate more and more utterances. Our system only require… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
8

Relationship

4
4

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 11 publications
0
5
0
Order By: Relevance
“…Unfortunately, most of these toolkits require both linguistic expertise and a large amount of annotated data. CRUISE (Shen et al, 2018) provide an utterance generation system to reduce the human workload of data annotation. However, CRUISE focuses on spoken language understanding.…”
Section: Related Workmentioning
confidence: 99%
“…Unfortunately, most of these toolkits require both linguistic expertise and a large amount of annotated data. CRUISE (Shen et al, 2018) provide an utterance generation system to reduce the human workload of data annotation. However, CRUISE focuses on spoken language understanding.…”
Section: Related Workmentioning
confidence: 99%
“…airline reservation), it can be both time-consuming and expensive to collect and annotate training utterances corresponding to each possible combination of slots. Secondly, for resource constrained cold-start skill developers [11], it is cheaper and easier to annotate a small number of short utterances (with just one or two slots) for training, than longer utterances with many slots which the SLU model may encounter after deployment. Building compositional SLU models which can generalize well under both these settings is vital for both scalable development, and reliability of future AI agents.…”
Section: Introductionmentioning
confidence: 99%
“…To address the challenges in open-world settings, previous works adopt varied strategies. Shen et al (2018aShen et al ( , 2019c) use a cold-start algorithm to generate additional training data to cover a larger variety of utterances. This strategy relies on the software developers to pre-build all possible skills.…”
Section: Introductionmentioning
confidence: 99%