Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-351
|View full text |Cite
|
Sign up to set email alerts
|

Disfluency Detection with Unlabeled Data and Small BERT Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 16 publications
(13 citation statements)
references
References 0 publications
0
13
0
Order By: Relevance
“…In addition to shorter word-level latency metrics presented in the results, the runtime latency of the BERTSV model is 80% lower than that of BERTBASE(Rocholl et al, 2021).…”
mentioning
confidence: 73%
See 2 more Smart Citations
“…In addition to shorter word-level latency metrics presented in the results, the runtime latency of the BERTSV model is 80% lower than that of BERTBASE(Rocholl et al, 2021).…”
mentioning
confidence: 73%
“…We specifically use the version from the Linguistic Data Corpus's Treebank-3 (Marcus et al, 1999) distribution, which additionally contains disfluency annotations and a standard train/dev/test split (Charniak and Johnson, 2001). We follow Rocholl et al (2021), training our models to classify both the reparanda and interregna as disfluent for future removal in a final post-processed transcript.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Disfluency detection is a widely studied area of research, with the most successful approaches leveraging BERT transformer models to achieve high accuracy (e.g. Bach and Huang, 2019;Jamshid Lou and Johnson, 2020;Rocholl et al, 2021). These models operate non-incrementally using whole sentences as inputs, often with a view to remove the disfluencies from transcripts all together.…”
Section: Related Workmentioning
confidence: 99%
“…For the filled pause, we simply relabel the silence following predefined filler words as <pause> . Finally, we exploit the disfluency detector [18] based on small vocabulary BERT to identify reparandum and interregnum (e.g. "you know", "well", "I mean") of disfluencies.…”
Section: Training Data Annotationmentioning
confidence: 99%