2022
DOI: 10.48550/arxiv.2202.12603
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Language technology practitioners as language managers: arbitrating data bias and predictive bias in ASR

Abstract: Despite the fact that variation is a fundamental characteristic of natural language, automatic speech recognition systems perform systematically worse on non-standardised and marginalised language varieties. In this paper we use the lens of language policy to analyse how current practices in training and testing ASR systems in industry lead to the data bias giving rise to these systematic error differences. We believe that this is a useful perspective for speech and language technology practitioners to underst… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 37 publications
0
1
0
Order By: Relevance
“…In automatic speech recognition, biases are in the direction of better recognition of 'standard' accents Harrington, 2023), one or other of male or female voices depending on the system (Markl and McNulty, 2022) as well as non-pathological voices (Benzeghiba et al, 2007;Markl and McNulty, 2022). Additionally, as noted by Benzeghiba et al (2007), children's voices and elderly voices are also generally not modelled well and cause performance issues with ASR.…”
Section: Fundingmentioning
confidence: 99%
“…In automatic speech recognition, biases are in the direction of better recognition of 'standard' accents Harrington, 2023), one or other of male or female voices depending on the system (Markl and McNulty, 2022) as well as non-pathological voices (Benzeghiba et al, 2007;Markl and McNulty, 2022). Additionally, as noted by Benzeghiba et al (2007), children's voices and elderly voices are also generally not modelled well and cause performance issues with ASR.…”
Section: Fundingmentioning
confidence: 99%
“…In automatic speech recognition, biases are in the direction of better recognition of 'standard' accents (Markl, 2022;Wassink et al, 2022;Harrington, 2023), one or other of male or female voices depending on the system (Markl and McNulty, 2022) as well as non-pathological voices (Benzeghiba et al, 2007;Markl and McNulty, 2022). Additionally, as noted by Benzeghiba et al (2007), children's voices and elderly voices are also generally not modelled well and cause performance issues with ASR.…”
Section: Introductionmentioning
confidence: 99%