Proceedings of the 17th SIGMORPHON Workshop on Computational Research in Phonetics, Phonology, and Morphology 2020
DOI: 10.18653/v1/2020.sigmorphon-1.2
|View full text |Cite
|
Sign up to set email alerts
|

The SIGMORPHON 2020 Shared Task on Multilingual Grapheme-to-Phoneme Conversion

Abstract: We describe the design and findings of the SIGMORPHON 2020 shared task on multilingual grapheme-to-phoneme conversion. Participants were asked to submit systems which consume a sequence of graphemes then emit output a sequence of phonemes representing the pronunciation of that grapheme sequence in one of fifteen languages. Nine teams submitted a total of 23 systems, at best achieving an 18% relative reduction in word error rate (macro-averaged over languages), versus strong neural sequence-to-sequence baseline… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
17
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 31 publications
(19 citation statements)
references
References 35 publications
0
17
0
Order By: Relevance
“…The SIGMORPHON 2020 Shared Task is composed of three tasks: Task 0 on typologically diverse morphological inflection (Vylomova et al, 2020), Task 1 on multilingual grapheme-tophoneme conversion (Gorman et al, 2020), and Task 2 on unsupervised morphological paradigm completion . We submit systems to Tasks 0 and 2.…”
Section: Sigmorphon 2020 Shared Taskmentioning
confidence: 99%
“…The SIGMORPHON 2020 Shared Task is composed of three tasks: Task 0 on typologically diverse morphological inflection (Vylomova et al, 2020), Task 1 on multilingual grapheme-tophoneme conversion (Gorman et al, 2020), and Task 2 on unsupervised morphological paradigm completion . We submit systems to Tasks 0 and 2.…”
Section: Sigmorphon 2020 Shared Taskmentioning
confidence: 99%
“…It features fifteen languages from various phylogenetic families and written in different scripts. We refer the reader to Gorman et al (2020) for an overview of the language data. Each language comes with 3,600 training and 450 development set examples.…”
Section: Introductionmentioning
confidence: 99%
“…• We show that sparse seq2seq techniques, previously used for morphological inflection and machine translation , are also effective for multilingual g2p. We make four submissions to Task 1 (Gorman et al, 2020), which differ based on their choice of softmax replacement (1.5-entmax or sparsemax) and their architecture (RNN or transformer). Our strongest models finish third in word error rate (WER) and second in phoneme error rate (PER).…”
Section: Introductionmentioning
confidence: 99%