2022
DOI: 10.1007/s40593-022-00291-5
|View full text |Cite
|
Sign up to set email alerts
|

Automated Speech Scoring System Under The Lens

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 39 publications
0
3
0
Order By: Relevance
“…To compare these results with recent works, (Singla et al, 2021) reports that their hierarchical model achieves an average QWK of 0.82 across four datasets, which is slightly lower our framework on EFSET. Another features-based approach provided by (Bamdev et al, 2023) reports that the system achieves a QWK of 0.81 on SLTI SOPI dataset, which is also lower than our model on EF-SET. These papers suggest that the multimodal multitask framework has a competitive performance in automated speech scoring compared to other recent works.…”
Section: Discussionmentioning
confidence: 63%
See 1 more Smart Citation
“…To compare these results with recent works, (Singla et al, 2021) reports that their hierarchical model achieves an average QWK of 0.82 across four datasets, which is slightly lower our framework on EFSET. Another features-based approach provided by (Bamdev et al, 2023) reports that the system achieves a QWK of 0.81 on SLTI SOPI dataset, which is also lower than our model on EF-SET. These papers suggest that the multimodal multitask framework has a competitive performance in automated speech scoring compared to other recent works.…”
Section: Discussionmentioning
confidence: 63%
“…Recently, Bamdev et al (2023) presents a machine learning-based approach to assess the English proficiency of non-native speakers from their speech samples. The paper uses the SLTI SOPI dataset, which contains 1200 speech samples with different proficiency levels, rated by human experts on a scale from 1 to 5.…”
Section: Features-based Approachmentioning
confidence: 99%
“…After encoding the tokens in a sentence, we enumerate through all the possible m spans J = {j1, • • • , ji, • • • , jm} upto a maximum specified length (in terms of number of tokens) for sentence s = {w1, • • • , wT } and then re-assign a label yi ∈ {I, O} for each span ji. For example, for the sentence "NLP is um important", all possible spans (or pairs of start and end indices) are {(1, 1), (2, 2), (3,3), (4,4), (1,2), (2,3), (2,4), (1,3), (1,4)}, and all these spans are labelled O except (3,3) which is labelled I. We denote bi and si as the start and end indices of span ji respectively.…”
Section: Span Representation Layermentioning
confidence: 99%
“…Thus disfluency detection and removal can output clean inputs for downstream NLP tasks, like dialogue systems, question answering, and machine translation. Moreover, disfluency detection also finds applications in automatic speech scoring [1,2].…”
Section: Introductionmentioning
confidence: 99%
“…Traditionally, autograding systems are built using manually crafted features used with machine learning based models (Kumar et al, 2019;Bamdev et al, 2022). Lately, these systems have been shifting to deep learning based models (Ke and Ng, 2019).…”
Section: Introductionmentioning
confidence: 99%