2022
DOI: 10.48550/arxiv.2203.16032
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ConferencingSpeech 2022 Challenge: Non-intrusive Objective Speech Quality Assessment (NISQA) Challenge for Online Conferencing Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…Non-intrusive objective speech quality assessment tools like ITU-T P.563 [14] do not require a reference, though it has low correlation to subjective quality [15]. Newer neural net-based methods such as [15,16,17,18] provide better correlations to subjective quality. NISQA [19] is an objective metric for P.804, though the correlation to subjective quality is not sufficient to use as a challenge metric.…”
Section: Related Workmentioning
confidence: 99%
“…Non-intrusive objective speech quality assessment tools like ITU-T P.563 [14] do not require a reference, though it has low correlation to subjective quality [15]. Newer neural net-based methods such as [15,16,17,18] provide better correlations to subjective quality. NISQA [19] is an objective metric for P.804, though the correlation to subjective quality is not sufficient to use as a challenge metric.…”
Section: Related Workmentioning
confidence: 99%
“…For the PSTN and Tencent corpora, our model performs comparably to the next best model (1.7% better and 5.6% worse). But for the TUB corpus, our model performed 27.4% better than the next-closest competitor for that corpus and 42.9% better than the overall secondplace model [10]. This suggests that the proposed model performs exceptionally well on unseen data.…”
Section: Test Resultsmentioning
confidence: 83%
“…Additionally, we test various combinations of the BiLSTM and AttPoolFF components placed on top of both the XLS-R and MFCC feature extractors. We show that the XLS-R feature extractor does not only outperform the classic MFCC feature extractor, but it also results in competitive model performances with regards to the root mean squared error (RMSE) on data from the ConferencingSpeech 2022 challenge [10].…”
Section: Introductionmentioning
confidence: 99%
“…Automatic audio quality measurement can be applied in many different audio related tasks, including detecting fake-quality audio [1], monitoring speech quality for online conference [2], or assessing the naturalness of synthesized speech [3]. Besides, since many people use mobile phones in their daily life, and audio quality is one of the considerations of choosing mobile phones [4], measuring the quality of audio playback for mobile phones would also be one of the important tasks.…”
Section: Introductionmentioning
confidence: 99%