Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-273
|View full text |Cite
|
Sign up to set email alerts
|

Visual Transformers for Primates Classification and Covid Detection

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
7
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 0 publications
0
7
0
Order By: Relevance
“…ANIMAL-SPOT demonstrated also great results in terms of single-stage multi-species primate classification, by outperforming the ComParE baseline system. The final result of 89.3% also exceeds the UAR of 88.3% presented by challenge competitor Illium et al 50 , who applied a vision transformer to the classification problem. Müller et al 51 report the same UAR of 89.3% while applying a Deep Recurrent Neural Network.…”
Section: Discussion and Future Outlookmentioning
confidence: 67%
“…ANIMAL-SPOT demonstrated also great results in terms of single-stage multi-species primate classification, by outperforming the ComParE baseline system. The final result of 89.3% also exceeds the UAR of 88.3% presented by challenge competitor Illium et al 50 , who applied a vision transformer to the classification problem. Müller et al 51 report the same UAR of 89.3% while applying a Deep Recurrent Neural Network.…”
Section: Discussion and Future Outlookmentioning
confidence: 67%
“…UAR is equivalent to balanced accuracy. We note that if F1 had been the evaluation metric, Illium, et al ( 25 ) would have infact won the cough sub challenge. This is thanks to their model’s superior precision performance, i.e., what proportion of the model’s positive predictions are correct.…”
mentioning
confidence: 78%
“…Team Casanova et al exploited a noise addition method and SpecAugment to augment the challenge dataset ( 23 ). Team Illium et al targeted spectrogram-level augmentations with temporal shifting, noise addition, SpecAugment and loudness adjustment ( 25 ). Instead of using a data augmentation method to manipulate the challenge dataset, team Klumpp et al used three auxiliary datasets in different languages aiming their deep acoustic model to better learn the properties of healthy speech ( 24 ).…”
Section: Overview Of Methodologies Used In Accepted Papers At Intersp...mentioning
confidence: 99%
See 2 more Smart Citations