Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1884
|View full text |Cite
|
Sign up to set email alerts
|

Global SNR Estimation of Speech Signals Using Entropy and Uncertainty Estimates from Dropout Networks

Abstract: This paper demonstrates two novel methods to estimate the global SNR of speech signals. In both methods, Deep Neural Network-Hidden Markov Model (DNN-HMM) acoustic model used in speech recognition systems is leveraged for the additional task of SNR estimation. In the first method, the entropy of the DNN-HMM output is computed. Recent work on bayesian deep learning has shown that a DNN-HMM trained with dropout can be used to estimate model uncertainty by approximating it as a deep Gaussian process. In the secon… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2019
2019
2025
2025

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 14 publications
0
4
0
Order By: Relevance
“…More recently, efficient strategies have been proposed to incorporate uncertainty estimation into deep neural networks [52,6]. Among them, MC-Dropout [12] and Deep Ensembles [21] are two of the most popular approaches given that they are agnostic to the specific network architecture [1,15,2,26]. More concretely, MC-Dropout adds stochastic dropout during inference into the intermediate network layers.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…More recently, efficient strategies have been proposed to incorporate uncertainty estimation into deep neural networks [52,6]. Among them, MC-Dropout [12] and Deep Ensembles [21] are two of the most popular approaches given that they are agnostic to the specific network architecture [1,15,2,26]. More concretely, MC-Dropout adds stochastic dropout during inference into the intermediate network layers.…”
Section: Related Workmentioning
confidence: 99%
“…In order to implicitly model the infinite set F, NeRF employs a neural network f θ (x, d) with parameters θ which outputs the density α and radiance r for any given input location-view pair {x, d}. Using this network, NeRF is able to estimate the color c(x o , d) for any given pixel defined by a 3D camera position x o and view direction d using the volumetric rendering function: (1) where x t = x o + td corresponds to 3D locations along a ray with direction d originated at the camera origin and intersecting with the pixel at x o .…”
Section: Deterministic and Stochastic Neural Radiance Fieldsmentioning
confidence: 99%
“…To address this limitation, other approaches have explored other strategies to implicitly learn the parameter distribution. For instance, dropout-based methods [9,2,12,6] introduce stochasticity over the intermediate neurons of the network in order to efficiently encode different possible solutions in the parameter space. By evaluating the model with different dropout configurations over the same input, the uncertainty can be quantified by computing the variance over the set of obtained outputs.…”
Section: Again)mentioning
confidence: 99%
“…The reason being that in a noisy environment audio loses its information and thus visual modality can be used to increase the overall performance of voice activity detection as visual modality is independent of noise. To detect that ambient surrounding is noisy, people have used SNR estimation approaches such as [17].…”
Section: Introductionmentioning
confidence: 99%