2003
DOI: 10.1109/lsp.2003.813679
|View full text |Cite
|
Sign up to set email alerts
|

Speech probability distribution

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
143
0
1

Year Published

2006
2006
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 258 publications
(144 citation statements)
references
References 12 publications
0
143
0
1
Order By: Relevance
“…central limit theorem). Previous work in speech processing has demonstrated that Laplacian (LD) and Gamma (ΓD) distributions are more suitable than GD for approximating active voice segments for many frame sizes (Gazor and Zhang, 2003;Martin, 2005). More specifically, LD fits well the highly correlated univariate space of speech amplitudes as well as the uncorrelated multivariate space of feature values after a Karhunen-Loeve Transformation (KLT) or Discrete Cosine Transformation (DCT).…”
Section: Speech Modelling With the Generalised Gamma Distributionmentioning
confidence: 99%
See 2 more Smart Citations
“…central limit theorem). Previous work in speech processing has demonstrated that Laplacian (LD) and Gamma (ΓD) distributions are more suitable than GD for approximating active voice segments for many frame sizes (Gazor and Zhang, 2003;Martin, 2005). More specifically, LD fits well the highly correlated univariate space of speech amplitudes as well as the uncorrelated multivariate space of feature values after a Karhunen-Loeve Transformation (KLT) or Discrete Cosine Transformation (DCT).…”
Section: Speech Modelling With the Generalised Gamma Distributionmentioning
confidence: 99%
“…More specifically, LD fits well the highly correlated univariate space of speech amplitudes as well as the uncorrelated multivariate space of feature values after a Karhunen-Loeve Transformation (KLT) or Discrete Cosine Transformation (DCT). It is also asserted that DCT coefficients of clean speech samples are better modelled using the generalised Gaussian distribution (GGD) since they have a distribution which is more peaky than a LD (Gazor and Zhang, 2003). Nakamura (2000) proposes a generalised Laplace distribution (GLD) for speech recognition, demonstrating significant improvement in word accuracy when MFCC is used.…”
Section: Speech Modelling With the Generalised Gamma Distributionmentioning
confidence: 99%
See 1 more Smart Citation
“…In specific, LD fits well the highly correlated univariate space of the speech amplitudes as well as the uncorrelated multivariate space of the feature values after a Karhunen-Loeve Transformation (KLT) or Discrete Cosine Transformation (DCT) [9]. While some reports attest that LD offers only a marginally better fit than GD, this is not valid when silence segments are absent from the testing [8]. The reason is that while clean speech segments best exhibit LD or ΓD properties the silence segments are Gaussian random processes.…”
Section: Speech Distributionsmentioning
confidence: 92%
“…Using a transformed feature space, it is possible to assume that these two Gaussian random processes are independent of each other and maximum a posteriori estimators can be used to determine the signal parameters. Nevertheless, previous work in speech processing has demonstrated that Laplacian (LD) and Gamma (ΓD) distributions are more suitable than GD for approximating active voice segments for most frame sizes [8]. In specific, LD fits well the highly correlated univariate space of the speech amplitudes as well as the uncorrelated multivariate space of the feature values after a Karhunen-Loeve Transformation (KLT) or Discrete Cosine Transformation (DCT) [9].…”
Section: Speech Distributionsmentioning
confidence: 99%