2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016
DOI: 10.1109/icassp.2016.7472814
|View full text |Cite
|
Sign up to set email alerts
|

Joint dictionary training for bandwidth extension of speech signals

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…Source-filter models have been also employed to extend the bandwidth using line spectral frequencies in [15]. Systems based on dictionary learning to map low-frequency patterns to high-frequency components have been proposed in [16,17]. Classic machine learning methods have also been explored for BWE, such as Gaussian mixture models (GMMs) [18], hidden Markov models (HMM) [19,20], or non-negative matrix factorization (NMF) [21,22].…”
Section: Signal Processing Approachesmentioning
confidence: 99%
“…Source-filter models have been also employed to extend the bandwidth using line spectral frequencies in [15]. Systems based on dictionary learning to map low-frequency patterns to high-frequency components have been proposed in [16,17]. Classic machine learning methods have also been explored for BWE, such as Gaussian mixture models (GMMs) [18], hidden Markov models (HMM) [19,20], or non-negative matrix factorization (NMF) [21,22].…”
Section: Signal Processing Approachesmentioning
confidence: 99%
“…The bandwidth extension is then divided into two separate tasks: spectral envelope extension and excitation signal extension. Early studies employ codebook mapping techniques for spectral envelope extension, using two codebooks representing narrowband and wideband spectra H. Wang is with the Department of Computer Science and Engineering, The Ohio State University, OH 43210, USA (e-mail: wang.11401@osu.edu) E. W. Healy is with the Department of Speech and Hearing Science and the Center for Cognitive and Brain Sciences, The Ohio State University, Columbus, OH 43210, USA (e-mail: healy.66@osu.edu) D. L. Wang is with the Department of Computer Science and Engineering and the Center for Cognitive and Brain Sciences, The Ohio State University, Columbus, OH 43210, USA (e-mail: dwang@cse.ohio-state.edu) [52], [58]. Through training, the best matching entry in the wideband codebook to the narrowband codebook is used to generate the spectral envelope.…”
Section: Introductionmentioning
confidence: 99%
“…In Kim et al [2008], temporal envelope and fine structure of sub-bands are taken into account for ABE. In Li and Lee [2015], Sadasivan et al [2016], Bin et al [2015], Bachhav et al [2017], the WB magnitude spectrum is taken directly for extracting bene-ficial HB and NB information. In Sunil and Sinha [2014], sparse modeling is used for ABE, wherein it requires a separate dictionary for voiced speech and unvoiced speech.…”
Section: Introductionmentioning
confidence: 99%