Interspeech 2016 2016
DOI: 10.21437/interspeech.2016-701
|View full text |Cite
|
Sign up to set email alerts
|

On Discriminative Framework for Single Channel Audio Source Separation

Abstract: Single channel source separation (SCSS) algorithms that utilise discriminative source models perform better in comparison to those that are trained independently. However, all the aspects of training discriminative models have not been addressed in the literature. For instance, the choice of dimensions of source models (number of columns of NMF, Dictionary etc) not only influences the fidelity of a given source but also impacts the interference introduced in it. Therefore choosing a right dimension parameter f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…Motivated by relevant applications and issues with joint separation paradigm, we introduced a 'one source at a time' separation framework in [6] for two sources when non-negative matrix factorization (NMF) based models for each source is used. The first contribution of this paper is the extension of this strategy to any number of sources when the sources are modeled using DNN.…”
Section: Contributionsmentioning
confidence: 99%
See 1 more Smart Citation
“…Motivated by relevant applications and issues with joint separation paradigm, we introduced a 'one source at a time' separation framework in [6] for two sources when non-negative matrix factorization (NMF) based models for each source is used. The first contribution of this paper is the extension of this strategy to any number of sources when the sources are modeled using DNN.…”
Section: Contributionsmentioning
confidence: 99%
“…Another modification that we propose is that the hyperparameters γ and µ, which are fixed in joint separation, are automatically searched in our formulation (10) using the ratios re, rs, rn, we first defined in [6]. The error ratio is defined as: where ỹns and ỹss are the estimates of ys when the input to the network is yn and ys respectively.…”
Section: Automatic Parameter Tuningmentioning
confidence: 99%
“…This problem has been a topic of study in recent years. Grais et al [18,19] achieved good performance by adding a penalty term to the objective function to minimize the cross-coherence between source-specific sub-dictionaries. Xu et al [20,21] proposed a discriminative dictionary learning (DDL) method to penalize the energy that contributes to a specific sub-dictionary but also originates from other sources.…”
Section: Introductionmentioning
confidence: 99%