2017 IEEE 27th International Workshop on Machine Learning for Signal Processing (MLSP) 2017
DOI: 10.1109/mlsp.2017.8168108
|View full text |Cite
|
Sign up to set email alerts
|

Neural network alternatives toconvolutive audio models for source separation

Abstract: Convolutive Non-Negative Matrix Factorization model factorizes a given audio spectrogram using frequency templates with a temporal dimension. In this paper, we present a convolutional auto-encoder model that acts as a neural network alternative to convolutive NMF. Using the modeling flexibility granted by neural networks, we also explore the idea of using a Recurrent Neural Network in the encoder. Experimental results on speech mixtures from TIMIT dataset indicate that the convolutive architecture provides a s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
9
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
2
2

Relationship

3
5

Authors

Journals

citations
Cited by 12 publications
(9 citation statements)
references
References 11 publications
(16 reference statements)
0
9
0
Order By: Relevance
“…Recent work has explored the connection between CDL and auto-encoders in deep learning [5,6]. Framing dictionary learning as training of a neural network opens the possibility of exploiting the GPU-based infrastructure that has been developed for training neural networks, to reduce computational requirements and decrease runtime.…”
Section: Introductionmentioning
confidence: 99%
“…Recent work has explored the connection between CDL and auto-encoders in deep learning [5,6]. Framing dictionary learning as training of a neural network opens the possibility of exploiting the GPU-based infrastructure that has been developed for training neural networks, to reduce computational requirements and decrease runtime.…”
Section: Introductionmentioning
confidence: 99%
“…This allows the network to learn a non-negative H and a non-negative reconstruction S of the input S. Unlike NMF, the decoder weights need not be strictly non-negative. But, under suitable sparsity constraints on the activations H, they can be shown to be non-negative like NMF bases [19,20].…”
Section: Non-negative Autoencodersmentioning
confidence: 99%
“…We can now take advantage of the modeling flexibility of neural networks and develop complex encoder and decoder architectures that adhere to the above format. In particular, multilayer [19] and convolutional extensions [20] have shown significant performance improvement compared to a single dense-layer encoder and decoder given in Eq. 1.…”
Section: Non-negative Autoencodersmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, this class of factorization techniques has started to fall under the deep learning umbrella. These methods interpret non-negative autoencoders as an extension of non-negative factorization methods [5,6]. These hybrid models benefit from the advances in deep learning while maintaining the interpretability that many deep models forgo.…”
Section: Introductionmentioning
confidence: 99%