2019
DOI: 10.48550/arxiv.1912.03884
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MITAS: A Compressed Time-Domain Audio Separation Network with Parameter Sharing

Abstract: Deep learning methods have brought substantial advancements in speech separation (SS). Nevertheless, it remains challenging to deploy deep-learning-based models on edge devices. Thus, identifying an effective way to compress these large models without hurting SS performance has become an important research topic. Recently, TasNet and Conv-TasNet have been proposed. They achieved state-ofthe-art results on several standardized SS tasks. Moreover, their low latency natures make them definitely suitable for real-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 22 publications
0
1
0
Order By: Relevance
“…We apply cross-layer parameter sharing to reduce parameters and make the model more compact. The cross-layer parameter sharing has shown to be helpful in NLP task pretraining [15] and source separation [16]. As shown in Figure 1(a), transformations in the compressed WaveGlow share the same affine coupling layer 1 .…”
Section: Highly Compressed Waveglowmentioning
confidence: 99%
“…We apply cross-layer parameter sharing to reduce parameters and make the model more compact. The cross-layer parameter sharing has shown to be helpful in NLP task pretraining [15] and source separation [16]. As shown in Figure 1(a), transformations in the compressed WaveGlow share the same affine coupling layer 1 .…”
Section: Highly Compressed Waveglowmentioning
confidence: 99%