2021
DOI: 10.48550/arxiv.2105.11681
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Neural Networks and End-to-End Learning for Audio Compression

Abstract: Recent achievements in end-to-end deep learning have encouraged the exploration of tasks dealing with highly structured data with unified deep network models. Having such models for compressing audio signals has been challenging since it requires discrete representations that are not easy to train with end-to-end backpropagation. In this paper, we present an end-to-end deep learning approach that combines recurrent neural networks (RNNs) within the training strategy of variational autoencoders (VAEs) with a bi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 15 publications
0
1
0
Order By: Relevance
“…sg(c) is the operation that detaches its input, c, from the gradient computation in backpropagation. Therefore, the final output, z, is a binary variable, z ∈ {0, 1}, but it can flow the gradient through the first term, p. As a variant of STE, this binary reparameterization trick is easy to implement and work with other network architectures (Rim et al, 2021). We propose a reparameterization trick for the categorical distribution inspired by this binary reparameterization trick.…”
Section: Binary Reparameterization Trickmentioning
confidence: 99%
“…sg(c) is the operation that detaches its input, c, from the gradient computation in backpropagation. Therefore, the final output, z, is a binary variable, z ∈ {0, 1}, but it can flow the gradient through the first term, p. As a variant of STE, this binary reparameterization trick is easy to implement and work with other network architectures (Rim et al, 2021). We propose a reparameterization trick for the categorical distribution inspired by this binary reparameterization trick.…”
Section: Binary Reparameterization Trickmentioning
confidence: 99%