2020
DOI: 10.1007/s00521-020-05424-2
|View full text |Cite
|
Sign up to set email alerts
|

Latent Timbre Synthesis

Abstract: We present the Latent Timbre Synthesis (LTS), a new audio synthesis method using Deep Learning. The synthesis method allows composers and sound designers to interpolate and extrapolate between the timbre of multiple sounds using the latent space of audio frames. We provide the details of two Variational Autoencoder architectures for LTS, and compare their advantages and drawbacks. The implementation includes a fully working application with graphical user interface, called interpolate two, which enables practi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 25 publications
0
8
0
Order By: Relevance
“…is synthesis method allows composers and sound designers to interpolate and extrapolate between the timbres of multiple sounds in the latent space of audio frames [2]. Banerjee et al believe that the analysis of sound signals in the linear deterministic approach reached a new dimension and developed many well-equipped software to precisely measure and control the basic parameters of sound, such as pitch, intensity, and rhythm [3].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…is synthesis method allows composers and sound designers to interpolate and extrapolate between the timbres of multiple sounds in the latent space of audio frames [2]. Banerjee et al believe that the analysis of sound signals in the linear deterministic approach reached a new dimension and developed many well-equipped software to precisely measure and control the basic parameters of sound, such as pitch, intensity, and rhythm [3].…”
Section: Related Workmentioning
confidence: 99%
“…e main innovations of this paper are as follows: (1) feature extraction-this part mainly analyzes the characteristics of the audio signal on the basis of preprocessing and then extracts the characteristic curve of the audio signal to pave the way for the subsequent audio melody matching. (2) Audio feature library construction-this part mainly studies the music in MIDI le format and uses the improved contour algorithm to build the audio feature library, which is used as the data source for melody matching. is plays an important role in the timbre analysis of various musical instruments in the text, and he can distinguish the timbres of various musical instruments.…”
Section: Introductionmentioning
confidence: 99%
“…For the SVM, tested parameter values for C and gamma were respectively [0.01, 0.1, 1, 10, 100, 1000] and [1, 0.1, 0.01, 0.001, 0.00001, 0.000001, 10]. For KNN, the tested N values were [1,2,3,4,5,6,7,8,9,10] and for the Random forest [1, 10, 100, 500,1000] trees were considered. In the case of the FuzzyRuLer, d was explored from 1 to half the number of features in the dataset and ratio with [0.9, 0.8, 0.7, 0.6, 0.5] values.…”
Section: Cross-validationmentioning
confidence: 99%
“…A deep learning based system that allows for interpolation and extrapolation between the timbre of multiple sounds is presented in [10]. Deep-learning systems are a promising path for sound synthesis applications, although their training times still do not allow for real-time feedback.…”
Section: Introductionmentioning
confidence: 99%
“…3). Th is type of corpus in a musical agent is known as a hybrid corpus [15]. In line with Respire's aesthetic choice of ambiguity, we aimed for a curation of abstract and ambient sounds for the musical agent's memory.…”
Section: Sound Memorymentioning
confidence: 99%