2020
DOI: 10.1109/tvcg.2020.2973058
|View full text |Cite
|
Sign up to set email alerts
|

Scene-Aware Audio Rendering via Deep Acoustic Analysis

Abstract: Fig. 1: Given a natural sound in a real-world room that is recorded using a cellphone microphone (left), we estimate the acoustic material properties and the frequency equalization of the room using a novel deep learning approach (middle). We use the estimated acoustic material properties for generating plausible sound effects in the virtual model of the room (right). Our approach is general and robust, and works well with commodity devices.Abstract-We present a new method to capture the acoustic characteristi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
22
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 32 publications
(22 citation statements)
references
References 58 publications
0
22
0
Order By: Relevance
“…For current acoustic-based material identification method, Tang et al [20] presented a new method to capture the acoustic characteristics of real-world rooms using commodity devices, and use the captured characteristics to generate similar sounding sources with virtual models. However, they required pre-constructed virtual room geometry as input for acoustic analysis and material optimization, which does not fit our VR/AR application and capability.…”
Section: Related Workmentioning
confidence: 99%
“…For current acoustic-based material identification method, Tang et al [20] presented a new method to capture the acoustic characteristics of real-world rooms using commodity devices, and use the captured characteristics to generate similar sounding sources with virtual models. However, they required pre-constructed virtual room geometry as input for acoustic analysis and material optimization, which does not fit our VR/AR application and capability.…”
Section: Related Workmentioning
confidence: 99%
“…Li et al [23] also generate spatial audio using a recorded acoustic impulse response. Tang et al [41] present a deep learning method to estimate the acoustic material characteristics of rooms and use them for real-world audio rendering. All these methods have been used for real-world rooms and are not directly applicable to arbitrary video recordings.…”
Section: Reconstructing Spatial Information From Audiomentioning
confidence: 99%
“…The conventional estimation methods use the room impulse response (RIR) [7][8][9] or room transfer function (RTF) [10,11], and convolve RIR and clean sound to create a virtual reverberant sound in a specific space. Among recent studies, there is a study of learning an artificial neural network using the structure of a room and acoustic signals acquired from the room, where RT60, which is an attenuation parameter of the sound pressure level, was estimated and used to construct a reverberant signal [12]. In a similar way, the methods for generating acoustic parameters using deep neural networks (DNNs) have been studied.…”
Section: Introductionmentioning
confidence: 99%