2021 29th European Signal Processing Conference (EUSIPCO) 2021
DOI: 10.23919/eusipco54536.2021.9616344
|View full text |Cite
|
Sign up to set email alerts
|

PyAWNeS-Codec: Speech and audio codec for ad-hoc acoustic wireless sensor networks

Abstract: Existing hardware with microphones can potentially be used as sensor networks to capture speech and audio signals for the benefit of better signal quality than possible with a single microphone. A central pre-requisite for such ad-hoc acoustic wireless sensor networks (ASWNs) is an efficient communication protocol with which to transmit audio data between nodes. For that purpose, we present the world's-first speech and audio codec especially designed for ASWNs, which has competitive quality also in single-chan… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 31 publications
0
5
0
Order By: Relevance
“…Particularly, for a constant sound pressure level, low frequencies are perceived more quietly in comparison to frequencies in the most sensitive area around 3000 Hz. One pre-processing technique that is commonly used for speech processing is pre-emphasis filtering [102]. This was initially applied to neural network modelling of guitar amplifiers by Dämskägg et al [75], where a first-order high-pass filter was applied to both model output and target data before computing the loss.…”
Section: Pre-emphasis Filteringmentioning
confidence: 99%
“…Particularly, for a constant sound pressure level, low frequencies are perceived more quietly in comparison to frequencies in the most sensitive area around 3000 Hz. One pre-processing technique that is commonly used for speech processing is pre-emphasis filtering [102]. This was initially applied to neural network modelling of guitar amplifiers by Dämskägg et al [75], where a first-order high-pass filter was applied to both model output and target data before computing the loss.…”
Section: Pre-emphasis Filteringmentioning
confidence: 99%
“…Since we are interested in speech data, our investigations are focused on selected speech audio codecs. Namely, we use the 3GPP Enhanced Voice Services (EVS) codec (Bruhn et al, 2012), Opus (Valin et al, 2012), our own PyAWNeS-codec (Bäckström et al, 2021), and the neural codec Lyra (Kleijn et al, 2021). In particular,…”
Section: Codecsmentioning
confidence: 99%
“…A delay compensation mode is integrated into the EVS codec, allowing the compensation of the integrated delay of about 32 ms within the encoded output signal. •The Python acoustic wireless network of sensors (PyAWNeS) codec is a speech and audio codec especially designed for distributed scenarios, where multiple independent devices sense and transmit the signal simultaneously (Bäckström et al, 2021). In contrast to prior codecs, it is designed to provide competitive quality in a single channel mode, but such that quality is improved with every added sensor.…”
Section: Codecsmentioning
confidence: 99%
See 2 more Smart Citations