A new method of extracting acoustic features based on auditory spike code is proposed. An auditory spike code represents the acoustic activities created by the signal, similar to sound encoding of the human auditory system. In the proposed method, an auditory spike code of the signal is computed using a 64-band Gammatone filterbank as the kernel functions. Then, for each spectral band, the sum and non-zero counts of the auditory spike code are determined, and the features corresponding to the population and occurrence rate of the acoustic activities for each band are computed. In addition, the distribution of the acoustic activities on a time axis is analysed based on the histogram of time intervals between the adjacent acoustic activities, and the features for expressing temporal properties of the signal are extracted. The reconstruction accuracy of the auditory spike code is also measured as the features. Different from most conventional features obtained by complex statistical modelling or learning, the features by the proposed method can directly show specific acoustic characteristics contained in the signal. These features are applied to a music genre classification, and it is confirmed that they provide a performance comparable to state-of-the-art features.
Speech captured by an in-ear microphone placed inside an occluded ear has a high signal-to-noise ratio; however, it has different sound characteristics compared to normal speech captured through air conduction. In this study, a method for blind speech quality enhancement is proposed that can convert speech captured by an in-ear microphone to one that resembles normal speech. The proposed method estimates an inputdependent enhancement function by using a neural network in the feature domain and enhances the captured speech via time-domain filtering. Subjective and objective evaluations confirm that the speech enhanced using our proposed method sounds more similar to normal speech than that enhanced using conventional equalizer-based methods.
SUMMARYA method for encoding detection and bit rate classification of AMR-coded speech is proposed. For each texture frame, 184 features consisting of the short-term and long-term temporal statistics of speech parameters are extracted, which can effectively measure the amount of distortion due to AMR. The deep neural network then classifies the bit rate of speech after analyzing the extracted features. It is confirmed that the proposed features provide better performance than the conventional spectral features designed for bit rate classification of coded audio. key words: bit rate, speech codec, AMR, deep neural network, feature vector
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.