Objective. Auditory Attention Decoding (AAD) determines which speaker the listener is focusing on by analyzing his/her EEG. CNN was adopted to extract Spectro-Spatial-Feature (SSF) from short-time-interval of EEG to detect auditory spatial attention without stimuli. However, the following factors are not considered in SSF-CNN scheme. i) Single-band frequency analysis cannot represent the EEG pattern precisely. ii) The power cannot represent the EEG feature related to the dynamic patterns of the attended auditory stimulus. iii) The temporal feature of EEG representing the relationship between EEG and attended stimulus is not extracted. To solve these problems, SSF-CNN scheme was modified. Approach. i) Multiple-frequency bands, but not a single alpha frequency band, of EEG, were analyzed to represent the EEG pattern more precisely. ii) Differential Entropy (DE), but not power, was extracted from each frequency band to represent the disorder degree of EEG, which was related to the dynamic patterns of the attended auditory stimulus. iii) CNN and Convolutional-Long-Short-Term-Memory (ConvLSTM) were combined to extract spectro-spatial-temporal features from the 3-D descriptor sequence constructed based on the topographical activity maps of multiple-frequency bands. Main results. Experimental results on KUL, DTU, and PKU with 0.1s, 1s, 2s, and 5s decision windows demonstrated that: i) The proposed model outperformed SSF-CNN and state-of-the-art AAD models. Specifically, when the auditory stimulus was unavailable, AAD accuracy could be enhanced by at least 3:25%, 3:96%, and 5:08% on KUL, DTU, and PKU, respectively, compared with the baselines. And, on KUL, the longer decision window corresponded to lower enhancement, while on both DTU and PKU, the longer decision window corresponded to higher enhancement, except for two cases when decision window length was 2s on PKU or 5s on DTU. ii) Each modification contributed to the performance enhancement. Significance. DE feature, multi-band frequency analysis, and ConvLSTM-based temporal analysis help to enhance AAD accuracy.
Objective. To take full advantage of both labeled data and unlabeled ones, the Graph Convolutional Network (GCN) was introduced in electroencephalography (EEG) based emotion recognition to achieve feature propagation. However, a single feature cannot represent the emotional state entirely and precisely due to the instability of the EEG signal and the complexity of the emotional state. In addition, the noise existing in the graph may affect the performance greatly. To solve these problems, it was necessary to introduce feature/similarity fusion and noise reduction strategies. Approach. A semi-supervised EEG emotion recognition model combining graph fusion, network enhancement, and feature fusion was proposed. Firstly, different features were extracted from EEG and then compacted by Principal Component Analysis (PCA), respectively. Secondly, a Sample-by-sample Similarity Matrix (SSM) was constructed based on each feature, and Similarity Network Fusion (SNF) was adopted to fuse the graphs corresponding to different SSMs to take advantage of their complementarity. Then, Network Enhancement (NE) was performed on the fused graph to reduce the noise in it. Finally, GCN was performed on the concatenated features and the enhanced fused graph to achieve feature propagation. Main results. Experimental results demonstrated that: i) When 5.30% of SEED and 7.20% of SEED-IV samples were chosen as the labeled samples, respectively, the minimum classification accuracy improvement achieved by the proposed scheme over state-of-the-art schemes were 1.52% on SEED and 13.14% on SEED-IV, respectively. ii) When 8.00% of SEED and 9.60% of SEED-IV samples were chosen as the labeled samples, respectively, the minimum training time reduction achieved by the proposed scheme over state-of-the-art schemes were 46.75s and 22.55s, respectively. iii) Graph fusion, network enhancement, and feature fusion all contributed to the performance enhancement. iv) The key hyperparameters that affect the performance were relatively few and easy to set to obtain outstanding performance. Significance. This paper demonstrated that the combination of graph fusion, network enhancement, and feature fusion help to enhance GCN-based EEG emotion recognition.
Similarity measurement plays an important role in various information retrieval tasks. In this paper, a music information retrieval scheme based on two-level similarity fusion and post-processing is proposed. At the similarity fusion level, to take full advantage of the common and complementary properties among different descriptors and different similarity functions, first, the track-by-track similarity graphs generated from the same descriptor but different similarity functions are fused with the similarity network fusion (SNF) technique. Then, the obtained first-level fused similarities based on different descriptors are further fused with the mixture Markov model (MMM) technique. At the post-processing level, diffusion is first performed on the two-level fused similarity graph to utilize the underlying track manifold contained within it. Then, a mutual proximity (MP) algorithm is adopted to refine the diffused similarity scores, which helps to reduce the bad influence caused by the “hubness” phenomenon contained in the scores. The performance of the proposed scheme is tested in the cover song identification (CSI) task on three cover song datasets (Covers80, Covers40, and Second Hand Songs (SHS)). The experimental results demonstrate that the proposed scheme outperforms state-of-the-art CSI schemes based on single similarity or similarity fusion.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.