Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/179031
Title: | Cross-modal credibility modelling for EEG-based multimodal emotion recognition | Authors: | Zhang, Yuzhe Liu, Huan Wang, Di Zhang, Dalin Lou, Tianyu Zheng, Qinghua Quek, Chai |
Keywords: | Computer and Information Science | Issue Date: | 2024 | Source: | Zhang, Y., Liu, H., Wang, D., Zhang, D., Lou, T., Zheng, Q. & Quek, C. (2024). Cross-modal credibility modelling for EEG-based multimodal emotion recognition. Journal of Neural Engineering, 21(2), 026040-. https://dx.doi.org/10.1088/1741-2552/ad3987 | Journal: | Journal of Neural Engineering | Abstract: | Objective. The study of emotion recognition through electroencephalography (EEG) has garnered significant attention recently. Integrating EEG with other peripheral physiological signals may greatly enhance performance in emotion recognition. Nonetheless, existing approaches still suffer from two predominant challenges: modality heterogeneity, stemming from the diverse mechanisms across modalities, and fusion credibility, which arises when one or multiple modalities fail to provide highly credible signals. Approach. In this paper, we introduce a novel multimodal physiological signal fusion model that incorporates both intra-inter modality reconstruction and sequential pattern consistency, thereby ensuring a computable and credible EEG-based multimodal emotion recognition. For the modality heterogeneity issue, we first implement a local self-attention transformer to obtain intra-modal features for each respective modality. Subsequently, we devise a pairwise cross-attention transformer to reveal the inter-modal correlations among different modalities, thereby rendering different modalities compatible and diminishing the heterogeneity concern. For the fusion credibility issue, we introduce the concept of sequential pattern consistency to measure whether different modalities evolve in a consistent way. Specifically, we propose to measure the varying trends of different modalities, and compute the inter-modality consistency scores to ascertain fusion credibility. Main results. We conduct extensive experiments on two benchmarked datasets (DEAP and MAHNOB-HCI) with the subject-dependent paradigm. For the DEAP dataset, our method improves the accuracy by 4.58%, and the F1 score by 0.63%, compared to the state-of-the-art baseline. Similarly, for the MAHNOB-HCI dataset, our method improves the accuracy by 3.97%, and the F1 score by 4.21%. In addition, we gain much insight into the proposed framework through significance test, ablation experiments, confusion matrices and hyperparameter analysis. Consequently, we demonstrate the effectiveness of the proposed credibility modelling through statistical analysis and carefully designed experiments. Significance. All experimental results demonstrate the effectiveness of our proposed architecture and indicate that credibility modelling is essential for multimodal emotion recognition. | URI: | https://hdl.handle.net/10356/179031 | ISSN: | 1741-2560 | DOI: | 10.1088/1741-2552/ad3987 | Schools: | School of Computer Science and Engineering | Rights: | © 2024 IOP Publishing Ltd. All rights reserved. | Fulltext Permission: | none | Fulltext Availability: | No Fulltext |
Appears in Collections: | SCSE Journal Articles |
SCOPUSTM
Citations
50
4
Updated on May 4, 2025
Page view(s)
95
Updated on May 6, 2025
Google ScholarTM
Check
Altmetric
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.