• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

EAV:用于对话情境中情绪识别的 EEG-Audio-Video 数据集。

EAV: EEG-Audio-Video Dataset for Emotion Recognition in Conversational Contexts.

机构信息

Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan.

Korea University, Department of Artificial Intelligence, Seoul, 02841, Republic of Korea.

出版信息

Sci Data. 2024 Sep 19;11(1):1026. doi: 10.1038/s41597-024-03838-4.

DOI:10.1038/s41597-024-03838-4
PMID:39300129
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11413008/
Abstract

Understanding emotional states is pivotal for the development of next-generation human-machine interfaces. Human behaviors in social interactions have resulted in psycho-physiological processes influenced by perceptual inputs. Therefore, efforts to comprehend brain functions and human behavior could potentially catalyze the development of AI models with human-like attributes. In this study, we introduce a multimodal emotion dataset comprising data from 30-channel electroencephalography (EEG), audio, and video recordings from 42 participants. Each participant engaged in a cue-based conversation scenario, eliciting five distinct emotions: neutral, anger, happiness, sadness, and calmness. Throughout the experiment, each participant contributed 200 interactions, which encompassed both listening and speaking. This resulted in a cumulative total of 8,400 interactions across all participants. We evaluated the baseline performance of emotion recognition for each modality using established deep neural network (DNN) methods. The Emotion in EEG-Audio-Visual (EAV) dataset represents the first public dataset to incorporate three primary modalities for emotion recognition within a conversational context. We anticipate that this dataset will make significant contributions to the modeling of the human emotional process, encompassing both fundamental neuroscience and machine learning viewpoints.

摘要

理解情绪状态对于下一代人机接口的发展至关重要。社交互动中的人类行为导致了受感知输入影响的心理生理过程。因此,努力理解大脑功能和人类行为有可能促进具有类人属性的人工智能模型的发展。在这项研究中,我们引入了一个多模态情感数据集,该数据集包含来自 42 名参与者的 30 通道脑电图 (EEG)、音频和视频记录的数据。每个参与者都参与了基于提示的对话场景,引发了五种不同的情绪:中性、愤怒、快乐、悲伤和平静。在整个实验过程中,每个参与者贡献了 200 次交互,包括听和说。这导致所有参与者的总交互次数达到 8400 次。我们使用已建立的深度神经网络 (DNN) 方法评估了每个模态的情感识别基线性能。EAV 数据集是第一个在对话环境中整合三种主要模态进行情感识别的公共数据集。我们预计,该数据集将对人类情感过程的建模做出重大贡献,涵盖基础神经科学和机器学习的观点。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/976f8fdbd392/41597_2024_3838_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/52c0d188d98f/41597_2024_3838_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/5bcc175ce602/41597_2024_3838_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/3790052e47df/41597_2024_3838_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/38043e67ab3a/41597_2024_3838_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/37ad171b1108/41597_2024_3838_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/976f8fdbd392/41597_2024_3838_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/52c0d188d98f/41597_2024_3838_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/5bcc175ce602/41597_2024_3838_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/3790052e47df/41597_2024_3838_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/38043e67ab3a/41597_2024_3838_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/37ad171b1108/41597_2024_3838_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e44a/11413008/976f8fdbd392/41597_2024_3838_Fig6_HTML.jpg

相似文献

1
EAV: EEG-Audio-Video Dataset for Emotion Recognition in Conversational Contexts.EAV:用于对话情境中情绪识别的 EEG-Audio-Video 数据集。
Sci Data. 2024 Sep 19;11(1):1026. doi: 10.1038/s41597-024-03838-4.
2
Emotion recognition from single-channel EEG signals using a two-stage correlation and instantaneous frequency-based filtering method.基于两级相关和基于瞬时频率的滤波方法从单通道 EEG 信号中进行情绪识别。
Comput Methods Programs Biomed. 2019 May;173:157-165. doi: 10.1016/j.cmpb.2019.03.015. Epub 2019 Mar 22.
3
Spatial-frequency-temporal convolutional recurrent network for olfactory-enhanced EEG emotion recognition.基于空间频率-时间卷积循环网络的嗅觉增强脑电情感识别
J Neurosci Methods. 2022 Jul 1;376:109624. doi: 10.1016/j.jneumeth.2022.109624. Epub 2022 May 16.
4
EEG-based emotion charting for Parkinson's disease patients using Convolutional Recurrent Neural Networks and cross dataset learning.基于 EEG 的帕金森病患者情绪图表分析,使用卷积循环神经网络和跨数据集学习。
Comput Biol Med. 2022 May;144:105327. doi: 10.1016/j.compbiomed.2022.105327. Epub 2022 Mar 11.
5
Investigating the Use of Pretrained Convolutional Neural Network on Cross-Subject and Cross-Dataset EEG Emotion Recognition.研究基于预训练卷积神经网络的跨被试和跨数据集 EEG 情绪识别
Sensors (Basel). 2020 Apr 4;20(7):2034. doi: 10.3390/s20072034.
6
Emotion Detection from EEG Signals Using Machine Deep Learning Models.使用机器深度学习模型从脑电图信号中进行情绪检测。
Bioengineering (Basel). 2024 Aug 2;11(8):782. doi: 10.3390/bioengineering11080782.
7
AVaTER: Fusing Audio, Visual, and Textual Modalities Using Cross-Modal Attention for Emotion Recognition.AVaTER:使用跨模态注意力融合音频、视觉和文本模态进行情感识别。
Sensors (Basel). 2024 Sep 10;24(18):5862. doi: 10.3390/s24185862.
8
Deep Learning-Based Approach for Emotion Recognition Using Electroencephalography (EEG) Signals Using Bi-Directional Long Short-Term Memory (Bi-LSTM).基于深度学习的脑电(EEG)信号情绪识别方法:使用双向长短时记忆网络(Bi-LSTM)。
Sensors (Basel). 2022 Apr 13;22(8):2976. doi: 10.3390/s22082976.
9
Spatio-Temporal Representation of an Electoencephalogram for Emotion Recognition Using a Three-Dimensional Convolutional Neural Network.使用三维卷积神经网络进行情感识别的脑电图的时空表示。
Sensors (Basel). 2020 Jun 20;20(12):3491. doi: 10.3390/s20123491.
10
Investigating EEG-based functional connectivity patterns for multimodal emotion recognition.研究基于 EEG 的功能连接模式进行多模态情感识别。
J Neural Eng. 2022 Jan 31;19(1). doi: 10.1088/1741-2552/ac49a7.

引用本文的文献

1
Stress and Emotion Open Access Data: A Review on Datasets, Modalities, Methods, Challenges, and Future Research Perspectives.压力与情绪开放获取数据:关于数据集、模态、方法、挑战及未来研究展望的综述
J Healthc Inform Res. 2025 Jun 18;9(3):247-279. doi: 10.1007/s41666-025-00200-0. eCollection 2025 Sep.

本文引用的文献

1
EEGformer: A transformer-based brain activity classification method using EEG signal.EEGformer:一种基于变压器的使用脑电图信号的脑活动分类方法。
Front Neurosci. 2023 Mar 24;17:1148855. doi: 10.3389/fnins.2023.1148855. eCollection 2023.
2
Spatio-Spectral Feature Representation for Motor Imagery Classification Using Convolutional Neural Networks.基于卷积神经网络的运动想象分类的时空频谱特征表示。
IEEE Trans Neural Netw Learn Syst. 2022 Jul;33(7):3038-3049. doi: 10.1109/TNNLS.2020.3048385. Epub 2022 Jul 6.
3
K-EmoCon, a multimodal sensor dataset for continuous emotion recognition in naturalistic conversations.
K-EmoCon,一个用于自然会话中连续情感识别的多模态传感器数据集。
Sci Data. 2020 Sep 8;7(1):293. doi: 10.1038/s41597-020-00630-y.
4
Subject-Independent Brain-Computer Interfaces Based on Deep Convolutional Neural Networks.基于深度卷积神经网络的独立于主体的脑机接口
IEEE Trans Neural Netw Learn Syst. 2020 Oct;31(10):3839-3852. doi: 10.1109/TNNLS.2019.2946869. Epub 2019 Nov 13.
5
EEG dataset and OpenBMI toolbox for three BCI paradigms: an investigation into BCI illiteracy.脑电数据集和 OpenBMI 工具箱,用于三种脑机接口范式:对脑机接口文盲现象的研究。
Gigascience. 2019 May 1;8(5). doi: 10.1093/gigascience/giz002.
6
EmotionMeter: A Multimodal Framework for Recognizing Human Emotions.情绪计量器:一种用于识别人类情绪的多模态框架。
IEEE Trans Cybern. 2019 Mar;49(3):1110-1122. doi: 10.1109/TCYB.2018.2797176. Epub 2018 Feb 8.
7
EEGNet: a compact convolutional neural network for EEG-based brain-computer interfaces.EEGNet:一种基于 EEG 的脑机接口用的紧凑卷积神经网络。
J Neural Eng. 2018 Oct;15(5):056013. doi: 10.1088/1741-2552/aace8c. Epub 2018 Jun 22.
8
The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English.瑞尔森情感语音和歌曲音频视频数据库(RAVDESS):一组具有北美英语特色的动态、多模态面部和声音表情数据集。
PLoS One. 2018 May 16;13(5):e0196391. doi: 10.1371/journal.pone.0196391. eCollection 2018.
9
Deep learning with convolutional neural networks for EEG decoding and visualization.基于卷积神经网络的 EEG 解码和可视化深度学习。
Hum Brain Mapp. 2017 Nov;38(11):5391-5420. doi: 10.1002/hbm.23730. Epub 2017 Aug 7.
10
DREAMER: A Database for Emotion Recognition Through EEG and ECG Signals From Wireless Low-cost Off-the-Shelf Devices.DREAMER:一个通过无线低成本现成设备的 EEG 和 ECG 信号进行情感识别的数据库。
IEEE J Biomed Health Inform. 2018 Jan;22(1):98-107. doi: 10.1109/JBHI.2017.2688239. Epub 2017 Mar 27.