• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CMU-MOSEAS:一个用于西班牙语、葡萄牙语、德语和法语的多模态语言数据集。

CMU-MOSEAS: A Multimodal Language Dataset for Spanish, Portuguese, German and French.

作者信息

Zadeh Amir, Cao Yan Sheng, Hessner Simon, Liang Paul Pu, Poria Soujanya, Morency Louis-Philippe

机构信息

LTI, SCS, CMU.

SCS, CMU.

出版信息

Proc Conf Empir Methods Nat Lang Process. 2020 Nov;2020:1801-1812. doi: 10.18653/v1/2020.emnlp-main.141.

DOI:10.18653/v1/2020.emnlp-main.141
PMID:33969362
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8106386/
Abstract

Modeling multimodal language is a core research area in natural language processing. While languages such as English have relatively large multimodal language resources, other widely spoken languages across the globe have few or no large-scale datasets in this area. This disproportionately affects native speakers of languages other than English. As a step towards building more equitable and inclusive multimodal systems, we introduce the first large-scale multimodal language dataset for Spanish, Portuguese, German and French. The proposed dataset, called CMU-MOSEAS (CMU Multimodal Opinion Sentiment, Emotions and Attributes), is the largest of its kind with 40, 000 total labelled sentences. It covers a diverse set topics and speakers, and carries supervision of 20 labels including sentiment (and subjectivity), emotions, and attributes. Our evaluations on a state-of-the-art multimodal model demonstrates that CMU-MOSEAS enables further research for multilingual studies in multimodal language.

摘要

多模态语言建模是自然语言处理中的一个核心研究领域。虽然像英语这样的语言拥有相对丰富的多模态语言资源,但全球其他广泛使用的语言在这一领域几乎没有或根本没有大规模数据集。这对非英语母语者产生了不成比例的影响。作为构建更公平、更具包容性的多模态系统的一步,我们推出了首个针对西班牙语、葡萄牙语、德语和法语的大规模多模态语言数据集。这个名为CMU-MOSEAS(卡内基梅隆大学多模态观点、情感、情绪和属性)的数据集是同类数据集中最大的,共有40000个带标签的句子。它涵盖了各种不同的主题和说话者,并带有包括情感(和主观性)、情绪及属性在内的20种标签的标注。我们在一个先进的多模态模型上进行的评估表明,CMU-MOSEAS能够推动多模态语言中多语言研究的进一步发展。

相似文献

1
CMU-MOSEAS: A Multimodal Language Dataset for Spanish, Portuguese, German and French.CMU-MOSEAS:一个用于西班牙语、葡萄牙语、德语和法语的多模态语言数据集。
Proc Conf Empir Methods Nat Lang Process. 2020 Nov;2020:1801-1812. doi: 10.18653/v1/2020.emnlp-main.141.
2
Lexical simplification benchmarks for English, Portuguese, and Spanish.英语、葡萄牙语和西班牙语的词汇简化基准。
Front Artif Intell. 2022 Sep 22;5:991242. doi: 10.3389/frai.2022.991242. eCollection 2022.
3
Building lexicon-based sentiment analysis model for low-resource languages.为低资源语言构建基于词典的情感分析模型。
MethodsX. 2023 Oct 22;11:102460. doi: 10.1016/j.mex.2023.102460. eCollection 2023 Dec.
4
A multimodal approach to cross-lingual sentiment analysis with ensemble of transformer and LLM.一种结合Transformer和大语言模型集成的跨语言情感分析多模态方法。
Sci Rep. 2024 Apr 26;14(1):9603. doi: 10.1038/s41598-024-60210-7.
5
Integrating Multimodal Information in Large Pretrained Transformers.在大型预训练变压器中整合多模态信息。
Proc Conf Assoc Comput Linguist Meet. 2020 Jul;2020:2359-2369. doi: 10.18653/v1/2020.acl-main.214.
6
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model.AFR-BERT:基于注意力机制的特征相关融合多模态情感分析模型。
PLoS One. 2022 Sep 9;17(9):e0273936. doi: 10.1371/journal.pone.0273936. eCollection 2022.
7
Multilingual event extraction for epidemic detection.用于疫情检测的多语言事件提取
Artif Intell Med. 2015 Oct;65(2):131-43. doi: 10.1016/j.artmed.2015.06.005. Epub 2015 Jul 17.
8
Semi-supervised word polarity identification in resource-lean languages.资源匮乏语言中的半监督词极性识别
Neural Netw. 2014 Oct;58:50-9. doi: 10.1016/j.neunet.2014.05.018. Epub 2014 Jun 4.
9
Real and synthetic Punjabi speech datasets for automatic speech recognition.用于自动语音识别的真实和合成旁遮普语语音数据集。
Data Brief. 2023 Nov 27;52:109865. doi: 10.1016/j.dib.2023.109865. eCollection 2024 Feb.
10
iSentenizer-μ: multilingual sentence boundary detection model.iSentenizer-μ:多语言句子边界检测模型。
ScientificWorldJournal. 2014;2014:196574. doi: 10.1155/2014/196574. Epub 2014 Apr 15.

引用本文的文献

1
Decision level scheme for fusing multiomics and histology slide images using deep neural network for tumor prognosis prediction.使用深度神经网络融合多组学和组织学切片图像以进行肿瘤预后预测的决策水平方案。
Sci Rep. 2025 Jul 15;15(1):25479. doi: 10.1038/s41598-025-09869-0.
2
Developing a Machine Learning-Based Automated Patient Engagement Estimator for Telehealth: Algorithm Development and Validation Study.开发一种基于机器学习的远程医疗自动患者参与度评估器:算法开发与验证研究。
JMIR Form Res. 2025 Jan 20;9:e46390. doi: 10.2196/46390.
3
A dataset for multimodal music information retrieval of Sotho-Tswana musical videos.一个用于索托-茨瓦纳音乐视频多模态音乐信息检索的数据集。
Data Brief. 2024 Jun 26;55:110672. doi: 10.1016/j.dib.2024.110672. eCollection 2024 Aug.
4
MultiBench: Multiscale Benchmarks for Multimodal Representation Learning.多模态基准测试:用于多模态表示学习的多尺度基准测试
Adv Neural Inf Process Syst. 2021 Dec;2021(DB1):1-20.
5
Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion.基于凝聚注意力融合的对比学习的多模态情感分析表示学习。
Sensors (Basel). 2023 Mar 1;23(5):2679. doi: 10.3390/s23052679.

本文引用的文献

1
ArcFace: Additive Angular Margin Loss for Deep Face Recognition.ArcFace:用于深度人脸识别的附加角度间隔损失。
IEEE Trans Pattern Anal Mach Intell. 2022 Oct;44(10):5962-5979. doi: 10.1109/TPAMI.2021.3087709. Epub 2022 Sep 14.
2
Multimodal Transformer for Unaligned Multimodal Language Sequences.用于未对齐多模态语言序列的多模态变换器
Proc Conf Assoc Comput Linguist Meet. 2019 Jul;2019:6558-6569. doi: 10.18653/v1/p19-1656.
3
Mutual Correlation Attentive Factors in Dyadic Fusion Networks for Speech Emotion Recognition.用于语音情感识别的二元融合网络中的互相关注意因素
Proc ACM Int Conf Multimed. 2019 Oct;2019:157-166. doi: 10.1145/3343031.3351039.
4
Why We Should Study Multimodal Language.我们为何要研究多模态语言。
Front Psychol. 2018 Jun 28;9:1109. doi: 10.3389/fpsyg.2018.01109. eCollection 2018.
5
Normalized amplitude quotient for parametrization of the glottal flow.用于声门气流参数化的归一化幅度商。
J Acoust Soc Am. 2002 Aug;112(2):701-10. doi: 10.1121/1.1490365.
6
Vocal quality factors: analysis, synthesis, and perception.嗓音质量因素:分析、合成与感知。
J Acoust Soc Am. 1991 Nov;90(5):2394-410. doi: 10.1121/1.402044.
7
Vocal intensity in speakers and singers.演讲者和歌手的发声强度。
J Acoust Soc Am. 1992 May;91(5):2936-46. doi: 10.1121/1.402929.