• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基尼特分类在埃塞俄比亚圣歌、阿兹玛里斯和现代音乐中的应用:一个新数据集和 CNN 基准。

Kiñit classification in Ethiopian chants, Azmaris and modern music: A new dataset and CNN benchmark.

机构信息

School of Information Science and Technology, Northwest University, Xi'an, China.

School of Computer Science and Electronic Engineering, University of Essex, Colchester, United Kingdom.

出版信息

PLoS One. 2023 Apr 20;18(4):e0284560. doi: 10.1371/journal.pone.0284560. eCollection 2023.

DOI:10.1371/journal.pone.0284560
PMID:37079543
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10118148/
Abstract

In this paper, we create EMIR, the first-ever Music Information Retrieval dataset for Ethiopian music. EMIR is freely available for research purposes and contains 600 sample recordings of Orthodox Tewahedo chants, traditional Azmari songs and contemporary Ethiopian secular music. Each sample is classified by five expert judges into one of four well-known Ethiopian Kiñits, Tizita, Bati, Ambassel and Anchihoye. Each Kiñit uses its own pentatonic scale and also has its own stylistic characteristics. Thus, Kiñit classification needs to combine scale identification with genre recognition. After describing the dataset, we present the Ethio Kiñits Model (EKM), based on VGG, for classifying the EMIR clips. In Experiment 1, we investigated whether Filterbank, Mel-spectrogram, Chroma, or Mel-frequency Cepstral coefficient (MFCC) features work best for Kiñit classification using EKM. MFCC was found to be superior and was therefore adopted for Experiment 2, where the performance of EKM models using MFCC was compared using three different audio sample lengths. 3s length gave the best results. In Experiment 3, EKM and four existing models were compared on the EMIR dataset: AlexNet, ResNet50, VGG16 and LSTM. EKM was found to have the best accuracy (95.00%) as well as the fastest training time. However, the performance of VGG16 (93.00%) was found not to be significantly worse (P < 0.01). We hope this work will encourage others to explore Ethiopian music and to experiment with other models for Kiñit classification.

摘要

在本文中,我们创建了 EMIR,这是第一个针对埃塞俄比亚音乐的音乐信息检索数据集。EMIR 可供研究使用,其中包含 600 个样本录音,包括东正教特瓦赫多颂歌、传统的阿兹马里歌曲和当代埃塞俄比亚世俗音乐。每个样本由五名专家评委分为四个著名的埃塞俄比亚 Kiñit 之一,即 Tizita、Bati、Ambassel 和 Anchihoye。每个 Kiñit 使用自己的五声音阶,也有自己的风格特点。因此,Kini 分类需要将音阶识别与流派识别结合起来。在描述数据集之后,我们提出了基于 VGG 的 Ethio Kiñits Model (EKM),用于对 EMIR 剪辑进行分类。在实验 1 中,我们研究了使用 EKM 时,Filterbank、Mel-spectrogram、Chroma 或 Mel-frequency Cepstral coefficient (MFCC) 特征是否最适合 Kiñit 分类。结果发现 MFCC 效果最佳,因此在实验 2 中采用了 MFCC,比较了使用三种不同音频样本长度的 EKM 模型的性能。3s 长度的结果最佳。在实验 3 中,我们在 EMIR 数据集上比较了 EKM 和四个现有模型:AlexNet、ResNet50、VGG16 和 LSTM。结果发现 EKM 的准确率最高(95.00%),训练时间最快。然而,VGG16(93.00%)的性能并不明显更差(P < 0.01)。我们希望这项工作将鼓励其他人探索埃塞俄比亚音乐,并尝试使用其他模型进行 Kiñit 分类。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/786a991fe8f3/pone.0284560.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/f99782b04148/pone.0284560.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/c66f25f9836c/pone.0284560.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/73b7845f914d/pone.0284560.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/23aa1e1c03b1/pone.0284560.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/3caf4ff690fa/pone.0284560.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/786a991fe8f3/pone.0284560.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/f99782b04148/pone.0284560.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/c66f25f9836c/pone.0284560.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/73b7845f914d/pone.0284560.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/23aa1e1c03b1/pone.0284560.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/3caf4ff690fa/pone.0284560.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/deaa/10118148/786a991fe8f3/pone.0284560.g006.jpg

相似文献

1
Kiñit classification in Ethiopian chants, Azmaris and modern music: A new dataset and CNN benchmark.基尼特分类在埃塞俄比亚圣歌、阿兹玛里斯和现代音乐中的应用:一个新数据集和 CNN 基准。
PLoS One. 2023 Apr 20;18(4):e0284560. doi: 10.1371/journal.pone.0284560. eCollection 2023.
2
A Music Emotion Classification Model Based on the Improved Convolutional Neural Network.基于改进卷积神经网络的音乐情绪分类模型。
Comput Intell Neurosci. 2022 Feb 14;2022:6749622. doi: 10.1155/2022/6749622. eCollection 2022.
3
Speaker-based language identification for Ethio-Semitic languages using CRNN and hybrid features.使用卷积循环神经网络(CRNN)和混合特征的基于说话者的埃塞俄比亚闪米特语语言识别
Network. 2024 Jun 4:1-23. doi: 10.1080/0954898X.2024.2359610.
4
Optimizing the configuration of deep learning models for music genre classification.优化用于音乐流派分类的深度学习模型配置。
Heliyon. 2024 Jan 17;10(2):e24892. doi: 10.1016/j.heliyon.2024.e24892. eCollection 2024 Jan 30.
5
Rigdelet neural network and improved partial reinforcement effect optimizer for music genre classification from sound spectrum images.用于基于声谱图像的音乐流派分类的Rigdelet神经网络和改进的部分强化效应优化器。
Heliyon. 2024 Jul 4;10(14):e34067. doi: 10.1016/j.heliyon.2024.e34067. eCollection 2024 Jul 30.
6
A hierarchical approach for speech-instrumental-song classification.一种用于语音-器乐-歌曲分类的分层方法。
Springerplus. 2013 Oct 17;2(1):526. doi: 10.1186/2193-1801-2-526. eCollection 2013.
7
Pop Music Singing in Education with Modern Innovative Technologies: How the Chinese Language Shapes the Creation of Popular Singing.现代创新技术在流行音乐演唱教育中的应用:中文如何塑造流行演唱创作
J Psycholinguist Res. 2023 Dec;52(6):2677-2691. doi: 10.1007/s10936-023-10014-z. Epub 2023 Sep 14.
8
A Multimodal Convolutional Neural Network Model for the Analysis of Music Genre on Children's Emotions Influence Intelligence.用于分析音乐类型对儿童情绪智力影响的多模态卷积神经网络模型。
Comput Intell Neurosci. 2022 Aug 29;2022:5611456. doi: 10.1155/2022/5611456. eCollection 2022.
9
Construction of Music Intelligent Creation Model Based on Convolutional Neural Network.基于卷积神经网络的音乐智能创作模型构建。
Comput Intell Neurosci. 2022 Jul 5;2022:2854066. doi: 10.1155/2022/2854066. eCollection 2022.
10
A computational lens into how music characterizes genre in film.计算镜头下的电影音乐如何塑造流派特征。
PLoS One. 2021 Apr 8;16(4):e0249957. doi: 10.1371/journal.pone.0249957. eCollection 2021.

引用本文的文献

1
Apple varieties, diseases, and distinguishing between fresh and rotten through deep learning approaches.通过深度学习方法对苹果品种、病害以及新鲜与腐烂进行区分。
PLoS One. 2025 May 15;20(5):e0322586. doi: 10.1371/journal.pone.0322586. eCollection 2025.

本文引用的文献

1
A computational lens into how music characterizes genre in film.计算镜头下的电影音乐如何塑造流派特征。
PLoS One. 2021 Apr 8;16(4):e0249957. doi: 10.1371/journal.pone.0249957. eCollection 2021.
2
A computational study on outliers in world music.世界音乐中异常值的计算研究。
PLoS One. 2017 Dec 18;12(12):e0189399. doi: 10.1371/journal.pone.0189399. eCollection 2017.
3
Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms.用于比较监督分类学习算法的近似统计检验
Neural Comput. 1998 Sep 15;10(7):1895-1923. doi: 10.1162/089976698300017197.