School of Music, Henan Vocational Institute of Arts, Zhengzhou, Henan, China.
Comput Intell Neurosci. 2022 May 18;2022:4439738. doi: 10.1155/2022/4439738. eCollection 2022.
Wireless networks are commonly employed for ambient assisted living applications, and artificial intelligence-enabled event detection and classification processes have become familiar. However, music is a kind of time-series data, and it is challenging to design an effective music genre classification (MGC) system due to a large quantity of music data. Robust MGC techniques necessitate a massive amount of data, which is time-consuming, laborious, and requires expert knowledge. Few studies have focused on the design of music representations extracted directly from input waveforms. In recent times, deep learning (DL) models have been widely used due to their characteristics of automatic extracting advanced features and contextual representation from actual music or processed data. This paper aims to develop a novel deep learning-enabled music genre classification (DLE-MGC) technique. The proposed DLE-MGC technique effectively classifies the music genres into multiple classes by using three subprocesses, namely preprocessing, classification, and hyperparameter optimization. At the initial stage, the Pitch to Vector (Pitch2vec) approach is applied as a preprocessing step where the pitches in the input musical instrument digital interface (MIDI) files are transformed into the vector sequences. Besides, the DLE-MGC technique involves the design of a cat swarm optimization (CSO) with bidirectional long-term memory (BiLSTM) model for the classification process. The DBTMPE technique has gained a moderately increased accuracy of 94.27%, and the DLE-MGC technique has accomplished a better accuracy of 95.87%. The performance validation of the DLE-MGC technique was carried out using the Lakh MIDI music dataset, and the comparative results verified the promising performance of the DLE-MGC technique over current methods.
无线网络常用于环境辅助生活应用,人工智能支持的事件检测和分类过程已经变得熟悉。然而,音乐是一种时间序列数据,由于音乐数据量大,设计有效的音乐类型分类 (MGC) 系统具有挑战性。稳健的 MGC 技术需要大量的数据,这既耗时、费力,又需要专业知识。很少有研究关注从输入波形中直接提取音乐表示的设计。近年来,由于深度学习 (DL) 模型从实际音乐或处理后的数据中自动提取高级特征和上下文表示的特点,它们得到了广泛的应用。本文旨在开发一种新的基于深度学习的音乐类型分类 (DLE-MGC) 技术。所提出的 DLE-MGC 技术通过使用三个子过程,即预处理、分类和超参数优化,有效地将音乐类型分类为多个类别。在初始阶段,应用音高到向量 (Pitch2vec) 方法作为预处理步骤,将输入乐器数字接口 (MIDI) 文件中的音高转换为向量序列。此外,DLE-MGC 技术还涉及设计具有双向长短期记忆 (BiLSTM) 模型的猫群优化 (CSO) 进行分类过程。DBTMPE 技术的准确率提高到了 94.27%,而 DLE-MGC 技术的准确率提高到了 95.87%。使用 Lakh MIDI 音乐数据集对 DLE-MGC 技术进行了性能验证,比较结果验证了 DLE-MGC 技术优于现有方法的良好性能。