Music genre classification with parallel convolutional neural networks and capuchin search algorithm

基于并行卷积神经网络和卷尾猴搜索算法的音乐流派分类

阅读:1

Abstract

With the primary objective of creating playlists that suggest songs, interest in music genre categorization has grown thanks to high-tech multimedia tools. To develop a strong music classifier that can quickly classify unlabeled music and enhance consumers' experiences with media players and music files, machine learning and deep learning ideas are required. This study presents a unique method that blends convolutional neural network (CNN) models as an ensemble system to detect musical genres. The method makes use of discrete wavelet transform (DWT), mel frequency cepstral coefficients (MFCC), and short-time fourier transform (STFT) characteristics to provide a comprehensive framework for expressing stylistic qualities in music. To do this, each model's hyperparameters are generated using the capuchin search algorithm (CapSA). Preprocessing the original signals, feature description utilizing DWT, MFCC, and STFT signal matrices, CNN model optimization to extract signal features, and music genre identification based on combined features make up the four main components of the technique. By integrating many signal processing techniques and CNN models, this study advances the field of music genre classification and provides possible insights into the blending of diverse musical components for improved classification accuracy. The GTZAN and Extended-Ballroom datasets were the two used in the studies. The average classification accuracy of 96.07 and 96.20 for each database, respectively, show how well our suggested strategy performs when compared to earlier, comparable methods.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。