A music structure analysis method based on beat feature and improved residual networks

一种基于节拍特征和改进残差网络的音乐结构分析方法

阅读:1

Abstract

In response to the issues of insufficient audio feature representation and insufficient model generalization ability in music structure analysis methods, a music structure analysis method based on beat feature fusion and an improved residual network was designed. Music structure analysis (MSA) contains two tasks, boundary detection and segment labeling. Boundary detection and Segment labeling are expected to accurately divide music segments and clarify the function. In this paper, a method is studied to accomplish the two tasks. First, the labels of music structure are refactored into 9 types and refined to each beat. Secondly, a beat-wise feature extraction is studied to segment music according to its beats and incorporates various acoustic features to achieve highly accurate segmentation. Then, a Resnet-34 based on a self-attentive mechanism is used to predict the category of each beat. Finally, a post-processing step is designed to filter predicted labels. The method is evaluated on the dataset SALAMI-IA, and experiments show that it is 3 percentage points higher than the current optimal method on HR3F. It is better than the most advanced methods on PWF and Sf.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。