Automated Detection and Classification of Marine Species Vocalizations Using a YOLO-Based Deep Learning Framework

基于YOLO深度学习框架的海洋物种发声自动检测与分类

阅读:2

Abstract

The underwater acoustic environment is highly complex, where signals from various natural and anthropogenic sources interact and overlap, making monitoring efforts very challenging. Thus, effective detection and classification mechanisms are vital, as they provide key information about marine species and help in understanding how human activities influence the overall marine environment. This study proposes a deep learning-based framework for the automatic detection and classification of marine species vocalizations, inspired by the YOLO (You Only Look Once) architecture. However, a major limitation in developing such frameworks is the limited availability of continuous, well-annotated monitoring datasets that contain multi-species recordings. To address this limitation, synthetic monitoring datasets were constructed by combining single-species vocalizations to simulate realistic monitoring conditions under both non-overlapping and overlapping scenarios. Augmentation techniques, including CutMix, were implemented to enhance dataset diversity and improve the model's robustness against signal overlap. Experimental results demonstrated that the proposed model achieves strong performance under non-overlapping conditions and maintains stable detection and classification performance even in overlapping scenarios. These findings suggest that YOLO-inspired architectures can achieve effective performance across various acoustic conditions. Future studies should focus on incorporating continuous, long-term field recordings to further improve detection and classification reliability.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。