Fused RGB and IR image based deep learning detection of dried laver bugak for robotic automation systems

基于融合RGB和红外图像的深度学习方法检测干紫菜,用于机器人自动化系统

阅读:1

Abstract

In Kimbugak manufacturing system, sorting and loading before frying are still performed manually, which imposes a heavy workload on workers and limits scalability. This study focuses on detecting and classifying the physical characteristics of dried laver bugak to enable robotic pick-and-place operations by developing a You Only Look Once (YOLO) and Real-time Detection Transformer (RT-DETR) deep learning detection model based on fused RGB and infrared (IR) images for integration into a robotic automation system. Through experiments, it was found that at least five physical classes are needed for effective robotic handling. A novel approach combining RGB and IR image fusion using the Visual Geometry Group 19-layer (VGG19) network is introduced to enhance the input quality for detection. Experimental results show that the YOLOv11l model significantly outperforms the YOLOv8s model, achieving an F1 score of 0.94 and a mean Average Precision at 0.5 (mAP@0.5) of 0.95. These results demonstrate that VGG-based image fusion with YOLOv11l is an optimal solution for classifying and locating dried laver bugak. This research highlights the importance of physical class definition, multimodal image fusion, and detector selection in developing an effective automated sorting and loading system for Kimbugak production.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。