Deep Learning Model Compression and Hardware Acceleration for High-Performance Foreign Material Detection on Poultry Meat Using NIR Hyperspectral Imaging

基于近红外高光谱成像的禽肉异物高性能检测:深度学习模型压缩与硬件加速

阅读:1

Abstract

Ensuring the safety and quality of poultry products requires efficient detection and removal of foreign materials during processing. Hyperspectral imaging (HSI) offers a non-invasive mechanism to capture detailed spatial and spectral information, enabling the discrimination of different types of contaminants from poultry muscle and non-muscle external tissues. When integrated with advanced deep learning (DL) models, HSI systems can achieve high accuracy in detecting foreign materials. However, the high dimensionality of HSI data, the computational complexity of DL models, and the high-paced nature of poultry processing environments pose challenges for real-time implementation in industrial settings, where the speed of imaging and decision-making is critical. In this study, we address these challenges by optimizing DL inference for HSI-based foreign material detection through a combination of post-training quantization and hardware acceleration techniques. We leveraged hardware acceleration utilizing the TensorRT module for NVIDIA GPU to enhance inference speed. Additionally, we applied half-precision (called FP16) post-training quantization to reduce the precision of model parameters, decreasing memory usage and computational requirements without any loss in model accuracy. We conducted simulations using two hypothetical hyperspectral line-scan cameras to evaluate the feasibility of real-time detection in industrial conditions. The simulation results demonstrated that our optimized models could achieve inference times compatible with the line speeds of poultry processing lines between 140 and 250 birds per minute, indicating the potential for real-time deployment. Specifically, the proposed inference method, optimized through hardware acceleration and model compression, achieved reductions in inference time of up to five times compared to unoptimized, traditional GPU-based inference. In addition, it resulted in a 50% decrease in model size while maintaining high detection accuracy that was also comparable to the original model. Our findings suggest that the integration of post-training quantization and hardware acceleration is an effective strategy for overcoming the computational bottlenecks associated with DL inference on HSI data.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。