Optimizing MobileNetV3 for multimodal eye gaze and emotion recognition via advanced pruning and quantisation techniques

通过先进的剪枝和量化技术优化 MobileNetV3,以实现多模态眼动和情绪识别。

阅读:1

Abstract

This study introduces an innovative approach for optimising visual understanding by leveraging MobileNet V3 for eye-gazing, eye-blinks, and emotional expressions recognition. The importance of visual cues, such as gaze direction, blinking patterns, and facial expressions, in various applications, including human-computer interaction and behavioural analysis, is driving the need for efficient and accurate models that operate effectively in resource-constrained environments. MobileNet V3 offers a strong foundation for such tasks due to its lightweight architecture; however, it can be enhanced further to provide even greater performance. To achieve this, we apply advanced model optimisation techniques, including pruning and quantisation, to reduce computational complexity without compromising accuracy. We validate our approach using three distinct datasets: EyeGaze, Emotions, and Closed Eye, which offer diverse visual inputs across different scenarios. The results demonstrate that our optimised MobileNet V3 model accurately detects and analyses eye gaze, blinks, and emotional expressions, making it a robust algorithm for real-world applications. All the codes for reprehensibility and trained models can be found at our github repository .

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。