Driver facial emotion tracking using an enhanced residual network with weighted fusion of channel and spatial attention.

阅读:4
作者:Duongthang, Long
Facial expression recognition (FER) plays a crucial role in interpreting human emotions and intentions in real-life applications, such as advanced driver assistance systems. However, it faces challenges due to subtle facial variations, environmental factors, and occlusions. In this paper, we propose a novel CNN-based model for driver facial emotion tracking, named FARNet, which incorporates residual connections and is inspired by vision transformer architectures. The model integrates a fusion of channel and spatial attention mechanisms with learnable weights to enhance FER performance while maintaining moderate complexity. It comprises four stages with residual blocks in a 2:2:4:2 ratio and approximately 3.05 million parameters, making it parameter-efficient compared to existing models. We evaluate FARNet on five popular FER datasets: CK+, OuluCASIA, RAF-DB, FER+, and AffectNet. The model achieves the highest accuracy on three datasets and the second-highest on the rest, with results ranging from 57.03% on AffectNet to 100% on CK + and OuluCASIA, remaining competitive against other methods.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。