Facial expression recognition (FER) plays a crucial role in interpreting human emotions and intentions in real-life applications, such as advanced driver assistance systems. However, it faces challenges due to subtle facial variations, environmental factors, and occlusions. In this paper, we propose a novel CNN-based model for driver facial emotion tracking, named FARNet, which incorporates residual connections and is inspired by vision transformer architectures. The model integrates a fusion of channel and spatial attention mechanisms with learnable weights to enhance FER performance while maintaining moderate complexity. It comprises four stages with residual blocks in a 2:2:4:2 ratio and approximately 3.05Â million parameters, making it parameter-efficient compared to existing models. We evaluate FARNet on five popular FER datasets: CK+, OuluCASIA, RAF-DB, FER+, and AffectNet. The model achieves the highest accuracy on three datasets and the second-highest on the rest, with results ranging from 57.03% on AffectNet to 100% on CKâ+âand OuluCASIA, remaining competitive against other methods.
Driver facial emotion tracking using an enhanced residual network with weighted fusion of channel and spatial attention.
阅读:4
作者:Duongthang, Long
| 期刊: | Scientific Reports | 影响因子: | 3.900 |
| 时间: | 2025 | 起止号: | 2025 Apr 12; 15(1):12675 |
| doi: | 10.1038/s41598-025-97451-z | ||
特别声明
1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。
2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。
3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。
4、投稿及合作请联系:info@biocloudy.com。
