Facial expression recognition in-the-wild using blended feature attention network

dc.contributor.authorKarnati, Mohan
dc.contributor.authorSeal, Ayan
dc.contributor.authorJaworek-Korjakowska, Joanna
dc.contributor.authorKrejcar, Ondřej
dc.date.accessioned2024-03-28T15:20:00Z
dc.date.available2024-03-28T15:20:00Z
dc.date.issued2023
dc.description.abstractFacial expression (FE) analysis plays a crucial role in various fields, such as affective computing, marketing, and clinical evaluation. Despite numerous advances, research on FE recognition (FER) has recently been proceeding from confined laboratory circumstances to in-the-wild environments. FER is still an arduous and demanding problem due to occlusion and pose changes, intraclass and intensity variations caused by illumination, and insufficient training data. Most state-of-the-art (SOTA) approaches use the entire face for FER. However, past studies on psychology and physiology reveal that the mouth and eyes reflect the variations of various FEs, which are closely related to the manifestation of emotion. A novel method is proposed in this study to address some of the issues mentioned above. First, modified homomorphic filtering (MHF) is employed to normalize the illumination, then the normalized face image is cropped into five local regions to emphasize expression-specific characteristics. Finally, a unique blended feature attention network (BFAN) is designed for FER. BFAN consists of both residual dilated multiscale (RDMS) feature extraction modules and spatial and channel-wise attention (CWA) modules. These modules help to extract the most relevant and discriminative features from the high-level (HL) and low-level (LL) features. Then, both feature maps are integrated and passed on to the dense layers followed by a softmax layer to compute probability scores. Finally, the Choquet fuzzy integral is applied to the computed probability scores to get the final outcome. The superiority of the proposed method is exemplified by comparing it with 18 existing approaches on seven benchmark datasets.cs
dc.description.firstpageart. no. 5026416cs
dc.description.sourceWeb of Sciencecs
dc.description.volume72cs
dc.identifier.citationIEEE Transactions on Instrumentation and Measurement. 2023, vol. 72, art. no. 5026416.cs
dc.identifier.doi10.1109/TIM.2023.3314815
dc.identifier.issn0018-9456
dc.identifier.issn1557-9662
dc.identifier.urihttp://hdl.handle.net/10084/152484
dc.identifier.wos001083291000016
dc.language.isoencs
dc.publisherIEEEcs
dc.relation.ispartofseriesIEEE Transactions on Instrumentation and Measurementcs
dc.relation.urihttps://doi.org/10.1109/TIM.2023.3314815cs
dc.rightsCopyright © 2023, IEEEcs
dc.subjectattention mechanismcs
dc.subjectfacial expression recognition (FER)cs
dc.subjectfuzzy integralcs
dc.subjectilluminationcs
dc.subjectintensity variationscs
dc.subjectocclusion and pose robustcs
dc.subjectstatistical significancecs
dc.titleFacial expression recognition in-the-wild using blended feature attention networkcs
dc.typearticlecs
dc.type.statusPeer-reviewedcs

Files

License bundle

Now showing 1 - 1 out of 1 results
Loading...
Thumbnail Image
Name:
license.txt
Size:
718 B
Format:
Item-specific license agreed upon to submission
Description: