Zobrazit minimální záznam

dc.contributor.authorKarnati, Mohan
dc.contributor.authorSeal, Ayan
dc.contributor.authorJaworek-Korjakowska, Joanna
dc.contributor.authorKrejcar, Ondřej
dc.date.accessioned2024-03-28T15:20:00Z
dc.date.available2024-03-28T15:20:00Z
dc.date.issued2023
dc.identifier.citationIEEE Transactions on Instrumentation and Measurement. 2023, vol. 72, art. no. 5026416.cs
dc.identifier.issn0018-9456
dc.identifier.issn1557-9662
dc.identifier.urihttp://hdl.handle.net/10084/152484
dc.description.abstractFacial expression (FE) analysis plays a crucial role in various fields, such as affective computing, marketing, and clinical evaluation. Despite numerous advances, research on FE recognition (FER) has recently been proceeding from confined laboratory circumstances to in-the-wild environments. FER is still an arduous and demanding problem due to occlusion and pose changes, intraclass and intensity variations caused by illumination, and insufficient training data. Most state-of-the-art (SOTA) approaches use the entire face for FER. However, past studies on psychology and physiology reveal that the mouth and eyes reflect the variations of various FEs, which are closely related to the manifestation of emotion. A novel method is proposed in this study to address some of the issues mentioned above. First, modified homomorphic filtering (MHF) is employed to normalize the illumination, then the normalized face image is cropped into five local regions to emphasize expression-specific characteristics. Finally, a unique blended feature attention network (BFAN) is designed for FER. BFAN consists of both residual dilated multiscale (RDMS) feature extraction modules and spatial and channel-wise attention (CWA) modules. These modules help to extract the most relevant and discriminative features from the high-level (HL) and low-level (LL) features. Then, both feature maps are integrated and passed on to the dense layers followed by a softmax layer to compute probability scores. Finally, the Choquet fuzzy integral is applied to the computed probability scores to get the final outcome. The superiority of the proposed method is exemplified by comparing it with 18 existing approaches on seven benchmark datasets.cs
dc.language.isoencs
dc.publisherIEEEcs
dc.relation.ispartofseriesIEEE Transactions on Instrumentation and Measurementcs
dc.relation.urihttps://doi.org/10.1109/TIM.2023.3314815cs
dc.rightsCopyright © 2023, IEEEcs
dc.subjectattention mechanismcs
dc.subjectfacial expression recognition (FER)cs
dc.subjectfuzzy integralcs
dc.subjectilluminationcs
dc.subjectintensity variationscs
dc.subjectocclusion and pose robustcs
dc.subjectstatistical significancecs
dc.titleFacial expression recognition in-the-wild using blended feature attention networkcs
dc.typearticlecs
dc.identifier.doi10.1109/TIM.2023.3314815
dc.type.statusPeer-reviewedcs
dc.description.sourceWeb of Sciencecs
dc.description.volume72cs
dc.description.firstpageart. no. 5026416cs
dc.identifier.wos001083291000016


Soubory tohoto záznamu

SouboryVelikostFormátZobrazit

K tomuto záznamu nejsou připojeny žádné soubory.

Tento záznam se objevuje v následujících kolekcích

Zobrazit minimální záznam