Facial Expression Recognition in-the-wild using Blended Feature Attention Network
The result's identifiers
Result code in IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F62690094%3A18450%2F23%3A50020670" target="_blank" >RIV/62690094:18450/23:50020670 - isvavai.cz</a>
Result on the web
<a href="https://ieeexplore.ieee.org/document/10250983/" target="_blank" >https://ieeexplore.ieee.org/document/10250983/</a>
DOI - Digital Object Identifier
<a href="http://dx.doi.org/10.1109/TIM.2023.3314815" target="_blank" >10.1109/TIM.2023.3314815</a>
Alternative languages
Result language
angličtina
Original language name
Facial Expression Recognition in-the-wild using Blended Feature Attention Network
Original language description
Facial expression analysis plays a crucial role in various fields, such as affective computing, marketing, and clinical evaluation. Despite numerous advances, research on facial expression recognition (FER) has recently been proceeding from confined lab circumstances to <italic>in-the-wild</italic> environments. FER is still an arduous and demanding problem due to occlusion and pose changes, intra-class and intensity variations caused by illumination, and insufficient training data. Most state-of-the-art approaches use entire face for FER. However, the past studies on psychology and physiology reveals that mouth and eyes reflect the variations of various facial expressions, which are closely related to the manifestation of emotion. A novel method is proposed in this study to address some of the issues mentioned above. Firstly, modified homomorphic filtering is employed to normalize the illumination, then the normalized face image is cropped into five local regions to emphasize expression-specific characteristics. Finally, a unique blended feature attention network (BFAN) is designed for FER. BFAN consists of both residual dilated multi-scale feature extraction module and spatial and channel-wise attention modules. These modules help to extract the most relevant and discriminative features from the high-level and low-level features. Then, both feature maps are integrated and passed on to the dense layers followed by a softmax layer to compute probability scores. Finally, the Choquet fuzzy integral is applied to the computed probability scores to get the final outcome. The superiority of the proposed method is exemplified by comparing it with eighteen existing approaches on seven benchmark datasets. IEEE
Czech name
—
Czech description
—
Classification
Type
J<sub>imp</sub> - Article in a specialist periodical, which is included in the Web of Science database
CEP classification
—
OECD FORD branch
10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)
Result continuities
Project
—
Continuities
I - Institucionalni podpora na dlouhodoby koncepcni rozvoj vyzkumne organizace
Others
Publication year
2023
Confidentiality
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Data specific for result type
Name of the periodical
IEEE Transactions on Instrumentation and Measurement
ISSN
0018-9456
e-ISSN
1557-9662
Volume of the periodical
72
Issue of the periodical within the volume
September
Country of publishing house
US - UNITED STATES
Number of pages
16
Pages from-to
"Article number: 5026416"
UT code for WoS article
001083291000016
EID of the result in the Scopus database
2-s2.0-85171576697