Towards a Better Performance in Facial Expression Recognition: A Data-Centric Approach

Please use this identifier to cite or link to this item: http://hdl.handle.net/10045/138462
Información del item - Informació de l'item - Item information
Title: Towards a Better Performance in Facial Expression Recognition: A Data-Centric Approach
Authors: Mejia-Escobar, Christian | Cazorla, Miguel | Martinez-Martin, Ester
Research Group/s: Robótica y Visión Tridimensional (RoViT)
Center, Department or Service: Universidad de Alicante. Departamento de Ciencia de la Computación e Inteligencia Artificial
Keywords: Facial expression recognition | Deep learning | Convolutional Neural Networks | Data-centric method
Issue Date: 3-Nov-2023
Publisher: Hindawi
Citation: Computational Intelligence and Neuroscience. 2023, Article ID 1394882. https://doi.org/10.1155/2023/1394882
Abstract: Facial expression is the best evidence of our emotions. Its automatic detection and recognition are key for robotics, medicine, healthcare, education, psychology, sociology, marketing, security, entertainment, and many other areas. Experiments in the lab environments achieve high performance. However, in real-world scenarios, it is challenging. Deep learning techniques based on convolutional neural networks (CNNs) have shown great potential. Most of the research is exclusively model-centric, searching for better algorithms to improve recognition. However, progress is insufficient. Despite being the main resource for automatic learning, few works focus on improving the quality of datasets. We propose a novel data-centric method to tackle misclassification, a problem commonly encountered in facial image datasets. The strategy is to progressively refine the dataset by successive training of a CNN model that is fixed. Each training uses the facial images corresponding to the correct predictions of the previous training, allowing the model to capture more distinctive features of each class of facial expression. After the last training, the model performs automatic reclassification of the whole dataset. Unlike other similar work, our method avoids modifying, deleting, or augmenting facial images. Experimental results on three representative datasets proved the effectiveness of the proposed method, improving the validation accuracy by 20.45%, 14.47%, and 39.66%, for FER2013, NHFI, and AffectNet, respectively. The recognition rates on the reclassified versions of these datasets are 86.71%, 70.44%, and 89.17% and become state-of-the-art performance.
Sponsor: This work was funded by grant CIPROM/2021/17 awarded by the Prometeo program from Conselleria de Innovación, Universidades, Ciencia y Sociedad Digital of Generalitat Valenciana (Spain), and partially funded by the grant awarded by the Central University of Ecuador through budget certification no. 34 of March 25, 2022, for the development of the research project with code: DOCT-DI-2020-37.
URI: http://hdl.handle.net/10045/138462
ISSN: 1687-5265 (Print) | 1687-5273 (Online)
DOI: 10.1155/2023/1394882
Language: eng
Type: info:eu-repo/semantics/article
Rights: © 2023 Christian Mejia-Escobar et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Peer Review: si
Publisher version: https://doi.org/10.1155/2023/1394882
Appears in Collections:INV - RoViT - Artículos de Revistas

Files in This Item:
Files in This Item:
File Description SizeFormat 
ThumbnailMejia-Escobar_etal_2023_ComputIntelligNeurosci.pdf3,65 MBAdobe PDFOpen Preview


Items in RUA are protected by copyright, with all rights reserved, unless otherwise indicated.