Towards a Better Performance in Facial Expression Recognition: A Data-Centric Approach

Por favor, use este identificador para citar o enlazar este ítem: http://hdl.handle.net/10045/138462
Información del item - Informació de l'item - Item information
Título: Towards a Better Performance in Facial Expression Recognition: A Data-Centric Approach
Autor/es: Mejia-Escobar, Christian | Cazorla, Miguel | Martinez-Martin, Ester
Grupo/s de investigación o GITE: Robótica y Visión Tridimensional (RoViT)
Centro, Departamento o Servicio: Universidad de Alicante. Departamento de Ciencia de la Computación e Inteligencia Artificial
Palabras clave: Facial expression recognition | Deep learning | Convolutional Neural Networks | Data-centric method
Fecha de publicación: 3-nov-2023
Editor: Hindawi
Cita bibliográfica: Computational Intelligence and Neuroscience. 2023, Article ID 1394882. https://doi.org/10.1155/2023/1394882
Resumen: Facial expression is the best evidence of our emotions. Its automatic detection and recognition are key for robotics, medicine, healthcare, education, psychology, sociology, marketing, security, entertainment, and many other areas. Experiments in the lab environments achieve high performance. However, in real-world scenarios, it is challenging. Deep learning techniques based on convolutional neural networks (CNNs) have shown great potential. Most of the research is exclusively model-centric, searching for better algorithms to improve recognition. However, progress is insufficient. Despite being the main resource for automatic learning, few works focus on improving the quality of datasets. We propose a novel data-centric method to tackle misclassification, a problem commonly encountered in facial image datasets. The strategy is to progressively refine the dataset by successive training of a CNN model that is fixed. Each training uses the facial images corresponding to the correct predictions of the previous training, allowing the model to capture more distinctive features of each class of facial expression. After the last training, the model performs automatic reclassification of the whole dataset. Unlike other similar work, our method avoids modifying, deleting, or augmenting facial images. Experimental results on three representative datasets proved the effectiveness of the proposed method, improving the validation accuracy by 20.45%, 14.47%, and 39.66%, for FER2013, NHFI, and AffectNet, respectively. The recognition rates on the reclassified versions of these datasets are 86.71%, 70.44%, and 89.17% and become state-of-the-art performance.
Patrocinador/es: This work was funded by grant CIPROM/2021/17 awarded by the Prometeo program from Conselleria de Innovación, Universidades, Ciencia y Sociedad Digital of Generalitat Valenciana (Spain), and partially funded by the grant awarded by the Central University of Ecuador through budget certification no. 34 of March 25, 2022, for the development of the research project with code: DOCT-DI-2020-37.
URI: http://hdl.handle.net/10045/138462
ISSN: 1687-5265 (Print) | 1687-5273 (Online)
DOI: 10.1155/2023/1394882
Idioma: eng
Tipo: info:eu-repo/semantics/article
Derechos: © 2023 Christian Mejia-Escobar et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Revisión científica: si
Versión del editor: https://doi.org/10.1155/2023/1394882
Aparece en las colecciones:INV - RoViT - Artículos de Revistas

Archivos en este ítem:
Archivos en este ítem:
Archivo Descripción TamañoFormato 
ThumbnailMejia-Escobar_etal_2023_ComputIntelligNeurosci.pdf3,65 MBAdobe PDFAbrir Vista previa


Todos los documentos en RUA están protegidos por derechos de autor. Algunos derechos reservados.