Towards a Better Performance in Facial Expression Recognition: A Data-Centric Approach
Please use this identifier to cite or link to this item:
http://hdl.handle.net/10045/138462
Title: | Towards a Better Performance in Facial Expression Recognition: A Data-Centric Approach |
---|---|
Authors: | Mejia-Escobar, Christian | Cazorla, Miguel | Martinez-Martin, Ester |
Research Group/s: | Robótica y Visión Tridimensional (RoViT) |
Center, Department or Service: | Universidad de Alicante. Departamento de Ciencia de la Computación e Inteligencia Artificial |
Keywords: | Facial expression recognition | Deep learning | Convolutional Neural Networks | Data-centric method |
Issue Date: | 3-Nov-2023 |
Publisher: | Hindawi |
Citation: | Computational Intelligence and Neuroscience. 2023, Article ID 1394882. https://doi.org/10.1155/2023/1394882 |
Abstract: | Facial expression is the best evidence of our emotions. Its automatic detection and recognition are key for robotics, medicine, healthcare, education, psychology, sociology, marketing, security, entertainment, and many other areas. Experiments in the lab environments achieve high performance. However, in real-world scenarios, it is challenging. Deep learning techniques based on convolutional neural networks (CNNs) have shown great potential. Most of the research is exclusively model-centric, searching for better algorithms to improve recognition. However, progress is insufficient. Despite being the main resource for automatic learning, few works focus on improving the quality of datasets. We propose a novel data-centric method to tackle misclassification, a problem commonly encountered in facial image datasets. The strategy is to progressively refine the dataset by successive training of a CNN model that is fixed. Each training uses the facial images corresponding to the correct predictions of the previous training, allowing the model to capture more distinctive features of each class of facial expression. After the last training, the model performs automatic reclassification of the whole dataset. Unlike other similar work, our method avoids modifying, deleting, or augmenting facial images. Experimental results on three representative datasets proved the effectiveness of the proposed method, improving the validation accuracy by 20.45%, 14.47%, and 39.66%, for FER2013, NHFI, and AffectNet, respectively. The recognition rates on the reclassified versions of these datasets are 86.71%, 70.44%, and 89.17% and become state-of-the-art performance. |
Sponsor: | This work was funded by grant CIPROM/2021/17 awarded by the Prometeo program from Conselleria de Innovación, Universidades, Ciencia y Sociedad Digital of Generalitat Valenciana (Spain), and partially funded by the grant awarded by the Central University of Ecuador through budget certification no. 34 of March 25, 2022, for the development of the research project with code: DOCT-DI-2020-37. |
URI: | http://hdl.handle.net/10045/138462 |
ISSN: | 1687-5265 (Print) | 1687-5273 (Online) |
DOI: | 10.1155/2023/1394882 |
Language: | eng |
Type: | info:eu-repo/semantics/article |
Rights: | © 2023 Christian Mejia-Escobar et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
Peer Review: | si |
Publisher version: | https://doi.org/10.1155/2023/1394882 |
Appears in Collections: | INV - RoViT - Artículos de Revistas |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Mejia-Escobar_etal_2023_ComputIntelligNeurosci.pdf | 3,65 MB | Adobe PDF | Open Preview | |
Items in RUA are protected by copyright, with all rights reserved, unless otherwise indicated.