Improving Facial Expression Recognition Through Data Preparation and Merging

Por favor, use este identificador para citar o enlazar este ítem: http://hdl.handle.net/10045/136107
Registro completo de metadatos
Registro completo de metadatos
Campo DCValorIdioma
dc.contributorRobótica y Visión Tridimensional (RoViT)es_ES
dc.contributor.authorMejia-Escobar, Christian-
dc.contributor.authorCazorla, Miguel-
dc.contributor.authorMartinez-Martin, Ester-
dc.contributor.otherUniversidad de Alicante. Departamento de Ciencia de la Computación e Inteligencia Artificiales_ES
dc.date.accessioned2023-07-13T09:44:15Z-
dc.date.available2023-07-13T09:44:15Z-
dc.date.issued2023-07-10-
dc.identifier.citationIEEE Access. 2023, 11: 71339-71360. https://doi.org/10.1109/ACCESS.2023.3293728es_ES
dc.identifier.issn2169-3536-
dc.identifier.urihttp://hdl.handle.net/10045/136107-
dc.description.abstractHuman emotions present a major challenge for artificial intelligence. Automated emotion recognition based on facial expressions is important to robotics, medicine, psychology, education, security, arts, entertainment and more. Deep learning is promising for capturing complex emotional features. However, there is no training dataset that is large and representative of the full diversity of emotional expressions in all populations and contexts. Current facial datasets are incomplete, biased, unbalanced, error-prone and have different properties. Models learn these limitations and become dependent on specific datasets, hindering their ability to generalize to new data or real-world scenarios. Our work addresses these difficulties and provides the following contributions to improve emotion recognition: 1) a methodology for merging disparate in-the-wild datasets that increases the number of images and enriches the diversity of people, gestures, and attributes of resolution, color, background, lighting and image format; 2) a balanced, unbiased, and well-labeled evaluator dataset, built with a gender, age, and ethnicity predictor and the successful Stable Diffusion model. Single- and cross-dataset experimentation show that our method increases the generalization of the FER2013, NHFI and AffectNet datasets by 13.93%, 24.17% and 7.45%, respectively; and 3) we propose the first and largest artificial emotion dataset, which can complement real datasets in tasks related to facial expression.es_ES
dc.description.sponsorshipThis work has been funded by grant CIPROM/2021/017 awarded by the MEEBAI Project (Prometheus Programme for Research Groups on R&D Excellence) from Conselleria de Innovación, Universidades, Ciencia y Sociedad Digital of Generalitat Valenciana (Spain), and partially by the grant awarded by the Central University of Ecuador through budget certification No. 34 of March 25, 2022 for the development of the research project with code DOCT-DI-2020-37.es_ES
dc.languageenges_ES
dc.publisherIEEEes_ES
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4.0/es_ES
dc.subjectArtificial datasetes_ES
dc.subjectDeep Learninges_ES
dc.subjectConvolutional Neural Networkes_ES
dc.subjectEmotion Recognitiones_ES
dc.subjectFacial Expression Recognitiones_ES
dc.subjectStable Diffusiones_ES
dc.titleImproving Facial Expression Recognition Through Data Preparation and Merginges_ES
dc.typeinfo:eu-repo/semantics/articlees_ES
dc.peerreviewedsies_ES
dc.identifier.doi10.1109/ACCESS.2023.3293728-
dc.relation.publisherversionhttps://doi.org/10.1109/ACCESS.2023.3293728es_ES
dc.rights.accessRightsinfo:eu-repo/semantics/openAccesses_ES
Aparece en las colecciones:INV - RoViT - Artículos de Revistas

Archivos en este ítem:
Archivos en este ítem:
Archivo Descripción TamañoFormato 
ThumbnailMejia-Escobar_etal_2023_IEEE-Access.pdf3,3 MBAdobe PDFAbrir Vista previa


Este ítem está licenciado bajo Licencia Creative Commons Creative Commons