The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)

Based on a comprehensive literature review, this study highlights the critical role of the temporal dimension of facial dynamics in understanding facial expressions and improving the accuracy and robustness of automatic emotion recognition systems (machine-FER). While deep learning (DL) techniques l...

Full description

Saved in:
Bibliographic Details
Main Authors: Williams Contreras-Higuera, Lucrezia Crescenzi-Lanna
Format: Article
Language:English
Published: Wiley 2025-01-01
Series:Human Behavior and Emerging Technologies
Online Access:http://dx.doi.org/10.1155/hbe2/7777949
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850107936355909632
author Williams Contreras-Higuera
Lucrezia Crescenzi-Lanna
author_facet Williams Contreras-Higuera
Lucrezia Crescenzi-Lanna
author_sort Williams Contreras-Higuera
collection DOAJ
description Based on a comprehensive literature review, this study highlights the critical role of the temporal dimension of facial dynamics in understanding facial expressions and improving the accuracy and robustness of automatic emotion recognition systems (machine-FER). While deep learning (DL) techniques like convolutional neural networks (CNNs) and long short-term memory (LSTM) networks offer significant advances, they face challenges such as gradient vanishing and overfitting, particularly in long and complex sequences. Vision transformers (ViTs) show promise but require integration with algorithms to mitigate spatial noise. Conventional machine learning (CML) methods like support vector machine (SVM) remain robust, especially in smaller datasets. The study underscores the importance of multimodal data synchronization (e.g., video, voice) in classifying emotions more accurately, reflecting both human and machine learning capabilities. It also addresses the limitations of current models, including cultural biases and the need for large, diverse datasets. The findings suggest that future research should focus on real-world conditions, integrating sequential multimodal data and employing supervised models based on theoretical and empirical frameworks. This approach is aimed at enhancing the understanding and classification of facial emotions, ensuring data quality and acceptable results through systematic human observations. The study provides valuable insights for selecting appropriate algorithms that are tailored to specific research objectives and contexts.
format Article
id doaj-art-6d9988278e0b4aa3aa2ae719dfa54dc2
institution OA Journals
issn 2578-1863
language English
publishDate 2025-01-01
publisher Wiley
record_format Article
series Human Behavior and Emerging Technologies
spelling doaj-art-6d9988278e0b4aa3aa2ae719dfa54dc22025-08-20T02:38:29ZengWileyHuman Behavior and Emerging Technologies2578-18632025-01-01202510.1155/hbe2/7777949The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)Williams Contreras-Higuera0Lucrezia Crescenzi-Lanna1Networks and Information TechnologiesFaculty of Psychology and Education SciencesBased on a comprehensive literature review, this study highlights the critical role of the temporal dimension of facial dynamics in understanding facial expressions and improving the accuracy and robustness of automatic emotion recognition systems (machine-FER). While deep learning (DL) techniques like convolutional neural networks (CNNs) and long short-term memory (LSTM) networks offer significant advances, they face challenges such as gradient vanishing and overfitting, particularly in long and complex sequences. Vision transformers (ViTs) show promise but require integration with algorithms to mitigate spatial noise. Conventional machine learning (CML) methods like support vector machine (SVM) remain robust, especially in smaller datasets. The study underscores the importance of multimodal data synchronization (e.g., video, voice) in classifying emotions more accurately, reflecting both human and machine learning capabilities. It also addresses the limitations of current models, including cultural biases and the need for large, diverse datasets. The findings suggest that future research should focus on real-world conditions, integrating sequential multimodal data and employing supervised models based on theoretical and empirical frameworks. This approach is aimed at enhancing the understanding and classification of facial emotions, ensuring data quality and acceptable results through systematic human observations. The study provides valuable insights for selecting appropriate algorithms that are tailored to specific research objectives and contexts.http://dx.doi.org/10.1155/hbe2/7777949
spellingShingle Williams Contreras-Higuera
Lucrezia Crescenzi-Lanna
The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)
Human Behavior and Emerging Technologies
title The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)
title_full The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)
title_fullStr The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)
title_full_unstemmed The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)
title_short The Role of Time in Facial Dynamics and Challenges in Automatic Emotion Recognition (2019–2024)
title_sort role of time in facial dynamics and challenges in automatic emotion recognition 2019 2024
url http://dx.doi.org/10.1155/hbe2/7777949
work_keys_str_mv AT williamscontrerashiguera theroleoftimeinfacialdynamicsandchallengesinautomaticemotionrecognition20192024
AT lucreziacrescenzilanna theroleoftimeinfacialdynamicsandchallengesinautomaticemotionrecognition20192024
AT williamscontrerashiguera roleoftimeinfacialdynamicsandchallengesinautomaticemotionrecognition20192024
AT lucreziacrescenzilanna roleoftimeinfacialdynamicsandchallengesinautomaticemotionrecognition20192024