Your search

In authors or contributors
  • Facial expression recognition is a key topic in computer vision, playing a crucial role in non-verbal communication. With the rapid development of artificial intelligence, significant progress has been made in improving recognition accuracy and generalization abilities. Traditional methods often suffer from low precision and poor generalization, while deep learning models have substantially advanced the field. However, deploying deep and complex models on different platforms remains challenging due to their high computational demands and frameworks. Hence, developing an efficient, real-time, and lightweight facial expression recognition system is critical. This study focuses on creating an efficient, accurate, and lightweight real-time facial expression recognition system with an emphasis on cross-platform deployment. It integrates various deep learning and optimization techniques to demonstrate flexibility across platforms. In this context, this study evaluates the performance of 10 advanced CNN models (VGG16, VGG19, ResNet50, etc.) on the facial expression dataset FER2013. YOLOv8 combined with ResNet50 achieved 70.56% accuracy on FER2013, outperforming YOLOv8 alone by 2.1%. Multi-module fusion models (MobileFaceNet, IR50, HyViT, SE) achieved an accuracy of 92.58% and 74.8% on the RAF-DB and FER2013 datasets, respectively, showing superior performance in ablation experiments. Given the significant impact of data quality on model performance, this study performed data cleaning on the FER2013 dataset, resulting in a 3.25% accuracy improvement for the YOLOv8 + ResNet50 model, which reached 73.81%. The high-resolution RAF-DB dataset, with fewer errors, led to improved performance, achieving 92.56% accuracy with the fusion model. A multi-purpose facial expression recognition system, VISTA, was developed using Python and PyQt5. The system supports multiple data formats and provides real-time emotional feedback, thus enhancing its usability for both research and practical applications. Furthermore, the fusion model was quantized using the OpenVINO toolkit, reducing its parameters by 75% while maintaining an accuracy of 91.17%. Inference speed was improved, and XGrad-CAM was employed to enhance model interpretability, revealing that the YOLOv8 + ResNet50 combination more effectively captured facial features. Finally, the high-performance model was successfully deployed on Intel CPUs, NVIDIA GPUs, and embedded devices Raspberry Pi 4B, demonstrating the portability and flexibility of the VISTA system across various platforms. This research provides promising solutions for applications in human-computer interaction, affective computing, and real-time emotional analysis, with significant advancements made in improving system real-time performance, accuracy, and cross-platform deployment capabilities. It contributes to the development of facial expression recognition technology and lays the foundation for its widespread future applications in fields such as smart healthcare, business analytics, education, and mental health.

  • The classification of emotions based on facial expressions have been a new topic of research in recent years, especially in marketing and consumer behavior areas. However, there is lack of studies to understand how the research topic is developed in terms of bibliometric data. Therefore, the purpose of this work is to provide a bibliometric analysis of the research on the analysis of facial expressions for marketing and consumer behavior, identifying the state of the art, the latest research direction, and other indicators. We extracted data from Web of Science (WOS) platform, considering its core database, resulting in a total of 117 articles. The software Vosviewer was used to analyze the data and graphically visualize the results. This study indicates some of the most influential authors citations and coupling analysis in this specific field, identifies journals with the most published articles, and provide trends of the research area based on the analysis of keywords and corresponding number of articles per year. The results shows that 11 articles (9.4%) were cited more than 100 times, and the two most prolific authors published 5 articles, and the two most influential authors are Bouaziz Sofien and Pauly mark(270 citations) in this field. Of the 117 articles retrieved by WOS, more than 70% were published in high impact journals. The bibliometric analysis of the existing work in this study provides a valuable and reliable reference for researchers in this field and makes a reasonable prediction of the research direction trends.

  • Facial expression recognition (FER) is essential for discerning human emotions and is applied extensively in big data analytics, healthcare, security, and user experience enhancement. This paper presents an empirical study that evaluates four existing deep learning models—VGG16, DenseNet, ResNet50, and GoogLeNet—utilizing the Facial Expression Recognition 2013 (FER2013) dataset. The dataset contains seven distinct emotional expressions: angry, disgust, fear, happy, neutral, sad, and surprise. Each model underwent rigorous assessment based on metrics including test accuracy, training duration, and weight file size to test their effectiveness in FER tasks. ResNet50 emerged as the top performer with a test accuracy of 69.46%, leveraging its residual learning architecture to effectively address challenges inherent in training deep neural networks. Conversely, GoogLeNet exhibited the lowest test accuracy among the models, suggesting potential architectural constraints in FER applications. VGG16, while competitive in accuracy, demonstrated lengthier training times and a larger weight file size (512MB), highlighting the inherent balance between model complexity and computational efficiency.

  • <jats:p>Facial expression recognition (FER) is essential for discerning human emotions and is applied extensively in big data analytics, healthcare, security, and user experience enhancement. This study presents a comprehensive evaluation of ten state-of-the-art deep learning models—VGG16, VGG19, ResNet50, ResNet101, DenseNet, GoogLeNet V1, MobileNet V1, EfficientNet V2, ShuffleNet V2, and RepVGG—on the task of facial expression recognition using the FER2013 dataset. Key performance metrics, including test accuracy, training time, and weight file size, were analyzed to assess the learning efficiency, generalization capabilities, and architectural innovations of each model. EfficientNet V2 and ResNet50 emerged as top performers, achieving high accuracy and stable convergence using compound scaling and residual connections, enabling them to capture complex emotional features with minimal overfitting. DenseNet, GoogLeNet V1, and RepVGG also demonstrated strong performance, leveraging dense connectivity, inception modules, and re-parameterization techniques, though they exhibited slower initial convergence. In contrast, lightweight models such as MobileNet V1 and ShuffleNet V2, while excelling in computational efficiency, faced limitations in accuracy, particularly in challenging emotion categories like “fear” and “disgust”. The results highlight the critical trade-offs between computational efficiency and predictive accuracy, emphasizing the importance of selecting appropriate architecture based on application-specific requirements. This research contributes to ongoing advancements in deep learning, particularly in domains such as facial expression recognition, where capturing subtle and complex patterns is essential for high-performance outcomes.</jats:p>

Last update from database: 12/14/25, 7:01 PM (UTC)