A Comparative Analysis of CNNs and ResNet50 for Facial Emotion Recognition
Received: 6 December 2024 | Revised: 24 December 2024 | Accepted: 4 January 2025 | Online: 21 January 2025
Corresponding author: Milind Talele
Abstract
Anger, disgust, fear, happiness, sadness, surprise, and neutrality are some of the basic facial emotions that researchers worldwide consider for recognition. Detection of these emotions is important in the present era due to the digital transformation of many processes and human communication. This study analyzes emotion detection methods using the capabilities of deep learning techniques such as a Convolutional Neural Network (CNN) and Residual Network 50 (ResNet-50). The FER2013 benchmark dataset for facial emotion recognition was used for training and testing purposes, along with a few other private images. This study aimed to compare and analyze the performance of the two methods based on several comparative factors, such as architectural differences, feature extraction capability, training dynamics, model performance, computational efficiency, and hardware configuration. The experimental results showed that the ResNet-50 model was significantly more accurate than the CNN, with an accuracy of 85.75% compared to 74%. Although ResNet-50 has higher computational costs, its robustness and accuracy make it the optimal choice for facial emotion recognition tasks. This research provides valuable insight into the capabilities and trade-offs of these models for face emotion recognition techniques.
Keywords:
artificial intelligence, basic face emotions, convolution neural network, face emotion detection, deep learningDownloads
References
P. Ekman, "Facial expression and emotion," American Psychologist, vol. 48, no. 4, pp. 384–392, 1993.
D. Grandjean, D. Sander, and K. R. Scherer, "Conscious emotional experience emerges as a function of multilevel, appraisal-driven response synchronization," Consciousness and Cognition, vol. 17, no. 2, pp. 484–495, Jun. 2008.
R. Plutchik, "A psycho evolutionary theory of emotions," Social Science Information, 1982.
H. Zhang and M. Xu, "Modeling temporal information using discrete fourier transform for recognizing emotions in user-generated videos," in 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, Sep. 2016, pp. 629–633.
D. Cheng, Y. Fan, S. Fang, M. Wang, and H. Liu, "ResNet-AE for Radar Signal Anomaly Detection," Sensors, vol. 22, no. 16, Jan. 2022, Art. no. 6249.
J. Miao, S. Xu, B. Zou, and Y. Qiao, "ResNet based on feature-inspired gating strategy," Multimedia Tools and Applications, vol. 81, no. 14, pp. 19283–19300, Jun. 2022.
J. Deng, W. Dong, R. Socher, L.-J. Li, Kai Li, and Li Fei-Fei, "ImageNet: A large-scale hierarchical image database," in 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, Jun. 2009.
M. Oquab, L. Bottou, I. Laptev, and J. Sivic, "Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks," in 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, Jun. 2014, pp. 1717–1724.
J. Wu, Yinan Yu, Chang Huang, and Kai Yu, "Deep multiple instance learning for image classification and auto-annotation," in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, Jun. 2015, pp. 3460–3469.
K. He, X. Zhang, S. Ren, and J. Sun, "Deep Residual Learning for Image Recognition," in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, Jun. 2016.
J. Oheix, "Face expression recognition dataset." Kaggle, [Online]. Available: https://www.kaggle.com/datasets/jonathanoheix/face-expression-recognition-dataset.
M. F. Aza, N. Suciati, and S. C. Hidayati, "Performance Study of Facial Expression Recognition Using Convolutional Neural Network," in 2020 6th International Conference on Science in Information Technology (ICSITech), Palu, Indonesia, Oct. 2020, pp. 121–126.
Z. Y. Huang et al., "A study on computer vision for facial emotion recognition," Scientific Reports, vol. 13, no. 1, May 2023.
D. O. Melinte and L. Vladareanu, "Facial Expressions Recognition for Human–Robot Interaction Using Deep Convolutional Neural Networks with Rectified Adam Optimizer," Sensors, vol. 20, no. 8, Apr. 2020, Art. no. 2393.
K. Mistry, L. Zhang, S. C. Neoh, C. P. Lim, and B. Fielding, "A Micro-GA Embedded PSO Feature Selection Approach to Intelligent Facial Emotion Recognition," IEEE Transactions on Cybernetics, vol. 47, no. 6, pp. 1496–1509, Jun. 2017.
Y. Khaireddin and Z. Chen, "Facial Emotion Recognition: State of the Art Performance on FER2013." arXiv, May 08, 2021.
M. Dirik, "Optimized Anfis Model with Hybrid Metaheuristic Algorithms for Facial Emotion Recognition," International Journal of Fuzzy Systems, vol. 25, no. 2, pp. 485–496, Mar. 2023.
A. Alshammari and M. E. Alshammari, "Emotional Facial Expression Detection using YOLOv8," Engineering, Technology & Applied Science Research, vol. 14, no. 5, pp. 16619–16623, Oct. 2024.
M. Talele, R. Jain, and S. Mapari, "Complex Face Emotion Recognition Using Computer Vision and Machine Learning," in Harnessing Artificial Emotional Intelligence for Improved Human-Computer Interactions, IGI Global, 2024, pp. 180–196.
M. Tripathi, "Facial Emotion Recognition Using Convolutional Neural Network," ICTACT Journal on Image and Video Processing, vol. 12, no. 1, pp. 2531–2536, Aug. 2021.
S. Porcu, A. Floris, and L. Atzori, "Evaluation of Data Augmentation Techniques for Facial Expression Recognition Systems," Electronics, vol. 9, no. 11, Nov. 2020, Art. no. 1892.
M. a. H. Akhand, S. Roy, N. Siddique, M. A. S. Kamal, and T. Shimamura, "Facial Emotion Recognition Using Transfer Learning in the Deep CNN," Electronics, vol. 10, no. 9, Jan. 2021, Art. no. 1036.
C. Szegedy et al., "Going deeper with convolutions," in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1–9.
P. D. M. Fernandez, F. A. G. Pena, T. I. Ren, and A. Cunha, "FERAtt: Facial Expression Recognition With Attention Net," in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, Jun. 2019, pp. 837–846.
L. V. L. Pinto et al., "A Systematic Review of Facial Expression Detection Methods," IEEE Access, vol. 11, pp. 61881–61891, 2023.
S. W. Naidoo, N. Naicker, S. S. Patel, and P. Govender, "Computer Vision: The Effectiveness of Deep Learning for Emotion Detection in Marketing Campaigns," International Journal of Advanced Computer Science and Applications, vol. 13, no. 5, 2022.
A. Kaehler and G. Bradski, Learning OpenCV 3: Computer Vision in C++ with the OpenCV Library. O’Reilly Media, Inc., 2016.
Madhivarman, "How to calculate the number of parameters in the CNN?," Medium, May 30, 2018. https://medium.com/@iamvarman/
how-to-calculate-the-number-of-parameters-in-the-cnn-5bd55364d7ca.
H. A. Rowley, S. Baluja, and T. Kanade, "Neural network-based face detection," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20, no. 1, pp. 23–38, Jan. 1998.
B. Li and D. Lima, "Facial expression recognition via ResNet-50," International Journal of Cognitive Computing in Engineering, vol. 2, pp. 57–64, Jun. 2021.
Downloads
How to Cite
License
Copyright (c) 2025 Milind Talele, Rajashree Jain

This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain the copyright and grant the journal the right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) after its publication in ETASR with an acknowledgement of its initial publication in this journal.