Publication: Context-Aware Facial Expression Recognition in Group Photos
| dc.contributor.author | Çakir, Duygu | |
| dc.contributor.author | Arica, Nafiz | |
| dc.contributor.institution | Çakir, Duygu, Department of Software Engineering, Bahçeşehir Üniversitesi, Istanbul, Turkey | |
| dc.contributor.institution | Arica, Nafiz, Department of Management Information Systems, Pîrî Reis Üniversitesi, Istanbul, Turkey | |
| dc.date.accessioned | 2025-10-05T15:06:56Z | |
| dc.date.issued | 2023 | |
| dc.description.abstract | Facial expression recognition has made significant progress in controlled laboratory environments, thanks to advancements in computing power, method development, and data availability. However, recognizing expressions in real-world (in-the-wild) settings remains a challenge, particularly for expressions that are often confused with others. While misclassifications in single-faced images primarily depend on the model's performance, the dominant expression in images containing multiple faces can help correct these misclassified expressions. By employing a two-stage classification algorithm to identify the most common expression in the image and reclassify faces with different expressions accordingly, we observed a 4.29% increase in the success rate of expression recognition among 21,270 face images. This approach leverages collective information within group images to achieve comprehensive facial expression recognition, overcoming limitations in analyzing single static images. Such advancements are crucial for real-world scenarios where facial expression recognition faces challenges due to environmental factors, the presence of multiple faces, and low-resolution or degraded image quality. © 2023 Elsevier B.V., All rights reserved. | |
| dc.description.sponsorship | et al. | |
| dc.description.sponsorship | GENOPOLE | |
| dc.description.sponsorship | IEEE France Section | |
| dc.description.sponsorship | IEEE Signal Processing Society | |
| dc.description.sponsorship | IUT of Evry Val d�Essonne | |
| dc.description.sponsorship | Paris-Saclay University, Univ. Evry | |
| dc.identifier.conferenceName | 12th International Conference on Image Processing Theory, Tools and Applications, IPTA 2023 | |
| dc.identifier.conferencePlace | Paris | |
| dc.identifier.doi | 10.1109/IPTA59101.2023.10320068 | |
| dc.identifier.isbn | 9798350325416 | |
| dc.identifier.scopus | 2-s2.0-85179556359 | |
| dc.identifier.uri | https://doi.org/10.1109/IPTA59101.2023.10320068 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.14719/8189 | |
| dc.language.iso | en | |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | |
| dc.subject.authorkeywords | Deep Learning | |
| dc.subject.authorkeywords | Dominant Expression | |
| dc.subject.authorkeywords | Facial Expression Recognition | |
| dc.subject.authorkeywords | Computing Power | |
| dc.subject.authorkeywords | Deep Learning | |
| dc.subject.authorkeywords | Context-aware | |
| dc.subject.authorkeywords | Controlled Laboratories | |
| dc.subject.authorkeywords | Data Availability | |
| dc.subject.authorkeywords | Dominant Expression | |
| dc.subject.authorkeywords | Facial Expression Recognition | |
| dc.subject.authorkeywords | Laboratory Environment | |
| dc.subject.authorkeywords | Method Development | |
| dc.subject.authorkeywords | Powers Method | |
| dc.subject.authorkeywords | Face Recognition | |
| dc.subject.indexkeywords | Computing power | |
| dc.subject.indexkeywords | Deep learning | |
| dc.subject.indexkeywords | Context-Aware | |
| dc.subject.indexkeywords | Controlled laboratories | |
| dc.subject.indexkeywords | Data availability | |
| dc.subject.indexkeywords | Dominant expression | |
| dc.subject.indexkeywords | Facial expression recognition | |
| dc.subject.indexkeywords | Laboratory environment | |
| dc.subject.indexkeywords | Method development | |
| dc.subject.indexkeywords | Powers method | |
| dc.subject.indexkeywords | Face recognition | |
| dc.title | Context-Aware Facial Expression Recognition in Group Photos | |
| dc.type | Conference Paper | |
| dcterms.references | Ekman, Paul, Measuring facial movement, Environmental Psychology and Nonverbal Behavior, 1, 1, pp. 56-75, (1976), Journal of the Canadian Academy of Child and Adolescent Psychiatry, (2007), Wang, Fengyuan, Facial expression recognition from image based on hybrid features understanding, Journal of Visual Communication and Image Representation, 59, pp. 84-88, (2019), Zhao, Xiangyun, Peak-piloted deep network for facial expression recognition, Lecture Notes in Computer Science, 9906 LNCS, pp. 425-442, (2016), Shan, Caifeng, Facial expression recognition based on Local Binary Patterns: A comprehensive study, Image and Vision Computing, 27, 6, pp. 803-816, (2009), Huang, Mingwei, A new method for facial expression recognition based on sparse representation plus LBP, 4, pp. 1750-1754, (2010), Lyons, Michael J., Coding facial expressions with Gabor wavelets, pp. 200-205, (1998), Hu, Yuxiao, Multi-view facial expression recognition, (2008), Gritti, Tommaso, Local features based facial expression recognition with face registration errors, (2008), Zhao, Kaili, Joint patch and multi-label learning for facial action unit detection, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 07-12-June-2015, pp. 2207-2216, (2015) | |
| dspace.entity.type | Publication | |
| local.indexed.at | Scopus | |
| person.identifier.scopus-author-id | 42060965200 | |
| person.identifier.scopus-author-id | 56247026400 |
