Publication: BAUM-1: A Spontaneous Audio-Visual Face Database of Affective and Mental States
| dc.contributor.author | Zhalehpour, Sara | |
| dc.contributor.author | Önder, Onur | |
| dc.contributor.author | Akhtar, Zahid | |
| dc.contributor.author | Erdem, Cigdem Eroglu | |
| dc.contributor.institution | Zhalehpour, Sara, Centre Énergie Matériaux Télécommunications, Varennes, Canada | |
| dc.contributor.institution | Önder, Onur, Arçelik A.S., Istanbul, Turkey | |
| dc.contributor.institution | Akhtar, Zahid, Department of Mathematics and Computer Science, Università degli Studi di Udine, Udine, Italy | |
| dc.contributor.institution | Erdem, Cigdem Eroglu, Department of Electrical and Electronic Engineering, Bahçeşehir Üniversitesi, Istanbul, Turkey | |
| dc.date.accessioned | 2025-10-05T16:16:34Z | |
| dc.date.issued | 2017 | |
| dc.description.abstract | In affective computing applications, access to labeled spontaneous affective data is essential for testing the designed algorithms under naturalistic and challenging conditions. Most databases available today are acted or do not contain audio data. We present a spontaneous audio-visual affective face database of affective and mental states. The video clips in the database are obtained by recording the subjects from the frontal view using a stereo camera and from the half-profile view using a mono camera. The subjects are first shown a sequence of images and short video clips, which are not only meticulously fashioned but also timed to evoke a set of emotions and mental states. Then, they express their ideas and feelings about the images and video clips they have watched in an unscripted and unguided way in Turkish. The target emotions, include the six basic ones (happiness, anger, sadness, disgust, fear, surprise) as well as boredom and contempt. We also target several mental states, which are unsure (including confused, undecided), thinking, concentrating, and bothered. Baseline experimental results on the BAUM-1 database show that recognition of affective and mental states under naturalistic conditions is quite challenging. The database is expected to enable further research on audio-visual affect and mental state recognition under close-To-real scenarios. © 2017 Elsevier B.V., All rights reserved. | |
| dc.identifier.doi | 10.1109/TAFFC.2016.2553038 | |
| dc.identifier.endpage | 313 | |
| dc.identifier.issn | 19493045 | |
| dc.identifier.issue | 3 | |
| dc.identifier.scopus | 2-s2.0-85029943602 | |
| dc.identifier.startpage | 300 | |
| dc.identifier.uri | https://doi.org/10.1109/TAFFC.2016.2553038 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.14719/12032 | |
| dc.identifier.volume | 8 | |
| dc.language.iso | en | |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | |
| dc.relation.source | IEEE Transactions on Affective Computing | |
| dc.subject.authorkeywords | Affective Computing | |
| dc.subject.authorkeywords | Audio-visual Affective Database | |
| dc.subject.authorkeywords | Dynamic Facial Expression Database | |
| dc.subject.authorkeywords | Emotion Recognition From Speech | |
| dc.subject.authorkeywords | Emotional Corpora | |
| dc.subject.authorkeywords | Facial Expression Recognition | |
| dc.subject.authorkeywords | Mental State Recognition | |
| dc.subject.authorkeywords | Spontaneous Expressions | |
| dc.subject.authorkeywords | Cameras | |
| dc.subject.authorkeywords | Database Systems | |
| dc.subject.authorkeywords | Human Computer Interaction | |
| dc.subject.authorkeywords | State Estimation | |
| dc.subject.authorkeywords | Stereo Image Processing | |
| dc.subject.authorkeywords | Video Cameras | |
| dc.subject.authorkeywords | Affective Computing | |
| dc.subject.authorkeywords | Audio-visual | |
| dc.subject.authorkeywords | Dynamic Facial Expression | |
| dc.subject.authorkeywords | Emotion Recognition From Speech | |
| dc.subject.authorkeywords | Emotional Corpora | |
| dc.subject.authorkeywords | Facial Expression Recognition | |
| dc.subject.authorkeywords | Mental State | |
| dc.subject.authorkeywords | Spontaneous Expressions | |
| dc.subject.authorkeywords | Speech Recognition | |
| dc.subject.indexkeywords | Cameras | |
| dc.subject.indexkeywords | Database systems | |
| dc.subject.indexkeywords | Human computer interaction | |
| dc.subject.indexkeywords | State estimation | |
| dc.subject.indexkeywords | Stereo image processing | |
| dc.subject.indexkeywords | Video cameras | |
| dc.subject.indexkeywords | Affective Computing | |
| dc.subject.indexkeywords | Audio-visual | |
| dc.subject.indexkeywords | Dynamic facial expression | |
| dc.subject.indexkeywords | Emotion recognition from speech | |
| dc.subject.indexkeywords | Emotional corpora | |
| dc.subject.indexkeywords | Facial expression recognition | |
| dc.subject.indexkeywords | Mental state | |
| dc.subject.indexkeywords | spontaneous expressions | |
| dc.subject.indexkeywords | Speech recognition | |
| dc.title | BAUM-1: A Spontaneous Audio-Visual Face Database of Affective and Mental States | |
| dc.type | Article | |
| dcterms.references | Sebe, Niculae, Multimodal approaches for emotion recognition: A survey, Proceedings of SPIE - The International Society for Optical Engineering, 5670, pp. 56-67, (2005), Zeng, Zhihong, A survey of affect recognition methods: Audio, visual, and spontaneous expressions, IEEE Transactions on Pattern Analysis and Machine Intelligence, 31, 1, pp. 39-58, (2009), Ryan, Andrew, Automated facial expression recognition system, Proceedings - International Carnahan Conference on Security Technology, pp. 172-177, (2009), Littlewort, Gwen C., Automatic coding of facial expressions displayed during posed and genuine pain, Image and Vision Computing, 27, 12, pp. 1797-1803, (2009), Ashraf, Ahmed Bilal, The painful face - Pain expression recognition using active appearance models, Image and Vision Computing, 27, 12, pp. 1788-1796, (2009), Palo Alto CA Consulting Psychologists Press, (1976), Bassili, John N., Emotion recognition: The role of facial movement and the relative importance of upper and lower areas of the face, Journal of Personality and Social Psychology, 37, 11, pp. 2049-2058, (1979), Lucey, Patrick, The extended Cohn-Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression, pp. 94-101, (2010), Blueprint for Affective Computing A Sourcebook, (2010), Savran, Arman, Bosphorus database for 3D face analysis, Lecture Notes in Computer Science, 5372 LNCS, pp. 47-56, (2008) | |
| dspace.entity.type | Publication | |
| local.indexed.at | Scopus | |
| person.identifier.scopus-author-id | 58343878100 | |
| person.identifier.scopus-author-id | 24765325900 | |
| person.identifier.scopus-author-id | 46661628200 | |
| person.identifier.scopus-author-id | 55807016900 |
