Araştırma Çıktıları | WoS | Scopus | TR-Dizin | PubMed
Permanent URI for this communityhttps://hdl.handle.net/20.500.14719/1741
Browse
2 results
Search Results
Publication Metadata only Facial Emotion Classification Using Deep Embedding with Triplet Loss Function(IEEE, 2017) Bircanoglu, Cenk; Arica, Nafiz; Bahcesehir UniversityIn this paper, a deep embedding method using triplet loss function is proposed for classification of the emotions in face images. The originality of proposed method lies in the loss function, different from the other deep learning-based facial emotion classification approaches. The input face images are embedded into a lower dimensional feature space using a multi layer convolutional neural network. In this embedding process the loss function is calculated by taking triple samples in the training dataset. For each sample in batches, two samples are selected in such a way that one of them is from the same class and the other one from the different class. The loss function aims to close the samples belonging to the same class to each other in the Euclidean space and to move away from the samples in the different class. In the performance analysis of proposed method, two popular deep architectures, namely AlexNet and VGG are used on two datasets called CIFE and GaMo. The experiments show that the proposed method outperforms the other popular loss functions and the available studies on this particular datasets.Publication Metadata only Face Frontalization Enhanced by Deep Learning(IEEE, 2017) Celik, Anil; Arica, Nafiz; Bahcesehir UniversityIn this study, a new approach based on 3-D models and deep learning to frontalize face images is proposed. Specifically designed for facial expression analysis, the proposed approach aims to reduce possible negative effects that a posed face image can generate, by normalizing the face region. In the first phase, the face image is semi-frontalized, with a pre-established 3-D reference model based approach. Then, missing regions on semi-frontalized images due to geometric transformation are reconstructed with the help of a denoising stacked autoencoder network. In this phase, missing regions created by line of sight are learned, with a deep architecture, using numerous images. When examined, it can be said that, faces acquired with the proposed approach, are objectively better than the faces acquired with a deep learning or 3-D based method alone. Therefore, it is assumed that the proposed approach can be used in face based computer vision methods as a beneficial pre-processing step.
