Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/423737
Title: | Development of Framework for Facial Expression Analysis Using Representation Learning |
Researcher: | Singh, Vivek |
Guide(s): | Kumar, Vinay |
Keywords: | Engineering Engineering and Technology Engineering Electrical and Electronic |
University: | Thapar Institute of Engineering and Technology |
Completed Date: | 2020 |
Abstract: | Facial expressions play a crucial role in human social interaction; and this is the primary component needed to be integrated into machines to make human computer interaction more user friendly. Although, humans are very efficient in recognizing even the minute changes in facial expression but for machines it is a very complex task. Recently, this area of research has attracted much needed attention due to its broad spectrum of application. However, expression analysis in an unconstrained environment is a very difficult task. Variations in illumination, facial features, head pose and changes in background make it very difficult to correctly recognize emotions in an open setup for commercial applications. This thesis develops deep learning based representation learning methods for analyzing facial expressions. In this work, multiple frameworks are developed for different applications of facial expression analysis. First proposed framework analyzes the emotional sentiment represented by an image based on its content. The proposed system investigates the faces and background in the image, and extracts facial and scene features from them, respectively. Two different convolutional neural networks are used to extract facial and scene features. Conditional occurrence of these features is modeled using long short term memory networks to predict the sentiment represented by the image. The second framework, proposed in this work, predicts likability of the multimedia content based on the facial expression of the viewer. A database with two different sets of video samples was collected for the task under unconstrained environment. First set of samples consists of videos to be watched by recruited subjects called as stimulants. Second set of samples are recordings of facial expressions of subjects while watching stimulants. The proposed framework is a multimodal system which learns spatio-temporal features from the videos of subject to predict the likability. |
Pagination: | 117p. |
URI: | http://hdl.handle.net/10603/423737 |
Appears in Departments: | Department of Electronics and Communication Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 91.39 kB | Adobe PDF | View/Open |
02_prelim pages.pdf | 1.82 MB | Adobe PDF | View/Open | |
03_content.pdf | 274.02 kB | Adobe PDF | View/Open | |
04_abstract.pdf | 339.88 kB | Adobe PDF | View/Open | |
05_chapter 1.pdf | 1.04 MB | Adobe PDF | View/Open | |
06_chapter 2.pdf | 2.19 MB | Adobe PDF | View/Open | |
07_chapter 3.pdf | 2.81 MB | Adobe PDF | View/Open | |
08_chapter 4.pdf | 2.88 MB | Adobe PDF | View/Open | |
09_chapter 5.pdf | 3.18 MB | Adobe PDF | View/Open | |
10_chapter 6.pdf | 3.25 MB | Adobe PDF | View/Open | |
11_chapter 7.pdf | 545.9 kB | Adobe PDF | View/Open | |
12_annexures.pdf | 2.92 MB | Adobe PDF | View/Open | |
80_recommendation.pdf | 539.08 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: