Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/339409
Full metadata record
DC FieldValueLanguage
dc.coverage.spatialAn enhanced framework for identifying human emotions from multimodal signals
dc.date.accessioned2021-09-07T05:39:17Z-
dc.date.available2021-09-07T05:39:17Z-
dc.identifier.urihttp://hdl.handle.net/10603/339409-
dc.description.abstractThe notion of emotions is very vital when one person wants to communicate with another person. When a child is sorrowful, it tends to cry. The father or the mother of the child will try to make the child happy by giving what the child needs and the child laughs, which makes the father and mother happy. This happiness cannot be expressed because it will be beyond the measure. Similarly emotions play an important role when a human wants to interact with a machine. Take for example an automated vehicle assist system, making the vehicle understand the emotions of the driver will make the driving experience better. Another example is automated telephony wherein the emotions of the caller can be analyzed in order to improve the feedback of the system. Understanding emotions from image, speech signal, and video is of major concern in this research work. In this thesis we first work on image processing, for identifying emotions. Second we work on speech signal for emotion identification. Finally we use video to identify emotions. In the first framework the emotion of a person is identified from an image with the help of tensorflow. It is an upcoming technology in the field of artificial intelligence. Tensorflow has grown to help users to determine what they need with respect to their daily life. One of the most talked about tensorflow applications is the RankBrain developed by Google corporation which allows users to find the relevant pages with the help of deep neural nets. For our purpose we use tensorflow to identify emotions from an image using the features extracted from the image. Basically the feature extraction is done by identifying the geometry of the face after detecting the landmarks of the eyes and mouth of the face. The landmarks are constructed by using the proposed modified eyemap-mouthmap algorithm on an enhanced image which uses discrete wavelet transform and fuzzy for enhancement. Results of classification show that the proposed methodology is better when tensorflow is used. The second framework on em
dc.format.extentxvii,129 p.
dc.languageEnglish
dc.relationThe notion of emotions is very vital when one person wants to communicate with another person. When a child is sorrowful, it tends to cry. The father or the mother of the child will try to make the child happy by giving what the child needs and the child laughs, which makes the father and mother happy. This happiness cannot be expressed because it will be beyond the measure. Similarly emotions play an important role when a human wants to interact with a machine. Take for example an automated vehicle assist system, making the vehicle understand the emotions of the driver will make the driving experience better. Another example is automated telephony wherein the emotions of the caller can be analyzed in order to improve the feedback of the system. Understanding emotions from image, speech signal, and video is of major concern in this research work. In this thesis we first work on image processing, for identifying emotions. Second we work on speech signal for emotion identification. Finally we use video to identify emotions. In the first framework the emotion of a person is identified from an image with the help of tensorflow. It is an upcoming technology in the field of artificial intelligence. Tensorflow has grown to help users to determine what they need with respect to their daily life. One of the most talked about tensorflow applications is the RankBrain developed by Google corporation which allows users to find the relevant pages with the help of deep neural nets. For our purpose we use tensorflow to identify emotions from an image using the features extracted from the image. Basically the feature extraction is done by identifying the geometry of the face after detecting the landmarks of the eyes and mouth of the face. The landmarks are constructed by using the proposed modified eyemap-mouthmap algorithm on an enhanced image which uses discrete wavelet transform and fuzzy for enhancement. Results of classification show that the proposed methodology is better when tensorflow is used. The second framework on emotion identification involves speech signals. Here we propose four different methods for speech emotion identification. The first method uses the formants detected from the input speech signal for classification. The proposed method for formant detection involves preprocessing the input speech signal followed by single level discrete wavelet transform and then finding the linear predictor coefficients. The different classifiers namely the K*, neural net, and random forest were employed for classification. The second method uses the proposed transform based Euclidean distance measure for identifying emotions. Two different transforms namely the fast Fourier transform and the discrete wavelet transform were used to extract the sub features and then the main features were extracted with the help of Euclidean distance measure which were further normalized with the use of a Gaussian probability distribution. Classification was performed using three different classifiers viz., the bagging, support vector machine, and random forest, respectively. The third method uses the proposed emotional diversity index in order to identify emotions. Here after applying discrete wavelet transform, the detail coefficients of the preprocessed signal is given to the proposed emotional diversity index system which comprises of peak detection, mode calculation and diversity identification. The identified diversity is assigned a positive diversity index and a negative diversity index based on the calculations. The two different diversity values are further processed and then classified using multiple classifiers
dc.rightsuniversity
dc.titleAn enhanced framework for identifying human emotions from multimodal signals
dc.title.alternative
dc.creator.researcherAllen Joseph, R
dc.subject.keywordEngineering and Technology
dc.subject.keywordComputer Science
dc.subject.keywordComputer Science Software Engineering
dc.description.note
dc.contributor.guideGeetha, P
dc.publisher.placeChennai
dc.publisher.universityAnna University
dc.publisher.institutionFaculty of Information and Communication Engineering
dc.date.registeredn.d.
dc.date.completed2020
dc.date.awarded2020
dc.format.dimensions21cm
dc.format.accompanyingmaterialNone
dc.source.universityUniversity
dc.type.degreePh.D.
Appears in Departments:Faculty of Information and Communication Engineering

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File190.71 kBAdobe PDFView/Open
02_certificates.pdf162.85 kBAdobe PDFView/Open
03_vivaproceedings.pdf415.96 kBAdobe PDFView/Open
04_bonafidecertificate.pdf263.19 kBAdobe PDFView/Open
05_abstracts.pdf47.12 kBAdobe PDFView/Open
06_acknowledgements.pdf305.03 kBAdobe PDFView/Open
07_contents.pdf90.52 kBAdobe PDFView/Open
08_listoftables.pdf54.66 kBAdobe PDFView/Open
09_listoffigures.pdf83.39 kBAdobe PDFView/Open
10_listofabbreviations.pdf49.15 kBAdobe PDFView/Open
11_chapter1.pdf1.62 MBAdobe PDFView/Open
12_chapter2.pdf133.29 kBAdobe PDFView/Open
13_chapter3.pdf1.21 MBAdobe PDFView/Open
14_chapter4.pdf4.12 MBAdobe PDFView/Open
15_chapter5.pdf1.28 MBAdobe PDFView/Open
16_chapter6.pdf91.92 kBAdobe PDFView/Open
17_conclusion.pdf91.92 kBAdobe PDFView/Open
18_references.pdf122.54 kBAdobe PDFView/Open
19_listofpublications.pdf90.48 kBAdobe PDFView/Open
80_recommendation.pdf90.5 kBAdobe PDFView/Open


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: