Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/427373
Full metadata record
DC FieldValueLanguage
dc.coverage.spatialCrossmodal fusion of multiple input modalities for human computer interaction
dc.date.accessioned2022-12-18T09:03:42Z-
dc.date.available2022-12-18T09:03:42Z-
dc.identifier.urihttp://hdl.handle.net/10603/427373-
dc.description.abstractModern technologies provide efficient implementation of a large newlinevariety of human-centered ubiquitous computer applications in the emergent newlinee-Society such as intelligent networked spaces for living, working, health and newlineelderly care, shopping and entertainment. In order to augment the newlinefunctionality and user-satisfaction of these e-Society applications, it newlinenecessitates to develop more accessible and unobtrusive, user-oriented and newlineuser friendly, human-computer interaction modalities allowing people with newlinediverse abilities to seamlessly access information anytime, anywhere and to newlineinteract naturally and intuitively with these new intelligent spaces. The prime newlineobjective of this dissertation is the design of a generic scalable multimodal newlineframework that enables rapid development of multimodal human computer newlineinteraction based applications using variety of modalities and methods for newlineambiguity resolution. newlineThe multimodal interaction framework proposed in this dissertation newlinetakes multimodal utterances involving speech and continuous double hand newlinegestures with elbow movement and facial cues from the user. Semantic partial newlineframes of the likely user intentions are generated by the individual modality newlinerecognizer and disseminated to the inference layers. The inference layer takes newlinethe semantic partial frames from various recognizers and fuses their results newlinewith the help of application context into a complete semantic frame for the newlineapplication. Cross modal fusion technique is then used to fuse the features newlineemerging from different modality sources and a multimodal learning system newlineis used to infer the semantic interpretation for the higher level application newlineevents. The application accepts semantic frames and provides feedback to the newlineuser through multimodal output. newline
dc.format.extentxvi,134p.
dc.languageEnglish
dc.relationp.124-133
dc.rightsuniversity
dc.titleCrossmodal fusion of multiple input modalities for human computer interaction
dc.title.alternative
dc.creator.researcherDevi N
dc.subject.keywordHuman-computer interaction
dc.subject.keywordMultimodal interaction
dc.subject.keywordMultimodal human computer
dc.description.note
dc.contributor.guideEaswarakumar K S
dc.publisher.placeChennai
dc.publisher.universityAnna University
dc.publisher.institutionFaculty of Information and Communication Engineering
dc.date.registered
dc.date.completed2021
dc.date.awarded2021
dc.format.dimensions21cm
dc.format.accompanyingmaterialNone
dc.source.universityUniversity
dc.type.degreePh.D.
Appears in Departments:Faculty of Information and Communication Engineering

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File235.91 kBAdobe PDFView/Open
02_prelim pages.pdf2.31 MBAdobe PDFView/Open
03_contents.pdf203.97 kBAdobe PDFView/Open
04_abstracts.pdf170.44 kBAdobe PDFView/Open
05_chapter 1.pdf586.19 kBAdobe PDFView/Open
06_chapter2.pdf1.43 MBAdobe PDFView/Open
07_chapter3.pdf1.6 MBAdobe PDFView/Open
08_chapter 4.pdf884.18 kBAdobe PDFView/Open
09_chapter5.pdf871.7 kBAdobe PDFView/Open
10_chapter 6.pdf707.19 kBAdobe PDFView/Open
11_chapter 7.pdf900.94 kBAdobe PDFView/Open
12_annexures.pdf92.09 kBAdobe PDFView/Open
80_recommendation.pdf63.97 kBAdobe PDFView/Open


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: