Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/427373
Title: Crossmodal fusion of multiple input modalities for human computer interaction
Researcher: Devi N
Guide(s): Easwarakumar K S
Keywords: Human-computer interaction
Multimodal interaction
Multimodal human computer
University: Anna University
Completed Date: 2021
Abstract: Modern technologies provide efficient implementation of a large newlinevariety of human-centered ubiquitous computer applications in the emergent newlinee-Society such as intelligent networked spaces for living, working, health and newlineelderly care, shopping and entertainment. In order to augment the newlinefunctionality and user-satisfaction of these e-Society applications, it newlinenecessitates to develop more accessible and unobtrusive, user-oriented and newlineuser friendly, human-computer interaction modalities allowing people with newlinediverse abilities to seamlessly access information anytime, anywhere and to newlineinteract naturally and intuitively with these new intelligent spaces. The prime newlineobjective of this dissertation is the design of a generic scalable multimodal newlineframework that enables rapid development of multimodal human computer newlineinteraction based applications using variety of modalities and methods for newlineambiguity resolution. newlineThe multimodal interaction framework proposed in this dissertation newlinetakes multimodal utterances involving speech and continuous double hand newlinegestures with elbow movement and facial cues from the user. Semantic partial newlineframes of the likely user intentions are generated by the individual modality newlinerecognizer and disseminated to the inference layers. The inference layer takes newlinethe semantic partial frames from various recognizers and fuses their results newlinewith the help of application context into a complete semantic frame for the newlineapplication. Cross modal fusion technique is then used to fuse the features newlineemerging from different modality sources and a multimodal learning system newlineis used to infer the semantic interpretation for the higher level application newlineevents. The application accepts semantic frames and provides feedback to the newlineuser through multimodal output. newline
Pagination: xvi,134p.
URI: http://hdl.handle.net/10603/427373
Appears in Departments:Faculty of Information and Communication Engineering

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File235.91 kBAdobe PDFView/Open
02_prelim pages.pdf2.31 MBAdobe PDFView/Open
03_contents.pdf203.97 kBAdobe PDFView/Open
04_abstracts.pdf170.44 kBAdobe PDFView/Open
05_chapter 1.pdf586.19 kBAdobe PDFView/Open
06_chapter2.pdf1.43 MBAdobe PDFView/Open
07_chapter3.pdf1.6 MBAdobe PDFView/Open
08_chapter 4.pdf884.18 kBAdobe PDFView/Open
09_chapter5.pdf871.7 kBAdobe PDFView/Open
10_chapter 6.pdf707.19 kBAdobe PDFView/Open
11_chapter 7.pdf900.94 kBAdobe PDFView/Open
12_annexures.pdf92.09 kBAdobe PDFView/Open
80_recommendation.pdf63.97 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: