Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/427373
Title: | Crossmodal fusion of multiple input modalities for human computer interaction |
Researcher: | Devi N |
Guide(s): | Easwarakumar K S |
Keywords: | Human-computer interaction Multimodal interaction Multimodal human computer |
University: | Anna University |
Completed Date: | 2021 |
Abstract: | Modern technologies provide efficient implementation of a large newlinevariety of human-centered ubiquitous computer applications in the emergent newlinee-Society such as intelligent networked spaces for living, working, health and newlineelderly care, shopping and entertainment. In order to augment the newlinefunctionality and user-satisfaction of these e-Society applications, it newlinenecessitates to develop more accessible and unobtrusive, user-oriented and newlineuser friendly, human-computer interaction modalities allowing people with newlinediverse abilities to seamlessly access information anytime, anywhere and to newlineinteract naturally and intuitively with these new intelligent spaces. The prime newlineobjective of this dissertation is the design of a generic scalable multimodal newlineframework that enables rapid development of multimodal human computer newlineinteraction based applications using variety of modalities and methods for newlineambiguity resolution. newlineThe multimodal interaction framework proposed in this dissertation newlinetakes multimodal utterances involving speech and continuous double hand newlinegestures with elbow movement and facial cues from the user. Semantic partial newlineframes of the likely user intentions are generated by the individual modality newlinerecognizer and disseminated to the inference layers. The inference layer takes newlinethe semantic partial frames from various recognizers and fuses their results newlinewith the help of application context into a complete semantic frame for the newlineapplication. Cross modal fusion technique is then used to fuse the features newlineemerging from different modality sources and a multimodal learning system newlineis used to infer the semantic interpretation for the higher level application newlineevents. The application accepts semantic frames and provides feedback to the newlineuser through multimodal output. newline |
Pagination: | xvi,134p. |
URI: | http://hdl.handle.net/10603/427373 |
Appears in Departments: | Faculty of Information and Communication Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 235.91 kB | Adobe PDF | View/Open |
02_prelim pages.pdf | 2.31 MB | Adobe PDF | View/Open | |
03_contents.pdf | 203.97 kB | Adobe PDF | View/Open | |
04_abstracts.pdf | 170.44 kB | Adobe PDF | View/Open | |
05_chapter 1.pdf | 586.19 kB | Adobe PDF | View/Open | |
06_chapter2.pdf | 1.43 MB | Adobe PDF | View/Open | |
07_chapter3.pdf | 1.6 MB | Adobe PDF | View/Open | |
08_chapter 4.pdf | 884.18 kB | Adobe PDF | View/Open | |
09_chapter5.pdf | 871.7 kB | Adobe PDF | View/Open | |
10_chapter 6.pdf | 707.19 kB | Adobe PDF | View/Open | |
11_chapter 7.pdf | 900.94 kB | Adobe PDF | View/Open | |
12_annexures.pdf | 92.09 kB | Adobe PDF | View/Open | |
80_recommendation.pdf | 63.97 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: