Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/540379
Title: A framework for continuous indian sign language recognition using computer vision
Researcher: Amrutha, K
Guide(s): Prabu, P
Keywords: Computer Science
Computer Science Artificial Intelligence
Engineering and Technology
InceptionV3,
INCLUDE dataset,
ISL,
LSTM.
SLT,
University: CHRIST University
Completed Date: 2023
Abstract: Sign language is a non-vocal, visually oriented natural language used by the hearing newlineimpaired and the hard-for-hearing part of society. It combines multiple modalities newlinelike hand movements, facial expressions and body poses. Static gestures involve basic finger movements such as numbers and alphabets, dynamic signs include words, and a sign sentence consists of grammatically connected and meaningful dynamic words. Sign Language Translation (SLT) models have been an actively evolving research topic under computer vision. One of the most challenging aspects in earlier iterations of SLTs was accurately capturing the intricate and constantly changing hand movements and facial expressions characteristic of sign language. newlineHowever, the advent of deep learning models has facilitated significant advancements in the field, particularly in the realm of continuous sign language translation. newlineThe research endeavours to develop a lightweight deep-learning framework newlinespecifically tailored for the translation of Indian Sign Language (ISL) into text and newlineaudio. The proposed framework introduces two collaborative deep-learning components that extract and classify features synergistically. The ISL video sequence serves as the input, which undergoes feature extraction utilizing the Inception V3 architecture, enabling the extraction of features from each frame. Classification models tend to be bulky and intricate, consuming substantial memory space and requiring extended training periods. This challenge has been addressed by introducing a lightweight LSTM model, which effectively utilizes the feature map generated by the Inception model for accurate classification. It is important to note that each sign possesses unique characteristics yet exhibits similar feature maps. The performance of the framework is assessed based on the speed and accuracy achieved in converting the input video into text and audio formats.
Pagination: xix, 151p.;
URI: http://hdl.handle.net/10603/540379
Appears in Departments:Department of Computer Science

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File189.02 kBAdobe PDFView/Open
02_prelim pages.pdf1.08 MBAdobe PDFView/Open
03_abstract.pdf8.78 kBAdobe PDFView/Open
04_table_of_contents.pdf80.09 kBAdobe PDFView/Open
05_chapter1.pdf181.49 kBAdobe PDFView/Open
06_chapter2.pdf581.01 kBAdobe PDFView/Open
07_chapter3.pdf1.02 MBAdobe PDFView/Open
08_chapter4.pdf1.7 MBAdobe PDFView/Open
09_chapter5.pdf629.71 kBAdobe PDFView/Open
10_chapter6.pdf128.31 kBAdobe PDFView/Open
11_annexures.pdf157.15 kBAdobe PDFView/Open
80_recommendation.pdf309.66 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: