Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/519875
Title: Certain investigation for text classification on fake news data using optimized deep learning techniques
Researcher: Nithya, K
Guide(s): Krishnamoorthi, M and Sampth, P
Keywords: Computer Science
Computer Science Information Systems
Deep learning
Engineering and Technology
Fake news data
Text classification
University: Anna University
Completed Date: 2023
Abstract: In social media platforms, numerous data are generated every second newlineand the users of these platforms share the contents instantly when they like to newlineshare. The contents posted in social media are mostly unverified and there is a newlinehigh possibility of the presence of inaccurate information. Spread of fake newlineinformation on social media misleads the users about some situations or newlinetopics in political and social sectors. Rumour data classification in social newlinemedia seems to be an area of recent research due to its dependency on digital newlinecommunications, and this rumour data make social media unstable. newlineIn the field of Natural Language Processing (NLP), research on rumour newlinedetection has its challenges with word embedding, feature extraction and newlineclassification tasks. Previously, many researches have been conducted, but the newlineaccuracy of rumour detection is a big issue for identifying infrequent/rare newlinewords. The traditional word embedding models like Word2Vec, FastText and newlineso on., perform word level analysis for finding word vectors. In recent days, newlinethese models are replaced with the novel transformer-based BERT model newlinewhich considers the combination of word pieces for producing efficient newlinefeature vectors. Still, it is difficult to identify which word piece of each token newlineis to be used for vector formation. In the first part of the research, an automated fake news detection system is introduced with three different word embedding techniques such as Word2Vec, GloVe and FastText. Further, it is classified by using two newlineproposed algorithms namely Optimized Cost Sensitive Long Short-Term newlineMemory (OC-LSTM) and Optimized Cost Sensitive Gated Recurrent Unit newline(OC-GRU). For the LIAR and Fake and Real News (ISOT) datasets, the newlineFastText embedding with OC-GRU model yields improved accuracy. newlineTo improve the quality and accuracy of features, in the second part of newlinethe research, two phases of feature extraction techniques are involved for newlineperforming word embedding and deep feature extraction. In the first phase, newlineBidirectional Encoder Representations from Tran
Pagination: xix,129p.
URI: http://hdl.handle.net/10603/519875
Appears in Departments:Faculty of Information and Communication Engineering

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File13.24 kBAdobe PDFView/Open
02_prelim pages.pdf2.54 MBAdobe PDFView/Open
03_content.pdf625.3 kBAdobe PDFView/Open
04_abstract.pdf257.79 kBAdobe PDFView/Open
05_chapter 1.pdf482.23 kBAdobe PDFView/Open
06_chapter 2.pdf443.27 kBAdobe PDFView/Open
07_chapter 3.pdf838.49 kBAdobe PDFView/Open
08_chapter 4.pdf1.06 MBAdobe PDFView/Open
09_chapter 5.pdf1.08 MBAdobe PDFView/Open
10_chapter 6.pdf449.33 kBAdobe PDFView/Open
11_annexures.pdf1.91 MBAdobe PDFView/Open
80_recommendation.pdf43.31 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: