Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/302565
Title: Developing Efficient Deep Architectures for Classification Task
Researcher: Saduf
Guide(s): Wani, M. Arif
Keywords: Computer Science
Deep Architecture- Computers
Engineering and Technology
University: University of Kashmir
Completed Date: NA
Abstract: Deep architectures are enjoying an increasing popularity due to its success in solving complex problems. In particular, deep architectures have proven to be effective in a large variety of classification tasks. Contrary to previous research, which required engineered feature representations, designed by experts, in order to succeed, deep architectures attempt to learn representation hierarchies automatically from data. The multi-layer architecture of these networks is particularly useful in capturing the hierarchical structure of the given data: simple features are detected at lower layers and fed into higher layers for extracting abstract representations. Despite the remarkable representational power of deep networks, training these models is computationally expensive. In addition, considering the lack of enough labeled training data in many applications, over-fitting is a serious threat for deep models with large number of free parameters. Also, there are innate issues with the gradient-based optimization procedure used for parameter learning in these models. Therefore, the search for algorithms to optimize the learning of deep architectures is extensive and ongoing. newlineIn this thesis, we tackle the challenging problem of optimizing the learning of deep architectures. We first draw up a state-of-the-art review of the deep architectures specifically multilayered feed forward networks aiming to understand the various techniques currently used to train deep architectures. We then propose and explore the use of two phase strategy in training of deep architectures. We propose a deep architecture where successive layers of units are pretrained using unsupervised learning. The second phase involves supervised fine tuning of various layers. For performing supervised fine tuning we introduce a set of algorithms that alleviate the problems associated with the conventional approach of fine tuning. We study the applicability and potential of the proposed architecture on a number of benchmark datasets, highlighting the .....
Pagination: 
URI: http://hdl.handle.net/10603/302565
Appears in Departments:Department of Computer Science

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File33.68 kBAdobe PDFView/Open
02_certificate.pdf83.75 kBAdobe PDFView/Open
03_declaration.pdf67.89 kBAdobe PDFView/Open
04_acknowledgements.pdf228.53 kBAdobe PDFView/Open
05 abstract.pdf146.17 kBAdobe PDFView/Open
06_table_of_contents.pdf92.39 kBAdobe PDFView/Open
07_list_of_figures.pdf313.38 kBAdobe PDFView/Open
08_list_of_tables.pdf81.62 kBAdobe PDFView/Open
09_list of acronym.pdf151.5 kBAdobe PDFView/Open
10_chapter_1.pdf245.86 kBAdobe PDFView/Open
11_chapter_2.pdf809.39 kBAdobe PDFView/Open
12_chapter_3.pdf702.34 kBAdobe PDFView/Open
13_chapter_4.pdf621.74 kBAdobe PDFView/Open
14_chapter_5.pdf988.3 kBAdobe PDFView/Open
15_chapter_6.pdf739.59 kBAdobe PDFView/Open
16_chapter_7.pdf188.81 kBAdobe PDFView/Open
17_publications.pdf88.85 kBAdobe PDFView/Open
18_references.pdf203.43 kBAdobe PDFView/Open
80_recommendation.pdf188.81 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: