Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/302565
Title: | Developing Efficient Deep Architectures for Classification Task |
Researcher: | Saduf |
Guide(s): | Wani, M. Arif |
Keywords: | Computer Science Deep Architecture- Computers Engineering and Technology |
University: | University of Kashmir |
Completed Date: | NA |
Abstract: | Deep architectures are enjoying an increasing popularity due to its success in solving complex problems. In particular, deep architectures have proven to be effective in a large variety of classification tasks. Contrary to previous research, which required engineered feature representations, designed by experts, in order to succeed, deep architectures attempt to learn representation hierarchies automatically from data. The multi-layer architecture of these networks is particularly useful in capturing the hierarchical structure of the given data: simple features are detected at lower layers and fed into higher layers for extracting abstract representations. Despite the remarkable representational power of deep networks, training these models is computationally expensive. In addition, considering the lack of enough labeled training data in many applications, over-fitting is a serious threat for deep models with large number of free parameters. Also, there are innate issues with the gradient-based optimization procedure used for parameter learning in these models. Therefore, the search for algorithms to optimize the learning of deep architectures is extensive and ongoing. newlineIn this thesis, we tackle the challenging problem of optimizing the learning of deep architectures. We first draw up a state-of-the-art review of the deep architectures specifically multilayered feed forward networks aiming to understand the various techniques currently used to train deep architectures. We then propose and explore the use of two phase strategy in training of deep architectures. We propose a deep architecture where successive layers of units are pretrained using unsupervised learning. The second phase involves supervised fine tuning of various layers. For performing supervised fine tuning we introduce a set of algorithms that alleviate the problems associated with the conventional approach of fine tuning. We study the applicability and potential of the proposed architecture on a number of benchmark datasets, highlighting the ..... |
Pagination: | |
URI: | http://hdl.handle.net/10603/302565 |
Appears in Departments: | Department of Computer Science |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 33.68 kB | Adobe PDF | View/Open |
02_certificate.pdf | 83.75 kB | Adobe PDF | View/Open | |
03_declaration.pdf | 67.89 kB | Adobe PDF | View/Open | |
04_acknowledgements.pdf | 228.53 kB | Adobe PDF | View/Open | |
05 abstract.pdf | 146.17 kB | Adobe PDF | View/Open | |
06_table_of_contents.pdf | 92.39 kB | Adobe PDF | View/Open | |
07_list_of_figures.pdf | 313.38 kB | Adobe PDF | View/Open | |
08_list_of_tables.pdf | 81.62 kB | Adobe PDF | View/Open | |
09_list of acronym.pdf | 151.5 kB | Adobe PDF | View/Open | |
10_chapter_1.pdf | 245.86 kB | Adobe PDF | View/Open | |
11_chapter_2.pdf | 809.39 kB | Adobe PDF | View/Open | |
12_chapter_3.pdf | 702.34 kB | Adobe PDF | View/Open | |
13_chapter_4.pdf | 621.74 kB | Adobe PDF | View/Open | |
14_chapter_5.pdf | 988.3 kB | Adobe PDF | View/Open | |
15_chapter_6.pdf | 739.59 kB | Adobe PDF | View/Open | |
16_chapter_7.pdf | 188.81 kB | Adobe PDF | View/Open | |
17_publications.pdf | 88.85 kB | Adobe PDF | View/Open | |
18_references.pdf | 203.43 kB | Adobe PDF | View/Open | |
80_recommendation.pdf | 188.81 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: