Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/566313
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.coverage.spatial | ||
dc.date.accessioned | 2024-05-24T05:36:04Z | - |
dc.date.available | 2024-05-24T05:36:04Z | - |
dc.identifier.uri | http://hdl.handle.net/10603/566313 | - |
dc.description.abstract | Deep Neural Networks (DNNs) have achieved remarkable success across various machine learning and computer vision tasks, especially when abundant training samples are available. In Convolutional Neural Network (CNN) research, it has been established that a model s generalization capability improves with the combination of complex architectures, strong regularization, domain- specific loss functions, and extensive databases. However, training DNNs in environments with limited data remains a significant challenge, calling for attention from the research community. Many applications lack the requisite volume of data needed to train models effectively. Data constraint in this context is influenced by factors such as 1) a scarcity of domain experts, 2) long-tail distribution in large datasets, 3) insufficient domain-specific data, and 4) the challenge of mimicking human cognition and learning. The issues above are common challenges encountered while designing deep models, underscoring the importance of addressing Data Constrained Learning (DCL). This thesis investigates the formulation of deep learning strategies explicitly tailored for scenarios with DCL. The objective is to ensure that the training of numerous parameters does not adversely affect the model s ability to learn meaningful patterns, as this could elevate the risk of overfitting and result in suboptimal generalization performance. To address the DCL challenge, we introduce a novel strength parameter in deep learning named SSF-CNN, which concentrates on learning both the quotstructurequot and quotstrengthquot of filters. The filter structure is initialized using a dictionary-based filter learning algorithm, while the strength is learned under data-constrained settings. This architecture demonstrates adaptability, delivering robust performance even when used with small databases and consistently attaining high accuracy. We validate the effectiveness of our algorithm on databases such as MNIST, CIFAR10, and NORB, with varying training sample sizes. The results indicate | |
dc.format.extent | 168 p. | |
dc.language | English | |
dc.relation | ||
dc.rights | university | |
dc.title | Data constrained deep learning | |
dc.title.alternative | ||
dc.creator.researcher | Keshari, Rohit | |
dc.subject.keyword | Computer Science | |
dc.subject.keyword | Computer Science Artificial Intelligence | |
dc.subject.keyword | Engineering and Technology | |
dc.description.note | ||
dc.contributor.guide | Singh, Richa and Vatsa, Mayank | |
dc.publisher.place | Delhi | |
dc.publisher.university | Indraprastha Institute of Information Technology, Delhi (IIIT-Delhi) | |
dc.publisher.institution | Computer Science and Engineering | |
dc.date.registered | ||
dc.date.completed | 2024 | |
dc.date.awarded | 2024 | |
dc.format.dimensions | 29 cm. | |
dc.format.accompanyingmaterial | None | |
dc.source.university | University | |
dc.type.degree | Ph.D. | |
Appears in Departments: | Department of Computer Science and Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01-title.pdf | Attached File | 54.64 kB | Adobe PDF | View/Open |
02_prelim pages.pdf | 424.52 kB | Adobe PDF | View/Open | |
03_content.pdf | 99.43 kB | Adobe PDF | View/Open | |
04_abstract.pdf | 67.6 kB | Adobe PDF | View/Open | |
05_chapter 1.pdf | 2.17 MB | Adobe PDF | View/Open | |
06_chapter 2.pdf | 1.85 MB | Adobe PDF | View/Open | |
07_chapter 3.pdf | 1.5 MB | Adobe PDF | View/Open | |
08_chapter 4.pdf | 1.64 MB | Adobe PDF | View/Open | |
09_chapter 5.pdf | 380.81 kB | Adobe PDF | View/Open | |
10_annexures.pdf | 555.26 kB | Adobe PDF | View/Open | |
80_recommendation.pdf | 560.84 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: