Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/507463
Full metadata record
DC FieldValueLanguage
dc.coverage.spatial
dc.date.accessioned2023-08-16T05:39:11Z-
dc.date.available2023-08-16T05:39:11Z-
dc.identifier.urihttp://hdl.handle.net/10603/507463-
dc.description.abstractThere is a strong incentive to build intelligent machines that can understand and adapt to changes in the visual world without human supervision. While humans and animals learn to perceive the world on their own, almost all state-of-the-art vision systems heavily rely on external supervision from millions of manually annotated training examples. Gathering such large-scale manual annotations for structured vision tasks, such as monocular depth estimation, scene segmentation, human pose estimation, faces several practical limitations. Usually, the annotations are gathered in two broad ways; 1) via specialized instruments (sensors) or laboratory setups, 2) via manual annotations. Both processes have several drawbacks. While human annotations are expensive, scarce, or error-prone; instrument-based annotations are often noisy or limited to specific laboratory environments. Such limitations not only stand as a major bottleneck in our efforts to gather unambiguous ground-truth but also limit the diversity in the collected labeled dataset. This motivates us to develop innovative ways to utilize synthetic environments to create labeled synthetic datasets with noise-free unambiguous ground-truths. However, the performance of models trained on such synthetic data markedly degrades when tested on real-world samples due to input distribution shift (a.k.a. domain shift). Unsupervised domain adaptation (DA) seeks learning techniques that can minimize the domain discrepancy between a labeled source and an unlabeled target. However, it mostly remains unexplored for challenging structured prediction based vision tasks. Motivated by the above observations, my research focuses on addressing the following key aspects: (1) Developing algorithms that support improved transferability to domain and task shifts, (2) Leveraging inter-entity or cross-modal relationships to develop self-supervised objectives, and (3) Instilling natural priors to constrain the model output within the realm of natural distributions. First, we present AdaDep...
dc.format.extent
dc.languageEnglish
dc.relation
dc.rightsuniversity
dc.titleSelf Supervised Domain Adaptation Frameworks for Computer Vision Tasks
dc.title.alternativeSelf-Supervised Domain Adaptation Frameworks for Computer Vision Tasks
dc.creator.researcherKundu, Jogendra Nath
dc.subject.keywordComputer Science
dc.subject.keywordComputer Science Software Engineering
dc.subject.keywordEngineering and Technology
dc.description.note
dc.contributor.guideVenkatesh Babu, R
dc.publisher.placeBangalore
dc.publisher.universityIndian Institute of Science Bangalore
dc.publisher.institutionComputational and Data Sciences
dc.date.registered
dc.date.completed2022
dc.date.awarded2023
dc.format.dimensions
dc.format.accompanyingmaterialNone
dc.source.universityUniversity
dc.type.degreePh.D.
Appears in Departments:Computational and Data Sciences



Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: