Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/522233
Title: | Design and development of context aware deep learning models for image captioning |
Researcher: | Lalitha B |
Guide(s): | Gomathi V |
Keywords: | Computer Science Computer Science Information Systems Engineering and Technology Image captioning MDAA RNN |
University: | Anna University |
Completed Date: | 2023 |
Abstract: | Image captioning is a brand-new study area in the science of computer vision. The primary goal of picture captioning is to create a natural language description for the input image. In recent years, research on natural language processing and computer vision has become increasingly interested in the problem of automatically synthesising descriptive phrases for photos. Image captioning is a crucial task that demands both the ability to create precise and accurate description phrases as well as a semantic understanding of the images. The objective of image captioning is to develop a natural language representation of an image. It is a difficult task to define the content of an image. To facilitate elaborate definition, the detection and recognition of objects, people, correlations and corresponding features is required. In order to execute Image Captioning methods, these present computer vision tasks must not only collect information contained in images but also extract semantic associations of acquired visual information corresponding to verbal expressions. Open domain captioning is a big challenge, due to its requirement for a deep knowledge on the global and the local entities present in an image, in addition to their features and associations. The template-based techniques, at first maps every sentence pieces (e.g., subject, verb, object) with the words identified from image content and then the sentence is generated with predefined language samples. Apparently, a majority of them highly rely on the samples of sentence and always develop sentences that are syntactically structured newline |
Pagination: | xxii,163 |
URI: | http://hdl.handle.net/10603/522233 |
Appears in Departments: | Faculty of Information and Communication Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 118.4 kB | Adobe PDF | View/Open |
02_prelim_pages.pdf | 3.59 MB | Adobe PDF | View/Open | |
03_content.pdf | 214.35 kB | Adobe PDF | View/Open | |
04_abstract.pdf | 199.24 kB | Adobe PDF | View/Open | |
05_chapter 1.pdf | 1.13 MB | Adobe PDF | View/Open | |
06_chapter 2.pdf | 503.11 kB | Adobe PDF | View/Open | |
07_chapter 3.pdf | 1.3 MB | Adobe PDF | View/Open | |
08_chapter 4.pdf | 1.37 MB | Adobe PDF | View/Open | |
09_chapter 5.pdf | 1 MB | Adobe PDF | View/Open | |
10_annexures.pdf | 2.1 MB | Adobe PDF | View/Open | |
80_recommendation.pdf | 57.57 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: