Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/422616
Title: | Some approaches for detecting multi lingual and multi text from natural scene images |
Researcher: | Aparna Y |
Guide(s): | Valli S |
Keywords: | Engineering and Technology Engineering Engineering Multidisciplinary Multi-lingual Framework |
University: | Anna University |
Completed Date: | 2022 |
Abstract: | Content in the text format helps to communicate the relevant and specific newlineinformation to users meticulously. Text detection in natural scene images finds newlineseveral applications in computer vision systems such as reading license plates, newlinedetecting street signs, retrieving images, performing mobile visual searches, newlineidentifying image-based geolocation, and assisting visually impaired persons. newlineText, an important way of communication, provides significant information for newlineannotation, indexing, and image structuring. During communication, text can be newlineconveniently embedded in scenes or documents and text information in images is newlineeasily perceivable by everyone. But text extraction is a challenging problem due newlineto the inconsistency in text size, style, color, orientation, and alignment. Also, newlineimages with low contrast, blur, noise, varying illumination, complex background, newlineand multi-lingual environment add to this difficulty. This research aims to newlineovercome the existing difficulty in scene text extraction by developing approaches newlinefor detecting text from natural scene images for various quadrilateral-type and newlinepolygon-type datasets. newlineThe first work uses amended maximally stable extremal region (a-MSER) newlinetogether with deep learning framework, You Only Look Once (YOLOv2) network. newlineThe a-MSER method is used to identify the region of interest based on the newlinevariation of MSER. This algorithm considers intensity changes between text and newlinebackground very effectively.The drawback of original YOLOv2, the poor detection newlinerate for small-sized objects, is overcome by employing a 1 × 1 layer with image newlinesize enhanced from 13 × 13 to 26 × 26. Focal loss is applied to improve newlineupon the existing cross entropy classification loss of YOLOv2 newline |
Pagination: | xviii, 131 p. |
URI: | http://hdl.handle.net/10603/422616 |
Appears in Departments: | Faculty of Information and Communication Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 190.34 kB | Adobe PDF | View/Open |
02_prelim pages.pdf | 1.68 MB | Adobe PDF | View/Open | |
03_content.pdf | 82.88 kB | Adobe PDF | View/Open | |
04_abstract.pdf | 58.31 kB | Adobe PDF | View/Open | |
05_chapter 1.pdf | 2.45 MB | Adobe PDF | View/Open | |
06_chapter 2.pdf | 2.03 MB | Adobe PDF | View/Open | |
07_chapter 3.pdf | 3.66 MB | Adobe PDF | View/Open | |
08_chapter 4.pdf | 7.33 MB | Adobe PDF | View/Open | |
09_chapter 5.pdf | 8.88 MB | Adobe PDF | View/Open | |
10_chapter 6.pdf | 6.91 MB | Adobe PDF | View/Open | |
11_annexures.pdf | 550.94 kB | Adobe PDF | View/Open | |
80_recommendation.pdf | 192.74 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: