Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/566010
Title: A Deep Learning Framework for Detection and Segmentation of Multiple Artefacts in Endoscopic Images
Researcher: Kirthika N
Guide(s): Sargunam B
Keywords: Engineering and Technology
Engineering
Electrical and Communication
University: Avinashilingam Institute for Home Science and Higher Education for Women
Completed Date: 2024
Abstract: Endoscopy is a standard procedure for disease surveillance, monitoring inflammations, detect cancer and tumor. During the procedure the organs are visualized. Artefacts, an artificial effect is found to be present in the resultant images. They play a dominant role in increasing procedure time by more than an hour. Hence an efficient algorithm to detect, segment and restore could assist clinician. The artefacts present in an endoscopic image include saturation, specular reflections, blur, bubbles, contrast, blood, instruments and miscellaneous artefacts. The presence of these artefacts acts as a barrier when investigating the underlying tissue for identifying clinical abnormalities. It also affect post processing steps where most of the images captured are discarded due to the presence of artefacts which in turn affects information storage and extracting useful frame for report generation. Endoscopic artefact detection dataset is the only available public dataset holding endoscopic images with annotations for multiple artefacts. Hence, a custom dataset is annotated using the same annotation protocol of endoscopic artefact detection dataset to maintain homogeneity. The algorithms are trained and tested with images from both public and custom dataset for artefact detection. State of the art object detection algorithms such as YOLOv3, YOLOv4 and faster R-CNN are used for detecting artefacts in endoscopic images. The detection algorithm focusses on three important performance parameters namely mean average precision, intersection over union and inference time. The ensemble model outperformed well across all the performance parameters compared with literature. The inference time is reduced by 8.63%, whereas the mAP and IoU are increased by 61.67% and 63.47% respectively. newlineSegmentation algorithms like U-Net with EfficientNetB3 backbone, Link-Net with EfficientNetB3 backbone and U-Net with SE-ResNeXt101 backbone are used to segment the artefacts. The results are assessed with performance parameters like F2 score and Jacca
Pagination: 126 p.
URI: http://hdl.handle.net/10603/566010
Appears in Departments:Department of Electronics and Communication

Files in This Item:
File Description SizeFormat 
02_prelimpages.pdfAttached File1.76 MBAdobe PDFView/Open
03_contents.pdf568.19 kBAdobe PDFView/Open
04_abstract.pdf469.25 kBAdobe PDFView/Open
05_chapter 1.pdf5.94 MBAdobe PDFView/Open
07_chapter 3.pdf7.34 MBAdobe PDFView/Open
09_chapter 5.pdf2.23 MBAdobe PDFView/Open
10_chapter 6.pdf6.43 MBAdobe PDFView/Open
11_annexure.pdf319.62 kBAdobe PDFView/Open
80_recommendation.pdf200 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: