Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/473611
Title: An optimized analytical approach for Big Data processing in Cloud Computing Environment
Researcher: Joshiara, Hetal Anilkumar
Guide(s): Thaker, Chirag Surykant
Keywords: Computer Science
Computer Science Software Engineering
Engineering and Technology
University: Gujarat Technological University
Completed Date: 2022
Abstract: In analytics, the speed at which Big Data (BD) is processed has huge importance. For BD analysis, Hadoop is one of the popular frameworks. Map Reduce (MR) is a hugely accepted parallel computing system for larger-scale Data Processing (DP). While processing BD, chief performance metrics in a parallel computing framework is job Execution Time (ET) along with cluster throughput. Performance is severely impacted by the straggler machines while processing BD. The system that consumes prolonged time for completing the task is termed a straggler machine. Moreover, in a parallel computing framework, one single slow task or slow node can severely affect the performance of overall job execution. newlineBy merely backing up these slow-running tasks on substitute machines, the most common approach to deal with stragglers is Speculative Execution (SE). The researchers proposed many SE strategies; however, a few drawbacks were included: (A) average progress rate was used which might be misleading to detect the slow task, (B) Phase wise estimation of Remaining Time of the slow-running task and Processing Speed (PS) (C) calculation of slow task s backup time (D) to rerun the task, detect the proper backup node. newlineIn the Cloud Computing (CC) environment, handling the issue of SE would be more challenging in which clusters are heterogeneous as Hadoop believes that nodes are homogeneous. newlineTo enhance BD processing performance in the case of slow-running tasks in the heterogeneous cloud-based Hadoop Cluster (HC), the proposed Speculative Execution strategy renamed Hadoop Supreme Rate Performance (Hadoop-SRP) can be utilized. For precisely and swiftly identifying stragglers the following are wielded (1) To estimate the PS in the Map/Shuffle/Reduce phase considering parameters process speed and process bandwidth, the Modified EWMA methodology is used (2) Calculate the slow tasks RT (3) calculation of slow task s back up (4) By using the BM-LOA (Behavioral Model-based Lion Optimization Algorithm), re-run the slow task on the identified node. newlineB
Pagination: cxiv, 107
URI: http://hdl.handle.net/10603/473611
Appears in Departments:Computer/IT Engineering

Files in This Item:
File Description SizeFormat 
01_title.pdfAttached File73.7 kBAdobe PDFView/Open
02_prelim pages.pdf2.12 MBAdobe PDFView/Open
03_abstract.pdf179.26 kBAdobe PDFView/Open
06_contents.pdf197.37 kBAdobe PDFView/Open
10_annexures.pdf820.41 kBAdobe PDFView/Open
10_chapter1.pdf491.88 kBAdobe PDFView/Open
11_chapter2.pdf251.24 kBAdobe PDFView/Open
12_chapter3.pdf568.09 kBAdobe PDFView/Open
13_chapter4.pdf2.13 MBAdobe PDFView/Open
14_chapter5.pdf668.51 kBAdobe PDFView/Open
17_appendix-a.pdf1.07 MBAdobe PDFView/Open
80_recommendation.pdf137.72 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: