Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/473611
Title: | An optimized analytical approach for Big Data processing in Cloud Computing Environment |
Researcher: | Joshiara, Hetal Anilkumar |
Guide(s): | Thaker, Chirag Surykant |
Keywords: | Computer Science Computer Science Software Engineering Engineering and Technology |
University: | Gujarat Technological University |
Completed Date: | 2022 |
Abstract: | In analytics, the speed at which Big Data (BD) is processed has huge importance. For BD analysis, Hadoop is one of the popular frameworks. Map Reduce (MR) is a hugely accepted parallel computing system for larger-scale Data Processing (DP). While processing BD, chief performance metrics in a parallel computing framework is job Execution Time (ET) along with cluster throughput. Performance is severely impacted by the straggler machines while processing BD. The system that consumes prolonged time for completing the task is termed a straggler machine. Moreover, in a parallel computing framework, one single slow task or slow node can severely affect the performance of overall job execution. newlineBy merely backing up these slow-running tasks on substitute machines, the most common approach to deal with stragglers is Speculative Execution (SE). The researchers proposed many SE strategies; however, a few drawbacks were included: (A) average progress rate was used which might be misleading to detect the slow task, (B) Phase wise estimation of Remaining Time of the slow-running task and Processing Speed (PS) (C) calculation of slow task s backup time (D) to rerun the task, detect the proper backup node. newlineIn the Cloud Computing (CC) environment, handling the issue of SE would be more challenging in which clusters are heterogeneous as Hadoop believes that nodes are homogeneous. newlineTo enhance BD processing performance in the case of slow-running tasks in the heterogeneous cloud-based Hadoop Cluster (HC), the proposed Speculative Execution strategy renamed Hadoop Supreme Rate Performance (Hadoop-SRP) can be utilized. For precisely and swiftly identifying stragglers the following are wielded (1) To estimate the PS in the Map/Shuffle/Reduce phase considering parameters process speed and process bandwidth, the Modified EWMA methodology is used (2) Calculate the slow tasks RT (3) calculation of slow task s back up (4) By using the BM-LOA (Behavioral Model-based Lion Optimization Algorithm), re-run the slow task on the identified node. newlineB |
Pagination: | cxiv, 107 |
URI: | http://hdl.handle.net/10603/473611 |
Appears in Departments: | Computer/IT Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 73.7 kB | Adobe PDF | View/Open |
02_prelim pages.pdf | 2.12 MB | Adobe PDF | View/Open | |
03_abstract.pdf | 179.26 kB | Adobe PDF | View/Open | |
06_contents.pdf | 197.37 kB | Adobe PDF | View/Open | |
10_annexures.pdf | 820.41 kB | Adobe PDF | View/Open | |
10_chapter1.pdf | 491.88 kB | Adobe PDF | View/Open | |
11_chapter2.pdf | 251.24 kB | Adobe PDF | View/Open | |
12_chapter3.pdf | 568.09 kB | Adobe PDF | View/Open | |
13_chapter4.pdf | 2.13 MB | Adobe PDF | View/Open | |
14_chapter5.pdf | 668.51 kB | Adobe PDF | View/Open | |
17_appendix-a.pdf | 1.07 MB | Adobe PDF | View/Open | |
80_recommendation.pdf | 137.72 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: