Please use this identifier to cite or link to this item: http://hdl.handle.net/10603/264844
Title: Efficient Computing Of Big Data Harmonization
Researcher: Jigna Ashish Patel
Guide(s): Praiyanka Sharma
Keywords: Big data,OLAP,Data harmonization,OOHI
Engineering and Technology,Engineering,---Select---
University: Gujarat Technological University
Completed Date: 2019
Abstract: By the improvement and expansion of the internet, social media, internet of things and advanced technology in the fields of healthcare, infrastructure, Agriculture, Education, Scientific fields and in Data Analytics, data generation growth augmented exponentially. In the world of exploding data, storage and speed become the burning issues. Big Data was in existence since long back but due to hype of social media usage it is well-known now. Cost-effective and innovative methodology to process information which can be used for good decision making is in demand. To manage, process and to analyze Big Data both academia and industry work together for cost effective solutions. Big Data harmonization is the process of providing a single platform to all heterogeneous data and variety of data. Extraction, transformation and loading is the essential step in the process of data warehouse. Data harmonization is the alternate name for the data warehouse to provide the common level of granularity. It is the base platform to work upon for OLAP servers and data analytics. Computing of Big Data OLAP requires lot of challenges like scaling of data, speed of processing, storage of data, query performance and lot of others. Mainly Roll up, Drill down, Slice and Dice operations are performed on data. In Big Data Era ROLAP (Relational Online Analytical Processing) or HOLAP (Hybrid Online Analytical Processing) takes more space due to costly joint operation and takes more evolution time to process query. In this thesis MOLAP (Multidimensional OLAP) is adopted. Main focus is to work up on two challenges of Big Data as storage and velocity over OLAP. To work effectively and in parallel manner to deal with volume of Big Data we implemented distributed environment. Author has proposed and implemented a technique name as OOHI (OLAP on Hadoop by Indexing) that offer simplified and efficient multidimensional model. Overall work of OOHI is divided into Data Loading Module, Data Storage Module, Dimension Encoding Module, Dimension traversal M
URI: http://hdl.handle.net/10603/264844
Appears in Departments:Computer/IT Engineering

Files in This Item:
File Description SizeFormat 
01_title page.pdfAttached File100.25 kBAdobe PDFView/Open
02_declaration.pdf123.92 kBAdobe PDFView/Open
03_certificate.pdf44.73 kBAdobe PDFView/Open
04_abstract.pdf72.12 kBAdobe PDFView/Open
05_acknowledgement.pdf101.92 kBAdobe PDFView/Open
06_table of content.pdf121.38 kBAdobe PDFView/Open
07_list of abbreviations.pdf26.75 kBAdobe PDFView/Open
08_list of figures.pdf44.54 kBAdobe PDFView/Open
09_list of tables.pdf21.25 kBAdobe PDFView/Open
10_chapter 1.pdf165.88 kBAdobe PDFView/Open
11_chapter 2.pdf352.76 kBAdobe PDFView/Open
12_chapter 3.pdf400.79 kBAdobe PDFView/Open
13_chapter 4.pdf707.24 kBAdobe PDFView/Open
14_chapter 5.pdf1.17 MBAdobe PDFView/Open
15_chapter 6.pdf161.28 kBAdobe PDFView/Open
16_references.pdf253.14 kBAdobe PDFView/Open
Show full item record


Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).

Altmetric Badge: