ANALYZING WORLD BANK DATA USING HADOOP AND APACHE PIG
Main Author: | Vandana Jha*1 & Reshu Grover2 |
---|---|
Format: | Article eJournal |
Terbitan: |
, 2019
|
Subjects: | |
Online Access: |
https://zenodo.org/record/3455862 |
Daftar Isi:
- In today’s world the growth of digital data has led to the requirement for new technologies providing solutions for managing and handling that huge data. Earlier technology used were capable of handling a different variety of data like structured one also referred as well organised data. The RDBMS systems are designed in a way to handle only that variety of data. But in today’s word with the growth of digital data it requires technology solutions capable of handling and managing different variety of data such as un-structured and semi-structured one. Few such solutions have been proposed recently like Hadoop and Apache Pig or Hive. So the purpose of this work is, using technology like such for handling varied data make an analysis on different available datasets chosen at random so as to draw or make some diverse market decisions. The datasets selected are WDI (i.e. World development indicators)” and “GFDD (Global financial development database)” analysed using Hadoop(HDFS), HBase and Apache Pig.