Unit - I Introduction To Big Data
Unit - I Introduction To Big Data
Gartner:
“Big data” is high-volume, velocity, and variety information assets that demand
cost-effective, innovative forms of information processing for enhanced insight and
decision making.”
IBM:
"Big Data is characterized by its volume, velocity, and variety – the three
Vs. But the fourth V, veracity (quality of data), is also a critical factor.“
Solution
Storage: This huge amount of data, Hadoop uses HDFS (Hadoop Distributed File System) which uses
commodity hardware to form clusters and store data in a distributed fashion. It works on Write once, read
many times principle.
Processing: Map Reduce paradigm is applied to data distributed over network to find the required
output.
Analyze: Pig, Hive can be used to analyze the data.
Cost: Hadoop is open source so the cost is no more an issue.
Modules of Hadoop
HDFS: Hadoop Distributed File System. Google published its paper GFS and on the basis of that HDFS
was developed. It states that the files will be broken into blocks and stored in nodes over the distributed
architecture.
Yarn: Yet another Resource Negotiator is used for job scheduling and manage the cluster.
Map Reduce: This is a framework which helps Java programs to do the parallel computation on data
using key value pair. The Map task takes input data and converts it into a data set which can be computed
in Key value pair. The output of Map task is consumed by reduce task and then the out of reducer gives
the desired result.
Hadoop Common: These Java libraries are used to start Hadoop and are used by other Hadoop modules.
History of Hadoop
The Hadoop was started by Doug Cutting and Mike Cafarella in 2002. Its origin was the Google File System paper, published by Google.
Advantages of Hadoop
Fast: In HDFS the data distributed over the cluster and are mapped which helps in faster retrieval. Even
the tools to process the data are often on the same servers, thus reducing the processing time. It is able to
process terabytes of data in minutes and Peta bytes in hours.
Scalable: Hadoop cluster can be extended by just adding nodes in the cluster.
Cost Effective: Hadoop is open source and uses commodity hardware to store data so it really cost
effective as compared to traditional relational database management system.
Resilient to failure: HDFS has the property with which it can replicate data over the network, so if one
node is down or some other network failure happens, then Hadoop takes the other copy of data and use it.
Normally, data are replicated thrice but the replication factor is configurable.
Hadoop Architecture:
Hadoop Distributed File System
This architecture consist of a single NameNode performs the role of master, and multiple DataNodes
performs the role of a slave.
Both NameNode and DataNode are capable enough to run on commodity machines. The Java language is
used to develop HDFS. So any machine that supports Java language can easily run the NameNode and
DataNode software.
Namenode
Namenode and Datanode
•NameNode
o It is a single master server exist in the HDFS cluster.
o As it is a single node, it may become the reason of single point failure.
o It manages the file system namespace by executing an operation like the opening, renaming and
closing the files.
o It simplifies the architecture of the system.
•DataNode
o The HDFS cluster contains multiple DataNodes.
o Each DataNode contains multiple data blocks.
o These data blocks are used to store data.
o It is the responsibility of DataNode to read and write requests from the file system's clients.
o It performs block creation, deletion, and replication upon instruction from the NameNode.
•Job Tracker
o The role of Job Tracker is to accept the MapReduce jobs from client and process the data by using
NameNode.
o In response, NameNode provides metadata to Job Tracker.
•Task Tracker
o It works as a slave node for Job Tracker.
o It receives task and code from Job Tracker and applies that code on the file. This process can also be
called as a Mapper.
Big Data is different from traditional
databases
• While traditional data is based on a centralized database architecture, big
data uses a distributed architecture. Computation is distributed among
several computers in a network. This makes big data far more scalable than
traditional data, in addition to delivering better performance and cost
benefits.
• In 2014, Data Science Central, Kirk Born has defined big data in 10 V's
i.e. Volume, Variety, Velocity, Veracity, Validity, Value, Variability, Venue,
Vocabulary, and Vagueness
Year Event
2006
o Hadoop introduced.
o Yahoo deploys 300 machines and within this year reaches 600 machines.
2007
o Yahoo runs 2 clusters of 1000 machines.
2008
o YARN JIRA opened
o Hadoop becomes the fastest system to sort 1 terabyte of data on a 900 node
cluster within 209 seconds.
2009
o Yahoo runs 17 clusters of 24,000 machines.
2010
o Hadoop added the support for Kerberos.
8. Cybersecurity