Chapter 2 EmTe
Chapter 2 EmTe
2
What is data science?
Data science is a multi-disciplinary field that uses
scientific methods, processes, algorithms, and systems
to extract knowledge and insights from structured,
semi-structured and unstructured data.
Data science is much more than simply analyzing
data. It offers a range of roles and requires a range of
skills.
3
What are data and information?
Data can be defined as a representation of facts,
concepts, or instructions in a formalized manner, which
should be suitable for communication, interpretation, or
processing, by human or electronic machines.
It can be described as unprocessed facts and figures.
It is represented with the help of characters such as
alphabets (A-Z, a-z)
digits (0-9) or
special characters (+, -, /, *, <,>, =, etc.)
4
What are data and information?...
Whereas information is the processed data on which
decisions and actions are based.
It is data that has been processed into a form that is
meaningful to the recipient and is of real or perceived
value in the current or the prospective action or decision
of recipient.
Furtherer more, information is interpreted data; created
from organized, structured, and processed data in a
particular context.
5
2.1.2. Data Processing Cycle
Data processing is the re-structuring or re-ordering of
data by people or machines to increase their usefulness
and add values for a particular purpose.
Data processing consists of the following basic steps -
input, processing, and output. These three steps
constitute the data processing cycle.
6
Data Processing Cycle
Input − in this step, the input data is prepared in some convenient form
for processing. The form will depend on the processing machine. For
example, when electronic computers are used, the input data can be
recorded on any one of the several types of storage medium, such as hard
disk, CD, flash disk and so on.
Processing − in this step, the input data is changed to produce data in a
more useful form. For example, interest can be calculated on deposit to a
bank, or a summary of sales for the month can be calculated from the
sales orders.
Output − at this stage, the result of the proceeding processing step is
collected. The particular form of the output data depends on the use of the
data. For example, output data may be payroll for employees.
7
2.3 Data types and their representation
Data types can be described from diverse perspectives.
In computer science and computer programming, for
instance, a data type is simply an attribute of data that
tells the compiler or interpreter how the programmer
intends to use the data.
8
2.3.1. Data types from Computer programming perspective
9
Data types from Computer programming perspective
A data type makes the values that expression, such as a
variable or a function, might take.
This data type defines the operations that can be done
on the data, the meaning of the data, and the way values
of that type can be stored.
10
2.3.2. Data types from Data Analytics perspective
From a data analytics point of view, it is important to
understand that there are three common types of data
types or structures: Structured, Semi-structured, and
Unstructured data types. Fig. 2.2 below describes the
three types of data and metadata.
11
Structured Data
Structured data is data that adheres to a pre-defined
data model and is therefore straightforward to analyze.
Structured data conforms to a tabular format with a
relationship between the different rows and columns.
Common examples of structured data are Excel files or
SQL databases.
Each of these has structured rows and columns that
can be sorted.
12
Semi-structured Data
Semi-structured data is a form of structured data that
does not conform with the formal structure of data
models associated with relational databases or other
forms of data tables, but nonetheless, contains tags or
other markers to separate semantic elements and enforce
hierarchies of records and fields within the data.
Therefore, it is also known as a self-describing
structure. Examples of semi-structured data include
JSON and XML are forms of semi-structured data.
13
Unstructured Data
Unstructured data is information that either does not
have a predefined data model or is not organized in a
pre-defined manner.
Unstructured information is typically text-heavy but
may contain data such as dates, numbers, and facts as
well. This results in irregularities and ambiguities that
make it difficult to understand using traditional
programs as compared to data stored in structured
databases. Common examples of unstructured data
include audio, video files or No-SQL databases.
14
Metadata – Data about Data
The last category of data type is metadata. From a technical point
of view, this is not a separate data structure, but it is one of the
most important elements for Big Data analysis and big data
solutions.
Metadata is data about data. It provides additional information
about a specific set of data.
In a set of photographs, for example, metadata could describe
when and where the photos were taken.
The metadata then provides fields for dates and locations which,
by themselves, can be considered structured data. Because of this
reason, metadata is frequently used by Big Data solutions for
initial analysis.
15
2.4. Data value Chain
The Data Value Chain is introduced to describe the information
flow within a big data system as a series of steps needed to
generate value and useful insights from data. The Big Data Value
Chain identifies the following key high-level activities:
17
2.4.2. Data Analysis
It is concerned with making the raw data acquired
amenable to use in decision-making as well as domain-
specific usage.
Data analysis involves exploring, transforming, and
modeling data with the goal of highlighting relevant
data, synthesizing and extracting useful hidden
information with high potential from a business point of
view.
Related areas include data mining, business
intelligence, and machine learning.
18
2.4.3. Data Curation
It is the active management of data over its life cycle to ensure it
meets the necessary data quality requirements for its effective
usage.
Data curation processes can be categorized into different
activities such as content creation, selection, classification,
transformation, validation, and preservation.
Data curation is performed by expert curators that are responsible
for improving the accessibility and quality of data.
Data curators (also known as scientific curators or data
annotators) hold the responsibility of ensuring that data are
trustworthy, discoverable, accessible, reusable and fit their
purpose. A key trend for the duration of big data utilizes
community and crowdsourcing approaches.
19
2.4.4. Data Storage
It is the persistence and management of data in a scalable way that
satisfies the needs of applications that require fast access to the data.
Relational Database Management Systems (RDBMS) have been the
main, and almost unique, a solution to the storage paradigm for
nearly 40 years.
However, the ACID (Atomicity, Consistency, Isolation, and Durability)
properties that guarantee database transactions lack flexibility with
regard to schema changes and the performance and fault tolerance
when data volumes and complexity grow, making them unsuitable
for big data scenarios.
No SQL technologies have been designed with the scalability goal
in mind and present a wide range of solutions based on alternative
data models.
20
2.4.5. Data Usage
It covers the data-driven business activities that need
access to data, its analysis, and the tools needed to
integrate the data analysis within the business activity.
Data usage in business decision-making can enhance
competitiveness through the reduction of costs,
increased added value, or any other parameter that can
be measured against existing performance criteria.
21
2.5. Basic concepts of big data
Big data is a blanket term for the non-traditional
strategies and technologies needed to gather, organize,
process, and gather insights from large datasets.
In this section, we will talk about big data on a
fundamental level and define common concepts you
might come across.
We will also take a high-level look at some of the
processes and technologies currently being used in this
space.
22
What Is Big Data?
Big data is characterized by 10V’s and more:
Volume:
23
Cont’d
Veracity:
Veracity refers to the reliability of the data source. Numerous
factors can contribute to the reliability of the input they provide
at a particular time in a particular situation.
Veracity is particularly important for making data-driven
decisions for businesses as reproducibility of patterns relies
heavily on the credibility of initial data inputs.
Validity
Validity pertains to the accuracy of data for its intended use.
For example, you may acquire a dataset pertaining to data
related to your subject of inquiry, increasing the task of forming
a meaningful relationship and inquiry. Registered charity data
contact lists
24
Cont’d
Visualization
With a new data visualization tool being released every month
or so, visualizing data is key to insightful results.
Value
BIG data is nothing if it cannot produce meaningful value.
Consider, again, the example of Target using a 16-year-old’s
shopping habits to predict her pregnancy. While in this case, it
violates privacy, in most other cases, it can generate
incredible customer value by bombarding them with the
specific product advertisement they require.
25
Cont’d
Volatility
Volatility refers to the time considerations placed on a
particular data set. It involves considering if data acquired a
year ago would be relevant for analysis for predictive
modeling today
Vulnerability
Big data is often about consumers. We often overlook the
potential harm in sharing our shopping data, but the reality is
that it can be used to uncover confidential information about
an individual. For instance, Target accurately predicted a
teenage girl’s pregnancy before her own parents knew it. To
avoid such consequences, it’s important to be mindful of the
information we share online.
26
Cont’d
27
What Is Big Data?
28
2.5.2. Clustered Computing and Hadoop Ecosystem
Clustered Computing
Because of the qualities of big data, individual computers are often
inadequate for handling the data at most stages.
To better address the high storage and computational needs of big
High Availability
Easy Scalability
29
Cluster computing benefits
Resource Pooling
Combining the available storage space to hold data is a clear
benefit, but CPU and memory pooling are also extremely
important.
High Availability
Clusters can provide varying levels of fault tolerance and
availability guarantees to prevent hardware or software failures
from affecting access to data and processing.
Easy Scalability
Clusters make it easy to scale horizontally by adding additional
machines to the group.
30
Cluster computing benefits
Using clusters requires a solution for managing cluster
membership, coordinating resource sharing,and
scheduling actual work on individual nodes.
Cluster membership and resource allocation can be
handled by software like Hadoop’s YARN (which
stands for Yet Another Resource Negotiator).
The assembled computing cluster often acts as a
foundation that other software interfaces with to process
the data.
31
Hadoop and its Ecosystem
Hadoop is an open-source framework intended to make
interaction with big data easier.
It is a framework that allows for the distributed
processing of large datasets across clusters of computers
using simple programming models.
It is inspired by a technical document published by
Google. The four key characteristics of Hadoop are:
32
Hadoop and its Ecosystem
Economical: Its systems are highly economical as
ordinary computers can be used for data processing.
Reliable: It is reliable as it stores copies of the data on
different machines and is resistant to hardware failure.
Scalable: It is easily scalable both, horizontally and
vertically. A few extra nodes help in scaling up the
framework.
Flexible: It is flexible and you can store as much
structured and unstructured data as you need to and
decide to use them later.
33
Hadoop and its Ecosystem
Hadoop has an ecosystem that has evolved from its four core
components: data management, access, processing, and storage. It
is continuously growing to meet the needs of Big Data. It comprises
the following components and many others
HDFS: Hadoop Distributed File System
YARN: Yet Another Resource Negotiator
MapReduce: Programming based Data Processing
Spark: In-Memory data processing
PIG, HIVE: Query-based processing of data services
HBase: NoSQL Database
Mahout, Spark MLLib: Machine Learning algorithm libraries
Solar, Lucene: Searching and Indexing
Zookeeper: Managing cluster
Oozie: Job Scheduling
34
Hadoop and its Ecosystem
36
Big Data Life Cycle with Hadoop
37