Business Analytics
Business Analytics
Now, deciding which method to employ will depend on the business situation at
hand.
Business will need a lot of research skills or business research skills. You will
develop an analytical framework that can be implemented in your daily decision making
and help your company succeed by learning how to identify patterns, test hypothesis
and draw conclusions from population samples.
Business analytics as one of the skills companies need most in 2019. Learning
how to recognize trends, test hypotheses, and draw conclusions from population
samples, you can build an analytical framework that can be applied in your everyday
decision-making and help your organization thrive.
The term information processing refers to any time data needs to be transmitted or
processed in any way.
1. Transactional
Focus on data item processing
2. Analytical
Focus on reporting, analysis, transformation and decision support.
DIKW PYRAMID
The DIKW MODEL is used for data value extraction and information management. The
DIKW pyramid is a common method for explaining how we move from data to
information, knowledge and wisdom with a portion of behavior and decisions. It has its
origins in knowledge management.
1. Data – raw facts, value or elements. Data is conceived of as symbols or signs,
representing stimuli or signals.
2. Information – the results of organizing data that provides context and meaning.
3. Knowledge – information that provides insights thus making it useful and
actionable. Knowledge is valuable and actionable because it offers insights. It is
a fluid blend of frame experience, principles, contextual information, expert
perspective and grounded intuition that offers structure and atmosphere for
analyzing and integrating new experiences and information. Knowledge begins
and ends in the minds of those who know. It is often embedded in organizational
routines, systems, activities and norms in addition to documents and repositories.
4. Wisdom – the soundness of an action or decision with regard to the application
of experience, information, knowledge, and good judgement. Wisdom adds a
value which requires the mental function that we call judgement. The ethical and
aesthetic values that this implies are inherent to the actor and are unique and
personal. It is the ability to increase effectiveness.
DIKW PYRAMID
Just like DIKW model, the tacit and explicit knowledge are both knowledge
management tools used for data value extraction and information management.
Tacit Knowledge (knowing-how): knowledge embedded in the human mind
through experience and jobs. Know-how and learning embedded within the
minds of peoples. Personal wisdom and experience, context-specific, more
difficult to extract and codify. Tacit knowledge includes insights, intuitions.
Explicit Knowledge (knowing-that): knowledge codified and digitized in books,
documents, reports, memos, etc. Documented information that can facilitate
action. Knowledge what is easily identified, articulated, shared and employed.
LESSON 2: DATA COLLECTION
WHAT IS DATA?
TYPES OF DATA
1. Quantitative Data
is any data that is in numerical form like statistics and percentages.
2. Qualitative Data
is descriptive data like color, smell, appearance and quality.
Secondary data is typically quantitative in nature and has already been
collected by another party for different purpose.
DATA COLLECTION
Depending on the project companies can now collect data from mobile devices, website
traffic, server activity and other related sources with the aid of web and analytics
software.
DATA VALIDATION
Data validation primarily helps in ensuring that the data sent to connected
applications is complete, accurate, secure and consistent.
This is achieved through data validation’s checks and rules that routinely check
for the validity of data. These rules are generally defined in a data dictionary or
are implemented through data validation software.
Accurate
Coherent and Comparable
Clarity and Accessible
Timely
DATA PRESENTATION
1. Textual - basically putting the results into a logical order. The disadvantage of
this approach is that in order to get a clear image, one must read the entire text.
2. Tabular - this is where data is presented using tables and graphs. It is a format
for presenting data in rows and columns.
3. Diagrammatic - this method of data presentation and interpretation conveys a
great deal in a limited period of time. The types of diagrammatic presentation are
geometric diagram, bar diagram, pie chart, frequency diagram and histogram.
PRIMARY SCALES
1. Nominal Scale
Used to mark variables that have no numerical significance.
Used to describe categories in which there is no specific order.
For example, green, yellow, and red are three colors that in general are not
bound by an inherent order.
2. Ordinal Scale
Ordinal scale is used to describe categories in which there is an inherent
order.
Is where order of the values is important and meaningful in ordinal scales but
the variations between them are not well understood
3. Interval Scale
Interval scale is used to convey relative magnitude information such as
temperature. The term “Interval” comes about because rulers (and rating
scales) have intervals of uniform lengths.
Numeric scales in which we know both the order and the exact differences
between the values. The classic example of an interval scale is Celsius
temperature because the difference between each value is the same.
4. Ratio
Convey information on an absolute scale.
Informs us about the order, the exact value between units and have an
absolute zero. They are the ultimate data measurement scale. They can be
used for a wide variety of descriptive and inferential statistics.
When it comes to statistical analysis, ratio scale opens up a world of
possibilities. These variables can be added, subtracted, multiplied and
separated in meaningful ways or in short, ratio.
The mode, median and mean can be used to determine the central
tendencies. Ratio scales can also be used to quantify measures of dispersion
such as standard deviation and coefficient of a variation.
WHAT IS DATABASE?
Since their start in the early 1960s, databases have evolved considerably. The
first system to store and outer data are navigational databases such as the hierarchical
database which utilized a tree-like architecture and only permitted one-to-many
interactions or a one-to-many relationship. And the network database which use more
flexible model that permitted many associations or many-to-many. These early systems
are inflexible despite their simplicity. Relational database gained popularity in the 1980s
followed by object-oriented databases in the 1990s.
Now, SQL database were created more recently in response to the rapid growth
of the internet and the demand for faster processing of unstructured data. When it
comes to how data is acquired, stored, managed and used, cloud databases and self-
driving databases are forging new ground today.
TYPES OF DATABASES
Object-Oriented Database - objects are used to represent data just as they are in
object-oriented programming using classes and objects.
Distributed Database - made up of two or more files that are stored at various
locations. The database could be spread across numerous machines in the same
physical area or across numerous networks.
Data Warehouses - specifically designed for fast query and analysis. It serves as a
central repository for data.
Textual - is basically putting the results into logical order. The disadvantage of this
approach is that in order to get a clear image, one must read the entire text.
Data Validation - primarily helps in ensuring that the data sent to connected
applications is complete, accurate, secure and consistent.
Nominal Scales - are used to mark variables that have no numerical significance.
Ratio Scale - informs us about the order, the exact value between units, and have an
absolute zero, they are the ultimate data measurement scale.
Quantitative Data - any data in numerical form, such as numbers and percentages.
Qualitative Data - secondary data and had already been obtained for a separate
reason by another group.
Tabular - is where data is presented using tables and graphs. It is a format for
presenting data in rows and columns.
Data Collection - is a method of collecting and analyzing data from variety of sources
in order to obtain a full and accurate picture of a subject.
Data - facts, values, or components are all examples of raw data. Data is thought of as
a collection of symbols or signs that reflect stimuli or signals.
Ordinal Scale - is where order of the values is important and meaningful in ordinal
scales, but the variations between them are not well understood.
Interval Scale - are numeric scales in which we know both the order and the exact
differences between the values.
WHAT IS ANALYTICS?
1. Descriptive Analytics
Tells us what happened, which describes how things have changed over
time.
2. Diagnostic Analytics
Answers the question why did it happen, which focuses on the explanation
of an events occurrence. It necessitates, hypothesizing and utilizes a large
and diverse data set.
3. Predictive Analytics
Answers the question what could happen in the future. It focuses on
incidents that are likely to happen in the near future.
4. Prescriptive Analytics
Answers the question how should we respond to those potential future
events. This indicates that there is a strategy in place.
DESCRIPTIVE ANALYTICS
Reports
Inventory
Workflow
Sales
Revenue
Social Analytics
PREDICTIVE ANALYTICS
PRESCRIPTIVE ANALYTICS
Prescriptive analytics is applied to try identify the best outcome to events, given
the parameters, and suggest decision options to best take advantage of a future
opportunity or mitigate a future risk.
Prescriptive analytics is used to determine the best outcome or events given the
parameters and to propose decision strategies for taking advantage of potential
opportunities or mitigating future risk.
This method is the third, final and most advanced stage in the business analysis
process and that calls business to action helping executives, managers and
operational employees make the best possible decisions based on the data
available to them.
Prescriptive analytics expand on what has been learned through descriptive and
BIG DATA
1. Structured Data
Defined as any data that can be stored, accessed, and processed in a
fixed format.
2. Unstructured Data
Defined as any data with an unknown form or structure.
3. Semi-Structured Data
Both types of data that can be found in semi-structured data. Semi-
structured data appears structured but it is not defined in the same way
that a table definition in a rational database management system is.
CHARACTERISTICS OF BIG DATA
Volume
The term data refers to a massive amount of information. The size of the
data is very important in determining the value of the data. Furthermore,
whether a particular data set can be considered as big data or not is
determined by the volume of data.
As a result, volume is one of the characteristics that must be considered
when dealing with big data.
Variety
The next feature of big data which is diverse. The term variety refers to
wide range of data sources and data types both structured and
unstructured.
Previously, spreadsheets and databases were the only data considered by
most applications.
Velocity
Refers to the rate at which data is generated. The true potential of the
data is determined by how quickly it is generated and processed to meet
the demands.
Big data velocity is concerned with the rate at which data is ingested from
sources such as business processes, application logs, networks and
social media sites, sensors mobile devices and so on.
The data flow is massive in and continuous.
Variability
Refers to the inconsistency that data can exhibit at times. Impending the
process of effective handling and managing data.