Study Guide Quantitative Techniques
Study Guide Quantitative Techniques
BACHELORS OF COMMERCE
(GENERIC)
STUDY GUIDE
2021
1
Quantitative Techniques – STUDY GUIDE Damelin©
2
Quantitative Techniques – STUDY GUIDE Damelin©
1 Table of Contents
1 About DAMELIN .............................................................................................................................. 6
2 Our Teaching and Learning Methodology ...................................................................................... 6
2.1 Icons ........................................................................................................................................ 8
3 Introduction to the Module .......................................................................................................... 11
3.1 Module Information.............................................................................................................. 11
3.2 Module Purpose .................................................................................................................... 11
The purpose of this module is to instil critical thinking and analytical mind-set in making decisions
in a business and management settings. This will give learners the ability to logically analyse sets of
related issues and be able to come up with an informed decision. ................................................. 11
3.3 Outcomes .............................................................................................................................. 11
3.4 Assessment ........................................................................................................................... 12
3.5 Planning Your Studies / Resources Required for this Module: ............................................. 13
4 Prescribed Reading ....................................................................................................................... 13
4.1 Prescribed Book .................................................................................................................... 13
4.2 Recommended Articles ......................................................................................................... 13
4.3 Recommended Multimedia .................................................................................................. 13
5 Module Pacing .............................................................................................................................. 14
5.1 WEEK 1: STATISTICS IN MANAGEMENT ................................................................................ 17
5.1.1 Introduction .................................................................................................................. 17
5.1.2 Statistics in Management.............................................................................................. 17
5.1.3 The terminology of Statistics ........................................................................................ 19
5.1.4 Components of Statistics .............................................................................................. 20
5.1.5 Statistical Applications in Management ....................................................................... 21
5.1.6 Statistics and Computers .............................................................................................. 21
5.1.7 Data and Data Quality ................................................................................................... 21
5.1.8 Data Types..................................................................................................................... 22
5.1.9 Data Sources ................................................................................................................. 23
5.1.10 Self-Assessment ............................................................................................................ 25
5.2 WEEK 2: SUMMARISING DATA: SUMMARY TABLES AND GRAPHS ....................................... 26
5.2.1 Introduction .................................................................................................................. 26
5.2.2 Summarising Categorical Data ...................................................................................... 26
5.2.3 Summarising Numeric Data .......................................................................................... 30
5.2.4 Self-Assessment ............................................................................................................ 33
3
Quantitative Techniques – STUDY GUIDE Damelin©
4
Quantitative Techniques – STUDY GUIDE Damelin©
5
Quantitative Techniques – STUDY GUIDE Damelin©
1 About DAMELIN
Damelin knows that you have dreams and ambitions. You’re thinking about the future, and how the
next chapter of your life is going to play out. Living the career, you’ve always dreamed of takes some
planning and a little bit of elbow grease, but the good news is that Damelin will be there with you
every step of the way.
We’ve been helping young people to turn their dreams into reality for over 70 years, so rest assured,
you have our support.
As South Africa’s premier education institution, we’re dedicated to giving you the education
experience you need and have proven our commitment in this regard with a legacy of academic
excellence that’s produced over 500 000 world – class graduates! Damelin alumni are redefining
industry in fields ranging from Media to Accounting and Business, from Community Service to Sound
Engineering. We invite you to join this storied legacy and write your own chapter in Damelin’s history
of excellence in achievement.
A Higher Education and Training (HET) qualification provides you with the necessary step in the right
direction towards excellence in education and professional development.
• A learning-centred approach is one in which not only lecturers and students, but all
sections and activities of the institution work together in establishing a learning
community that promotes a deepening of insight and a broadening of perspective with
regard to learning and the application thereof.
• An outcomes-oriented approach implies that the following categories of outcomes are
embodied in the academic programmes:
• Culminating outcomes that are generic with specific reference to the critical cross-field
outcomes including problem identification and problem-solving, co-operation, self-
organisation and self-management, research skills, communication skills,
entrepreneurship and the application of science and technology.
• Empowering outcomes that are specific, i.e. the context specific competencies students
must master within specific learning areas and at specific levels before they exit or move
to a next level.
• Discrete outcomes of community service learning to cultivate discipline-appropriate
competencies.
Damelin actively strives to promote a research culture within which a critical-analytical approach and
competencies can be developed in students at undergraduate level. Damelin accepts that students’
learning is influenced by a number of factors, including their previous educational experience, their
cultural background, their perceptions of particular learning tasks and assessments, as well as
discipline contexts.
6
Quantitative Techniques – STUDY GUIDE Damelin©
Students learn better when they are actively engaged in their learning rather than when they are
passive recipients of transmitted information and/or knowledge. A learning-oriented culture that
acknowledges individual student learning styles and diversity and focuses on active learning and
student engagement, with the objective of achieving deep learning outcomes and preparing students
for lifelong learning, is seen as the ideal. These principles are supported through the use of an engaged
learning approach that involves interactive, reflective, cooperative, experiential, creative or
constructive learning, as well as conceptual learning via online-based tools.
• Well-designed and active learning tasks or opportunities to encourage a deep rather than
a surface approach to learning.
• Content integration that entails the construction, contextualization and application of
knowledge, principles and theories rather than the memorisation and reproduction of
information.
• Learning that involves students building knowledge by constructing meaning for
themselves.
• The ability to apply what has been learnt in one context to another context or problem.
• Knowledge acquisition at a higher level that requires self-insight, self-regulation and self-
evaluation during the learning process.
• Collaborative learning in which students work together to reach a shared goal and
contribute to one another’s learning at a distance.
• Community service learning that leads to collaborative and mutual acquisition of
competencies in order to ensure cross cultural interaction and societal development.
• Provision of resources such as information technology and digital library facilities of a high
quality to support an engaged teaching-learning approach.
• A commitment to give effect teaching-learning in innovative ways and the fostering of
digital literacy.
• Establishing a culture of learning as an overarching and cohesive factor within institutional
diversity.
• Teaching and learning that reflect the reality of diversity.
• Taking multi culturalism into account in a responsible manner that seeks to foster an
appreciation of diversity, build mutual respect and promote cross-cultural learning
experiences that encourage students to display insight into and appreciation of
differences.
7
Quantitative Techniques – STUDY GUIDE Damelin©
2.1 Icons
The icons below act as markers, that will help you make your way through the study guide.
Additional Information
All supplementary and recommended learning resources
Announcements
Important announcements made via myClass
Assessments
Continuous and Summative Assessments
Audio Material
Audio recordings and podcasts
Calculator
Activities that require calculation and equation base solutions
Case Study
Working examples of concepts and practices
Chat
A live chat with your Online Academic Tutor
Discussion Forum
Topic to be explored in the weekly discussion forum
Glossary
Learning activity centered on building a module glossary
8
Quantitative Techniques – STUDY GUIDE Damelin©
Group Assignment
Assignments to be completed with peers
Help
Instructions on how to receive academic support and guidance
Individual Assignment
Assignments to be completed individually
Lesson Material
Learning content in myClass as per the units below
Module Information
Important information regarding your module like outcomes, credits,
assessment, and textbooks
Module Welcome
A welcome to the module in myClass to introduce you to the module and
important module information
Outcomes
Learning outcomes you will meet at the end of a section or module
Survey
A poll, feedback form or survey to complete
Practice
Indicates an activity for you to practice what you’ve learnt
Lesson/Virtual Class
Virtual Class links available via myClass
9
Quantitative Techniques – STUDY GUIDE Damelin©
Quote
A thought, quote or important statement from a thought leader in the
specialist field
Reading
Prescribed reading material and module textbooks
Revision
Questions and activities that will support your module revision
Self-Assessment Quiz
Weekly quizzes to complete to self-measure if you have a complete
understanding of the lesson material
Lesson Material
Indicates sections of learning material in myClass
Thinking Point
A question, problem or example posed to you for deeper thinking,
interrogation, and reflection
Time
The allocated time required per week, unit and module related to the module
credit structure as per your factsheet
Video
Additional videos, video tutorials, desktop capture/screen recording and
other audiovisual supplementary material
Vocabulary
Important words and their definitions that aid the development of your
specialist vocabulary
10
Quantitative Techniques – STUDY GUIDE Damelin©
This course is a Business Statistics in nature and its part of every management education programme
offered today by academic institutions and business schools. Statistics provides evidence-based
information which makes it an important decision support tool in management.
Although students are encouraged to use this guide, it must be used in conjunction with other
prescribed and recommended text.
NQF Level 7
Credits 10
3.3 Outcomes
At the end of this module, you should be able to:
• Describe the role of Statistics in management decision making and the importance of data in
statistical analysis.
• Describe the meaning of and be able to calculate the mean, confidence intervals for the mean,
standard deviation, standard error, median, interquartile range, and mode.
• Summarise tables (pivot tables) and graphs providing a broad overview of the profile of random
variables, identifying the location, spread, and shape of the data.
• Understanding the basic concepts of probability to help a manager to understand and use
probabilities in decision making.
• Describe and make use of probability distributions that occur most often in management
situations, that describe patterns of outcomes for both discrete as well as continuous events
• Review the different methods of sampling and the concept of the sampling distribution.
• Describe the concept of interval estimation
• Describe hypothesis testing and construct the null and an alternate hypothesis.
11
Quantitative Techniques – STUDY GUIDE Damelin©
3.4 Assessment
You will be required to complete both formative and summative assessment activities.
Formative assessment:
These are activities you will do as you make your way through the course. They are designed to help
you learn about the concepts, theories, and models in this module. This could be through case studies,
practice activities, self-check activities, study group / online forum discussions and think points.
You may also be asked to blog / post your responses online.
Summative assessment:
These are activities you will do as you make your way through the course. They are designed to help
you learn about the concepts, theories, and models in this module. This could be through case studies,
practice activities, self-check activities, study group / online forum discussions and think points.
You may also be asked to blog / post your responses online.
You are required to do two individual assignments, online multiple-choice questions, and an online
exam.
Mark allocation
These are activities you will do as you make your way through the course. They are designed to help
you learn about the concepts, theories, and models in this module. This could be through case studies,
practice activities, self-check activities, study group / online forum discussions and think points.
You may also be asked to blog / post your responses online.
12
Quantitative Techniques – STUDY GUIDE Damelin©
TOTAL 100%
4 Prescribed Reading
13
Quantitative Techniques – STUDY GUIDE Damelin©
5 Module Pacing
Week Topics Study Guide Unit Textbook Chapter
Number Number
1 STATISTICS IN MANAGEMENT 1 1
2 SUMMARISING DATA: SUMMARY TABLES 2 2
AND GRAPHS
3 DESCRIBING DATA: NUMERIC DESCRIPTIVE
3 3
STATISTICS
4 BASIC PROBABILITY CONCEPTS 4 4
5 PROBABILITY DISTRIBUTIONS 5 5
6 CONFIDENCE INTERVAL ESTIMATION 6 7
7 HYPOTHESES TESTS –SINGLE POPULATION 7 8
(PROPORTIONS & MEANS)
8 SIMPLE LINEAR REGRESSION AND 8 9
CORRELATION ANALYSIS
9 TIME SERIES ANALYSIS: A FORECASTING 9 10
TOOL
10 REVISION 1 1 To 2
11 REVISION 2 3 To 5
12 REVISION 3 6&8
13 REVISION 4 9 & 10
NAME OF TOPIC FOR THE WEEK AS PER THIS GUIDE AND REFLECTIVE OF LMS
Each Unit should be thought of as a “week of content”. If the unit is larger, it can be split over two
weeks, but we should see that a week is a capsule or episode of learning that can have “consolidating”
14
Quantitative Techniques – STUDY GUIDE Damelin©
learning activities. As such, each “unit” will be required to have a prescribed amount of learning
activities and engagements. These are to be embedded within each week and not to be listed at the
end of the unit. PLEASE SEE THE EXAMPLE OF UNIT CONTENT.
Video Content At least ONE video resource to be in each subsection of content. This
is be embedded within the content at the appropriate time as per the
learning design.
Podcast At least ONE podcast to be in each unit.
The podcast must be seen as supplementary to the learning content
and if a podcast is not available on the specific topics at hand, an
adjunct concept/topic can be used that will broaden the general area
knowledge of the subject matter for the student.
Podcasts should not be selected that are only available on
streaming websites that require a subscription.
Thinking Point At least ONE thinking point should be used within each subsection of
content as a way to pause the movement through content and to
provide the chance for the student to think and concretize their
learning or what they have just read. A thinking point may be a
hypothetical, a personal reflection or a question regarding the
content within a different context (application). A thinking point
must be thorough and engaging enough to draw pause and focus
from the student.
Case Studies A case study should be within each unit and can be used in any
relevant subsection of content. The case study should be robust
enough for the student to understand how to apply something or to
see how a function/tool/theory or practice may work in a real world
environment. A case study should be seen as a way for the student
to be reflected in the learning experience and as such, it is advised
that case studies are selected from local/afrocentric contexts and
illustrate our commitment to intersectionality within our teaching
and learning approach and philosophy.
Discussion Forum Each unit of study/each week will require at least ONE discussion
forum topics. This can either be embedded within a certain section
of content or it can be at the end of the unit content depending on
the requirements of the module as per the subject matter. The
discussion forum topic/question should robust and dense enough for
the student to be engaged and a reference must be made to the fact
that the Discussion Forum topic is live and available within the
module page on myClass.
Example/Practice These are to be used within each section that deals with applied
learning – the application of a process, technique, equation or
function. The example is to be used when an example of a problem
15
Quantitative Techniques – STUDY GUIDE Damelin©
16
Quantitative Techniques – STUDY GUIDE Damelin©
Reading
Wegner, T (2016). Applied business statistics methods and Excel-based applications. Juta.
Cape Town South Africa. 978-1-48511-193-1. Chapter one
5.1.1 Introduction
This week, we focuses on describing what is statistics and the role it plays in management and decision
making. The importance of data in statistics is also discussed. Some basic statistical terms and
concepts are also explained
Let us Watch!
17
Quantitative Techniques – STUDY GUIDE Damelin©
As a manager decision making is one of the most crucial aspects of the jobs. Decision are made on all
business activities such as what to sell, how to sell, how much to buy, which market to target, which
equipment to buy; whether certain types of goods are of acceptable quality; where to locate stores
so as to maximise profits; and whether girls buy more of a particular product than boys etc. Therefore,
a well-informed decision based on quality information will need to be made. Wegner (2016)
Information
In order to make sound and viable business decisions, managers need high-quality information.
Information must be relevant, adequate timeous, accurate and of easy access. Information is
organised (collected, collated, summarized, analysed and presented) data values that are meaningful
and can be used to make business decisions. Most often the information is not readily available in the
formats required by the decision makers. Wegner (2016)
Data
Date constitute of individual values, for instance, observations or measurements on an issue e.g.
R400.50, 5 days, 70 meters, strongly agree, etc. Data is readily available and carries a little useful and
usable information to decision makers. Wegner (2016)
Statistics
It is a set of mathematically-based methods and techniques which transform small or large sets of raw
(unprocessed) data into a few meaningful summary measures, that may exhibit relationships, show
patterns and trends, which then contains very useful and usable information to support sound
decision making, whether we're sorting out the day's stock quotations to make a more informed
investment decision or unscrambling the latest market research data so we can better respond to
customer needs and wants. The understanding and use of statistics empower managers to become
confident and quantitative reasoning skills that enhance decision-making capabilities and provides an
advantage over colleagues who do not possess them Black (2013).
INPUT___________PROCESS____________OUTPUT ____________BENEFIT
18
Quantitative Techniques – STUDY GUIDE Damelin©
Statistics support the decision process by strengthening the quantifiable basis from which a well-
informed decision can be made.
Random Variable is any attribute or characteristic being measured or observed. It takes different
values at each point of measurement e.g. Years of experience of an employee.
Data, these are real values or outcomes drawn from a random variable e.g. Years of experience of an
employee might be (2, 4, 1, 3, 2, 6).
Sampling Unit, this will be the item being measured, observed or counted with respect to the random
variable under study. e.g. employees.
Population represents every possible item that contains a data value (measurement or observation)
of the random variable under study. The sampling units should possess the characteristics that are
relevant to the problem. e.g. all employees of Damelin Pretoria City.
19
Quantitative Techniques – STUDY GUIDE Damelin©
Population Parameter, the actual value of a random variable in a population. It’s derived from all data
values on the random variable in the population. It is constant. e.g. about 57% of MTN employees
have more than 5 years’ experience.
The sample is a subset of items drawn from a population. e.g. employees in the Finance department
of MTN.
Sample Statistic, It is a value of a random variable derived from sample data. It is NOT constant as its
value always depends on the values included in each sample drawn.
• Descriptive Statistics – condense large volumes of data into summary measures. It seeks
to paint a picture of a management problem scenario.
• Inferential Statistics – sample findings can be generalized to the broader population. It
extends the sample findings to the actual population.
• Statistical Modelling – builds relationships between variables to make predictions. It uses
equations to explain variables and to estimate or predict values of one or more of the
variables under different management scenarios.
20
Quantitative Techniques – STUDY GUIDE Damelin©
From these few examples, it is clear that there are a wide variety use and applications of statistics in
business i.e.
• Finance
• Marketing
• Human Resources
• Operation/Logistics
• Economics
The invention of computers has opened many new opportunities for statistical analysis. A computer
allows for storage, retrieval, and transfer of large data sets. Some widely used statistical techniques,
such as multiple regression, are so tedious and cumbersome to compute manually that they were of
little practical use to researchers before computers were developed. Some statistical software
packages, include i.e. R, Minitab, SAS, and SPSS. Wegner (2016)
Case Study
The link below takes you through various statistical packages that
are used in industries. Identify three statistical packages that are of
interest to you and carry out a brief research writing what features
they entail.
https://www.youtube.com/watch?v=hHywVkLwLzg
21
Quantitative Techniques – STUDY GUIDE Damelin©
is GIGO, which stands for ‘garbage in, garbage out’. It will be of great importance to understand
influences of the quality of data needed to produce meaningful and reliable statistical results. Data is
used to plan business activities and to make business decisions Black (2013) Data should be of good
quality and the quality of data depends on three aspects:
• Type of data
• Source of data
• Data collection method
Enormous numerical data are gathered in businesses every day, representing myriad items. For
instance, numbers represent rand costs of items produced, geographical locations of retail outlets,
weights of shipments, and rankings of subordinates at yearly reviews. All this data should not be
analysed the same way statistically because the entities represented by the numbers are different. In
such cases, the business researchers need to know the level of data measurement represented by the
numbers being analysed Black (2013). Four common levels of data measurement follow.
1. Nominal
2. Ordinal
3. Interval
4. Ratio
Nominal is the lowest level of data measurement followed by ordinal, interval, and ratio. Ratio is the
highest level of data
22
Quantitative Techniques – STUDY GUIDE Damelin©
• Nominal-scaled data: used on categorical data of equal importance e.g. gender – male or female
• Ordinal-scaled data: used on categorical data where ranking is implied e.g. Shirt size – small,
medium, large.
• Interval-scaled data: It is mainly from rating scales, which are used in questionnaires to measure
respondents’ attitudes, motivations, preferences, and perceptions e.g. attitudes – poor, unsure,
good etc.
• Ratio-scaled data: used on numeric data involving direct measurement where there is an absolute
origin of zero. E.g. length of service – 27 months, 45 months.
Data, of course, can be collected from a variety of sources—from vast government agencies charged
with maintaining public records to surveys conducted among a small group of customers or
prospective clients.
23
Quantitative Techniques – STUDY GUIDE Damelin©
The Internet
In these last 20 years, Internet has become an almost limitless source of business and economic data.
With the help of powerful search engines, even the casual user has instant access to data that once
would have required hours, if not weeks, of painstaking research (Bergquist et al, 2013).
Both governmental agencies and private-sector companies gather and make available a wealth of
business and economic data.
Thinking Point
Name both governmental agencies and private sector companies that gather and make
available business and economic data.
• Internal data refers to the availability of data from within an organization, examples are. Financial,
production, human resources etc.
• External data refers to data available from outside an organization, examples are. Employee
associations, research institutions, government bodies etc.
• Primary data is data which is taken at the point at which it is generated. i.e. surveys
• Secondary data: the data is collected and processed by others for various purposes other than the
problem at hand. i.e. publications
• Observation Methods
Direct observation –by directly observing the respondent or object in action data is collected. E.g.
vehicle traffic survey
Desk Research (Abstraction) – extracting secondary data from a variety of source documents. E.g.
books, publications, newspapers etc.
• Survey Methods:
Personal interviews: involves face-to-face with a respondent during which a questionnaire gets to be
completed.
24
Quantitative Techniques – STUDY GUIDE Damelin©
Let us watch
Watch the video on the link below about ideas on data collection
and list some instances where you think a particular data collection
method is suited for.
https://www.youtube.com/watch?v=8SHnJfPQ9qc
REVISION QUESTIONS
Before we progress to the following learning unit learning unit, make sure you are able
to understand and talk through the following concepts:
• Nominal data
• Ordinal data
• Interval data
• Ratio data
• Scale of Measurement
5.1.10 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
25
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 10 hours to make your way through this study week.
Reading
5.2.1 Introduction
This week aims to give the students an understanding of the common ways in which statistical findings
are conveyed. The most commonly used means of displaying statistical results are summary tables
and graphs. Summary tables can potentially be used to summarise single random variables as well as
examine the relationship of 2 random variables. The choice of the summary table and graphic to be
used depends on the type of data that need to be displayed. Managers do benefit from the statistical
findings if the information will easily be interpreted and communicated effectively to them. Tables
and graphs convey information much more efficiently and quick than a written report. In graphs and
tables, for instance, there is much truth in the old adage ‘a picture is worth a thousand words’. While
in practice, analysts’ should most if not all the times consider the use summary tables and graphical
displays more than written texts. Profile consisting of a single random variable (e.g. most-preferred
TV channel by viewers or pattern of delivery times) or to examining the relationship between two
random variables (e.g. between gender and newspaper readership) can be easily summarised by
summary tables and graphs.
26
Quantitative Techniques – STUDY GUIDE Damelin©
Categorical summary table (one-way pivot table) shows the count/percentage of responses that
belong to each category of a categorical variable. If it is given as a count it is called absolute frequency
and if given as a percentage or fraction it is referred to as relative frequency Wegener (2016).
Data from a categorical frequency table can be displayed as a pie chart or simple bar chart.
Charts or bar graphs contain 2 or more categories along one axis and a series of bars, one for each
particular category, along with the other axis. The length of the bar would represent the magnitude
of the measure (frequency, percentages amount, money, etc.) for each category. A bar graph is
qualitative since the categories are not numerical, they may be either horizontal or vertical. The same
type of data that is used to produce a bar graph is also used to construct a pie chart. The advantage
of a bar graph over a pie chart is that for some categories close in value, it is considered easier to see
the difference in the bars of bar graph than discriminating between pie slices Black (2011).
27
Quantitative Techniques – STUDY GUIDE Damelin©
In the example below, consider the data in Table 2.2 expenses by an average student on back-to-
school spending. When constructing a bar graph from the data, the categories would be Electronics,
Clothing and Accessories, Dorm Furnishings, School Supplies, and Misc. Bars for each of these
categories are made using the rand figures given in the table. Figure 2.1 is the resulting bar graph
produced by Excel
Electronics R211.89
Clothing R134.40
Misc. R93.72
Pie Chart
A circular display of data where the area of the whole circle represents 100% of the data and slices of
the circle represents a percentage breakdown of the different sublevels is known as a pie chart. Pie
charts are widely used in business, mostly in showing things such as budget, market share, ethnic
groups and time/resource allocations categories. Since pie charts can lead to less accuracy than are
possible with other types of graphs there are, however, their use is minimized in the sciences and
28
Quantitative Techniques – STUDY GUIDE Damelin©
technology. In general, it is harder for the viewer to interpret the relative sizes of angles in a pie chart
than judging the length of rectangles in a bar chart Black (2013).
27%
Mazda
40%
Toyota
Nissan
Isuzu
13%
20%
A cross-tabulation table (two-way pivot table) shows the number/percentage of observations that
jointly belong to each combination of categories between two categorical variables. For example car
type at company X in two years 2007 and 2008.
The first categorical variable is car type with four categories (Mazda, Toyota, Nissan, and Isuzu). The
second categorical variable is a year with two categories (2007 and 2008). The results can be displayed
as shown in Table 2.3:
Data from a cross-tabulation table can be displayed as a component bar chart or a multiple bar chart.
29
Quantitative Techniques – STUDY GUIDE Damelin©
A numeric frequency table (distribution) is a summary table which groups numeric data into
intervals and reports the frequency count of numbers assigned to each interval.
• Determine data range, often is defined as the difference between the largest and smallest
numbers
• Decide on a number of classes, one rule of thumb is to select between 5 and 15 classes. If the
frequency distribution contains too few classes, the data summary may be too general to be useful.
Too many classes may result in a frequency distribution that does not aggregate the data enough
to be helpful. The final number of classes is arbitrary. The business researcher arrives at a number
by examining the range and determining the number of classes that will span the range adequately
and also be meaningful to the user
• Determine class width, an approximation of the class width can be calculated by dividing the range
by the number of classes
• Determine class limits, are selected so that no value of the data can fit into more than one class
30
Quantitative Techniques – STUDY GUIDE Damelin©
Histogram
A Histogram is a graphic display of numeric frequency distribution. One of the more widely
used types of graphs for quantitative data is the histogram. A histogram is a series of
contiguous rectangles that represent the frequency of data in given class intervals. If the class
intervals used along the horizontal axis are equal, then the heights of the rectangles represent
the frequency of values in a given class interval. If the class intervals are unequal, then the
areas of the rectangles can be used for relative comparisons of class frequencies Black (2013).
31
Quantitative Techniques – STUDY GUIDE Damelin©
REVISION QUESTIONS
32
Quantitative Techniques – STUDY GUIDE Damelin©
5.2.4 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
33
Quantitative Techniques – STUDY GUIDE Damelin©
• Calculate the five-number summary table and construct its box plot
Time It will take you 10 hours to make your way through this study week.
Reading
5.3.1 Introduction
Summary table and their graphical displays, described in study week 2 are used to communicate broad
overviews of the profiles of random variables. Managers sometimes need numerical measures
(statistics) to convey more precise information about the behaviour of random variables. This precise
communication of data is the purpose of descriptive statistical measures.
34
Quantitative Techniques – STUDY GUIDE Damelin©
One type of measure that is used to describe a set of data is the measure of central tendency.
Measures of central tendency yield information about the centre, or middle part, of a group of
numbers, observations of a random variable tend to group about some central value. The statistical
measures, which quantify where the majority of observations are concentrated, are referred to as
measures of central location. A central location statistic represents a typical value or middle data point
of a set of observation and is useful for comparing data sets Wegner (2016).
1. Arithmetic mean
2. Mode
3. Mean
1. Mean
The arithmetic mean is the average of a group of numbers and is computed by summing all numbers
and dividing by the number of numbers. Because the arithmetic mean is so widely used, most
statisticians refer to it simply as the mean. The population mean is represented by the Greek letter
mu (μ). The sample mean is represented by
Formula:
∑ 𝑥𝑖
Population Mean 𝜇= 𝑁
∑ 𝑥𝑖
Sample Mean 𝑥̅ = 𝑛
2. The Mode
The mode is the most frequently occurring value in a set of data Organizing the data into an ordered
array (an ordering of the numbers from smallest to largest) helps to locate the mode i.e. The value
with the highest frequency. It is calculated by observation Wegner (2016).
3. The Median
The median is the middle value in an ordered array of numbers. For an array with an odd number of
terms, the median is the middle number. For an array with an even number of terms, the median is
the average of the two middle numbers Wegner (2016).
STEP 2. For an odd number of terms, find the middle term of the ordered array. It is the median.
STEP 3. For an even number of terms, find the average of the middle two terms. This average is the
median.
35
Quantitative Techniques – STUDY GUIDE Damelin©
1. Mean
For ungrouped data, the mean is computed by summing the data values and dividing by the number
of values. With grouped data, the specific values are unknown. What can be used to represent the
data values? The midpoint of each class interval is used to represent all the values in a class interval.
This midpoint is weighted by the frequency of values in that class interval. The mean for grouped data
is then computed by summing the products of the class midpoint and the class frequency for each
class and dividing that sum by the total number of frequencies. The formula for the mean of grouped
data follows.
∑(𝑓𝑥)
𝑥̅ =
𝑛
x = mid-point of each class interval
n = total frequency.
2. The Mode
The mode for grouped data is the class midpoint of the modal class. The modal class is the class interval
with the greatest frequency. The formula for the mode of grouped data follows
𝑐(⌊𝑓𝑚 −𝑓𝑚−1 ⌋)
Mo = O𝑚𝑜 +
2𝑓𝑚 − 𝑓𝑚−1 − 𝑓𝑚+1
3. The Median
The median for ungrouped or raw data is the middle value of an ordered array of numbers. For
grouped data, solving for the median is considerably more complicated. The calculation of the median
for grouped data is done by using the following formula.
Median Formula:
𝑐⌊𝑛2−𝑓(<)⌋
Me = O𝑚𝑒 +
𝑓𝑚𝑒
36
Quantitative Techniques – STUDY GUIDE Damelin©
𝑐 = class width
𝑛 = sample size (number of observations)
𝑓(<) = cumulative frequency count of all intervals before the median interval
• Interquartile range
• Quartile deviation
Interquartile range
It is the difference between the highest quartile (upper quartile) and the lowest quartile (lower
quartile). i.e. 𝑰𝑸𝑹 = 𝑸𝟑 − 𝑸𝟏
Quartile deviation
• Range
• Variance
• Standard deviation
• Coefficient of variation
• Coefficient of skewness
Range
37
Quantitative Techniques – STUDY GUIDE Damelin©
has any influence at all. Considering also the data set consisting of the values 3, 7, 5, 2, 4, 5, 1000. The
range would be 998, this would give a pretty misleading sense of the dispersion involved here since
all the values but one is clustered within 5 units of each other.) The measures described next are
intended to correct for this shortcomings Black (2013).
Variance
A variance is a measure of average squared deviation. It is calculated using all the data values in the
dataset.
2
∑ 𝑓𝑖 𝑥𝑖 2 − 𝑛𝑥̅ 2
𝑠 =
𝑛−1
Standard deviation
∑(𝑥𝑖 −𝑥̅ )2
s=√
𝑛−1
∑ 𝑓𝑖 𝑥𝑖 2 − 𝑛𝑥̅ 2
2
𝑠 = √
𝑛−1
Coefficient of variation
It is used to compare variability where data sets are given in different units. The coefficient of variation
essentially is a relative comparison of a standard deviation to its mean. The coefficient of variation can
be useful in comparing standard deviations that have been computed from data with different means.
𝑆
Coefficient of variation (CV) = 𝑥̅ × 100
38
Quantitative Techniques – STUDY GUIDE Damelin©
39
Quantitative Techniques – STUDY GUIDE Damelin©
Think point
Most of the statistics presented here are drawn from studies or surveys. Let’s
say a study of laundry usage is done in 50 South Africa households that have
washers and dryers. Water measurements are taken for the number of gallons
of water used by each washing machine in completing a cycle. The following
data presented are the number of litres used by each washing machine during
the washing cycle. Summarize the data so those study findings can be
reported.
Revision Questions
5.3.6 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
40
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 12 hours to make your way through this study week.
Reading
5.4.1 Introduction
Most business decisions are made under ambiguous conditions. Probability theory provides the
underpinning for quantifying and assessing uncertainty. It is used to estimate the dependability in
making inferences from samples to populations, as well as to quantify the uncertainty of future
occurrences.
• Subjective Probability
• Objective Probability
41
Quantitative Techniques – STUDY GUIDE Damelin©
Subjective Probability
The subjective method of assigning probability is based on the feelings or insights of the person
determining the probability. Subjective probability comes from the person's intuition or intellectual.
Although not a scientific approach to probability, the subjective method often is based on the accrual
of knowledge, understanding, and experience stored and processed in the human mind. At times it is
merely a supposition. At other times, the subjective probability can potentially yield accurate
probabilities. Subjective probability can be used to exploit the background of experienced workers
and managers in decision making. It is based on an educated guess, expert belief or value judgment.
This type of probability cannot be confirmed statistically, hence it has limited use. E.g. the probability
that it will rain in Cape town tomorrow is 0.15 Wegner (2016).
Objective Probability
Formula:
𝑟
P(A) = 𝑛
e.g. A container contains 3 red balls and 2 black balls. If a ball is picked at random from the bag, what
is the probability that it is: (i) red, (ii) black?
Solution:
i P(Red) = 3/5
ii P(Black) = 2/5
Experiment
with outcomes that can be statistically analysed might include the following.
• Interviewing 10 randomly selected consumers and asking them which brand of washing powder do
they prefer
• Sampling every 100th bottle of KOO beans from an assembly line and weighing the contents
• Testing new antibiotic drugs on samples of HIV patients and measuring the patients' improvement
• Auditing every 5th account to detect any errors
• Recording the S&P 500 index on the first Monday of every month for 5 years
42
Quantitative Techniques – STUDY GUIDE Damelin©
Event
Since an event is an outcome of an experiment, the experiment expresses the possibilities of the
event. If the experiment is to sample five bottles coming off a production line, an event could be to
get one defective and four good bottles. In an experiment to roll a die, one event could be to roll an
even number and another event could be to roll a number greater than two.
The intersection of events A and B is the set of outcomes that belong to both A and B altogether. The
key word is “AND” Wegner (2016).
The union of events A and B is the set of outcomes that belong to either event A or B or both. The key
word is “OR” Wegner (2016).
Events are mutually exclusive if they cannot happen together on a single trial of a random experiment.
Two or more events are mutually exclusive events if the happening of one event precludes the
occurrence of the other event(s). This characteristic means that mutually exclusive events cannot
occur at the same time and therefore can have no intersection Wegner (2016).
NB* The probability of two mutually exclusive events taking place at the same time is zero.
Events are collectively exhaustive when the union of all possible events is equal to the sample space.
i.e. at least one of the events is certain to occur in a randomly drawn object from the sample space
Wegner (2016).
Two events A and B are statistically independent if the happening of event A has no effect on the
outcome of event B and vice-versa. Two or more events are independent events if the occurrence or
non-occurrence of one of the events does not affect the occurrence or non-occurrence of the other
43
Quantitative Techniques – STUDY GUIDE Damelin©
event(s). Certain experiments, such as rolling dice, yield independent events; each die is independent
of the other. Whether a 6 is rolled on the first die has no effect on whether a 6 is rolled on the second
die. Coin tosses always are independent of each other. The event of getting ahead on the first toss of
a coin is independent of getting ahead on the second toss. It is generally believed that certain human
characteristics are independent of other events Wegner (2016).
• Marginal Probability
• Joint Probability
• Conditional Probability
Marginal Probability
A marginal probability is the probability of only a single event A occurring. i.e. the outcome of only
one random variable Wegner (2016).
Joint Probability
A joint probability is the probability of both event A and event B occurring simultaneously on a given
trial of a random experiment Wegner (2016).
Conditional Probability
A conditional probability is the probability of one event A occurring, given information about the
occurrence of a prior event Wegner (2016).
44
Quantitative Techniques – STUDY GUIDE Damelin©
• Addition Rule
• Multiplication Rule
Addition Rule
Multiplication Rule
Factorial Notation
It is used to find the total number of different ways in which n objects of a single event can be arranged
(ordered).
Permutations
A permutation is the number of distinct ways of arranging a subset of r objects selected from a group
of n objects where the order is important. Each possible arrangement (ordering) is called a
permutation.
Formula
45
Quantitative Techniques – STUDY GUIDE Damelin©
. 𝒏!
𝒏𝑷𝒓 = (𝒏−𝒓)!
Combinations Rule
A combination is the number of diverse ways of bringing together a subset of r objects selected from
a group of n objects where the order is not important. Each separate grouping is called a combination.
Formula
. 𝒏!
𝒏𝑪𝒓 = 𝒓!(𝒏−𝒓)!
Revision Questions
5.4.9 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
46
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 12 hours to make your way through this study week.
Reading
Wegner, T (2016). Applied business statistics methods and Excel-based applications. Juta.
Cape Town South Africa. 978-1-48511-193-1. Chapter 5
5.5.1 Introduction
This week makes known to us the probability distributions. Probabilities can also be derived using
mathematical functions known as probability distributions. Probability distributions quantify the
uncertain conduct of many random variables in business practice. Probability distributions can define
patterns of outcomes for both discrete as well as continuous events. important study unit, as it lays
the foundation for most of the economic analysis in Microeconomics.
47
Quantitative Techniques – STUDY GUIDE Damelin©
The word binomial indicates any single trial of a binomial experiment consists of only two possible
outcomes. These two outcomes are categorized as success or failure. Usually, the outcome of interest
to the analyst is labelled a success.
Example
1. If a quality analyst is looking for defective products, he would contemplate finding a defective
product a success even though the company would not consider a defective product a success.
2. If analysts are studying HIV patients, the outcome of getting an HIV person in a trial of an
experiment is a success.
The other possible outcome of a trial in a binomial experiment is called a failure. The word failure is
used only in opposition to success.
Characteristics:
• There are only two, mutually exclusive and collectively exhaustive outcomes of the random
variable, success & failure.
• Each outcome has an associated probability:
• Probability of success = p and probability of failure = q
• p + q = 1 (always)
• The random variable is observed n times/trials. Each trial generates either a success or failure.
• Then trials are independent of each other. i.e. p & q are constant.
The Binomial question: What is the probability that r successes will occur in n trials of the process
under study?
A measure of central location and a measure of dispersion can be calculated for any random variable
that follows a binomial distribution using the following formulae:
Mean 𝜇=𝑛𝑝
Standard deviation: 𝜎=√𝑛𝑝(1−𝑝)
How to select p
The success outcome is always associated with the probability, p. the outcome that must be labelled
as the success outcome is identified from the binomial question.
48
Quantitative Techniques – STUDY GUIDE Damelin©
• Key words such as at least, no more than, at most, no less than, smaller than, larger than,
greater than, no greater than always imply cumulative probabilities (i.e. the summing of
individual marginal probabilities.
• The complementary rule should be considered whenever practical to reduce the number of
probability calculation
Revision question
Example
1. Serious accidents at a chemical plant are rare, and the number per month might be described by
the Poisson distribution.
2. The number of random customer arrivals per five-minute interval at a small boutique on weekday
mornings.
The Poisson distribution is every so often used to explain the number of random arrivals per some
time interval. If the number of arrivals per interval is too recurrent, the time interval can be reduced
enough so that a rare number of occurrences is expected. In the field of management science, models
used in queuing theory are usually based on the assumption that the Poisson distribution is the proper
distribution to describe random arrival rates over a period of time. In statistical quality control, the
Poisson distribution is the basis for the c control chart used to track the number of non-conformances
per item or unit.
Characteristics:
Measures the number of occurrences of a particular event of a discrete random variable.
• There is a pre-determined time, space or volume interval.
• The average number of occurrences of the event is known or can be determined.
The Poisson question: What is the probability of x occurrences of a given event being observed in a
predetermined time, space or volume interval?
49
Quantitative Techniques – STUDY GUIDE Damelin©
Where:
a = the mean number of occurrences of a given event of the random variable for a predetermined
time, space or volume interval.
e = a mathematical constant
x = number of occurrences for which the probability is required.
𝑀𝑒𝑎𝑛: 𝜇 = 𝜆
Standard deviation: 𝜎 = √𝜆
Revision question
One study by CNNMoney reported that 60% of workers have less
than $25,000 in total savings and investments (excluding the value
of their home). If this is true and if a random sample of 20 workers is
selected, what is the probability that fewer than 10 have less than
$25,000 in total savings and investments?
Example,
If a worker is assembling a product component, the time it takes to accomplish that feat could be any
value within a reasonable range such as 3 minutes 36.4218 seconds or 5 minutes 17.5169 seconds.
A list of measures for which continuous random variables might be generated would include time,
height, weight, and volume.
The following are examples of experiments that could produce continuous random variables:
1. Sampling the volume of liquid nitrogen in a storage tank
2. Determining the time between customer arrivals at a retail outlet
3. Determining the lengths of newly designed automobiles
50
Quantitative Techniques – STUDY GUIDE Damelin©
Examples
Characteristics:
𝒙−𝝁
𝒛=
𝝈
51
Quantitative Techniques – STUDY GUIDE Damelin©
Revision Questions
1. Name two regularly used discrete probability distributions.
2. Specify whether each of the subsequent random variables is discrete
or continuous:
a) The mass of cans coming off a production line
b) The number of employees in a company
c) The number of households in Gauteng that have solar heating
panels
d) The distance traveled daily by a courier service truck.
5.5.9 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
52
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 15 hours to make your way through this study week.
Reading
5.6.1 Introduction
Inferential statistics’ role is to use sample evidence to ascertain population parameters. An important
and. reliable procedure to estimate a population measure will be to use the sample statistic as a
reference point and to create intervals of values around it. This would likely cover the true population
parameter with a stated level of confidence and the procedure would be called confidence interval
estimation.
53
Quantitative Techniques – STUDY GUIDE Damelin©
sample means statistic can be used when the population standard deviation. The parameter is known
regardless of the shape of the population when the sample size is large, or even for small sample sizes
if population standard deviation is known then would imply the population is normally distributed.
Because a sample mean can be greater than or less than the population mean, z can be positive or
negative. Thus the preceding expression takes the following form:
When this expression is rewritten it would yield the confidence interval formula for estimating μ with
large sample sizes when the population standard deviation is known.
For 95% confidence, α = .0.5 and α/2 = .0.25. The value of zα/2 or z.025 is found by looking in the
standard normal table under .0.5000 = .4750. This area in the table is associated with a z value of 1.96.
Another way can be used to locate the table z value. Because the distribution is symmetric and the
intervals are equal on each side of the population mean, ½(95%), or .4750, of the area is on each side
of the mean. It would yield a z value of 1.96 for this portion of the normal curve. Thus the z value for
a 95% confidence interval is always 1.96. In other words, of all the possible values along the horizontal
axis of the diagram, 95% of them should be within a z score of 1.96 from the population mean.
54
Quantitative Techniques – STUDY GUIDE Damelin©
Think Point
55
Quantitative Techniques – STUDY GUIDE Damelin©
Confidence Interval =
𝝈
̅ ± 𝒁𝜶⁄𝟐
𝒙
√𝒏
n = sample size
Most commonly used confidence intervals are shown in the table below
(n ≥ 30). In the business world, however, sample sizes may be small. While the central limit theorem
applies only when the sample size is large, the distribution of sample means is approximately normal
even for small sizes if the population is normally distributed Thus, if it is known that the population
56
Quantitative Techniques – STUDY GUIDE Damelin©
from which the sample is being drawn is normally distributed and if σ is known, the z formulas
presented in this previously can still be used to estimate a population mean even if the sample size is
small (n < 30).
Example
Suppose a South Africa. car rental firm wants to estimate the average number of
kms travelled per day by each of its cars rented in Cape Town. A random sample
of 20 cars rented in Cape Town reveals that the sample mean travel distance per
day is 85.5 km, with a population standard deviation of 19.3 km. Compute a 99%
confidence interval to estimate μ.
Here, n = 20, 𝑥̅ = 85.5, and σ = 19.3. For a 99% level of confidence, a z value of 2.575 is obtained.
Assume that number of kms traveled per day is normally distributed in the population. The confidence
interval is
The point estimate indicates that the average number of kms travelled per day by a rental car in Cape
Town is 85.5 with a margin of error of 11.1 kms. With 99% confidence, we estimate that the population
means is somewhere between 74.4 and 96.6 kms per day.
This formula is essentially the same as the z formula, but the distribution table values are different.
Confidence interval to estimate 𝜇: Population Standard deviation unknown and the population
𝑺
normally distributed, ̅ ± 𝒕𝜶⁄𝒏−𝟏
𝒙
𝟐 √𝒏
57
Quantitative Techniques – STUDY GUIDE Damelin©
Example
Here constructs a 90% confidence interval to estimate the average amount of extra
time per week worked by a manager in the aerospace industry. He assumes that
comp time is normally distributed in the population.
The sample size is 18, so df = 17. A 90% level of confidence results in α/2 = .05 area in each tail. The
table t value is
t.05, 17 =1,740
The subscripts in the t value denote to other researchers the area in the right tail of the t distribution
(for confidence intervals α/2) and the number of degrees of freedom. The sample mean is 13.56 hours,
and the sample standard deviation is 7.80 hours. The confidence interval is computed from this
information as
The point estimate for this problem is 13.56 hours, with a margin of error of ±3.20 hours. The
researcher is 90% confident that the average amount of comp time accumulated by a manager per
week in this industry is between 10.36 and 16.76 hours. From these figures, aerospace managers could
attempt to build a reward system for such extra work or evaluate the regular 40-hour week to
determine how to use the normal work hours more effectively and thus reduce comp time.
58
Quantitative Techniques – STUDY GUIDE Damelin©
Example
The central limit theorem for sample proportions leads to the following formula
𝑝(1 − 𝑝)
𝜎𝑝 ≈ √
𝑛
𝑝(1−𝑝) 𝑝(1−𝑝)
𝑝 − 𝑧√ ≤ 𝜋 ≤ 𝑝 + 𝑧√
𝑛 𝑛
Example
59
Quantitative Techniques – STUDY GUIDE Damelin©
This interval suggests that the population proportion of telemarketing firms that use their operation
to assist order processing is somewhere between .29 and .49, based on the point estimate of .39 with
a margin of error of ±.10. This result has a 95% level of confidence
Revision Questions
5.6.9 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
60
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 12 hours to make your way through this study week.
Reading
5.7.1 Introduction
In this unit, we focus on another approach of inferential statistics, where a claim made about the true
value of a population parameter is assessed for validity. The hypothesis testing is a statistical process
to test the validity of such claims using sample evidence.
General Procedure:
61
Quantitative Techniques – STUDY GUIDE Damelin©
Step4: Comparing the sample test statistic to the rejection criteria and make a conclusion.
Example
The formula below can be used to test hypotheses about a single population mean. When σ has known
if the sample size is large (n ≥ 30) for any population and for small samples (n < 30) If x is known to be
normally distributed in the population.
62
Quantitative Techniques – STUDY GUIDE Damelin©
measurement being studied. In this part of the unit, the t-test is discussed for a single population
mean. More often, the t-test is applicable whenever the analyst is drawing a single random sample to
test the value of a population mean (μ), the population standard deviation is unknown, and the
population is normally distributed for the measurement of interest.
Example
1. Suppose a company held a 36%, or .36, share of the market
for several years. Resulting from a massive marketing effort
and improved product quality, company officials believe that
the market share increased, and they want to prove it.
2. A market researcher analyst wishes to test to determine
whether the proportion of old car purchasers who are female
has increased.
3. A financial analyst wants to test to determine whether the
proportion of companies that were profitable last year in the
average investment officer's portfolio is 0.50.
4. A quality assurance manager for a large manufacturing firm
wishes to test to determine whether the proportion of
defective items in a batch is less than 0.04.
The formula below makes possible the testing of hypotheses about the population proportion in a
manner similar to that of the formula used to test sample means
𝑝−𝜋
𝑧-𝑠𝑡𝑎𝑡 =
𝜋(1−𝜋)
√
𝑛
63
Quantitative Techniques – STUDY GUIDE Damelin©
statistic. Decisions to reject or fail to reject the null hypothesis are made using a p-value technique,
which is the probability of getting a test statistic at least as extreme as the observed test statistic
computed under the assumption that the null hypothesis is true. P-value is sometimes referred to as
the observed significance level. The p-value technique has grown in importance with the increasing
use of statistical computer packages to test hypotheses. The good thing about this method is that
most computer statistical packages yield a p-value for every analysis. The p-value defines the smallest
value of alpha 𝛼 for which the null hypothesis could be rejected.
Example
Suppose an analyst is conducting a one-tailed test with a rejection region in the upper tail, and the
analyst obtains an observed test statistic of z = 2.04 from the sample data. Using the standard normal
table, we find that the probability of randomly obtaining a z value this great or greater by chance is
0.5000− 0.4793 = 0.0207. Thus, the p-value for this problem is 0.0207. Using this information, the
analyst would reject the null hypothesis for α = 0.05 or 0.10 or any value larger than 0.0207. The
analyst would not reject the null hypothesis for any alpha value less than 0.0207 (in particular, α =
0.01, 0.001, etc.). When conducting two-tailed tests, remember that alpha is split to determine the
critical value of the test statistic. For the two-tailed test, the p-value can be compared to α/2 to reach
a statistical conclusion. If the p-value is less than α/2, the decision would be to reject the null
hypothesis.
Revision Questions
64
Quantitative Techniques – STUDY GUIDE Damelin©
5.7.7 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
65
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 12 hours to make your way through this study week.
Reading
Wegner, T (2016). Applied business statistics methods and Excel-based applications. Juta.
Cape Town South Africa. 978-1-48511-193-1. Chapter 12
5.8.1 Introduction
Business decisions most often are made through predicting the unknown values of numeric variables
using other numeric variables that may be related to it and for which values could be known. A
statistical method that quantifies the relationship between a single response variable and one or more
predictor variables is called regression analysis. This relationship, which is referred to as a statistical
model, is used for prediction purposes. Correlation analysis, on the other hand, determines the
strength of the relationships and determines which variables are useful in predicting the response
variable.
66
Quantitative Techniques – STUDY GUIDE Damelin©
is referred to as the dependent variable (y). The predictor is referred to as the independent variable,
or explanatory variable, (x). In simple linear regression analysis, only a straight-line relationship
between two variables is assessed Nonlinear relationships and regression models with more than one
independent variable can be explored by using multiple regression models which are beyond the
scope of this module.
67
Quantitative Techniques – STUDY GUIDE Damelin©
The first step in determining the equation of the regression line that passes through the sample data
is to establish the equation's form. In regression analysis, analysts use the slope-intercept equation of
a line. In statistics, the slope-intercept form of the equation of the regression line through the
population points is:
̂ = 𝒃𝟎 + 𝒃𝟏 𝒙
𝒚
Where:
To construct the equation of the regression line for a sample of data, the analyst must determine the
values for b0 and b1. This procedure is sometimes referred to as least squares analysis. Least squares
analysis is a procedure whereby a regression model is constructed by obtaining the minimum sum of
the squared errors. On the basis of this premise and calculus, a particular set of equations has been
developed to produce components of the regression model.
68
Quantitative Techniques – STUDY GUIDE Damelin©
It Is a mathematical technique that determines the values 𝑏0 and 𝑏1 , such that: the sum of the squared
deviations of data points from the fitted line is minimised
The mathematical calculations of coefficients 𝑏0 and 𝑏1 that results from the method of least squares
is as follows:
𝑛 ∑ 𝑥𝑦 − ∑ 𝑥 ∑ 𝑦
𝑏1 =
𝑛 ∑ 𝑥 2 − (∑ 𝑥)2
∑ 𝑦 − 𝑏1 ∑ 𝑥
𝑏0 =
𝑛
Interpreting the 𝒃𝟏 Coefficient
The 𝑏1 regression coefficient is the slope of the regression line. It is a marginal rate of change measure.
It is interpreted as follows: for a unit change in 𝑥, 𝑦 will change by the value of 𝑏1 .
Extrapolation
Extrapolation occurs when 𝑥-values are chosen from outside the domain to substitute into the
regression equation to estimate 𝑦.
69
Quantitative Techniques – STUDY GUIDE Damelin©
Several measures of correlation are available, the selection of which depends mostly on the level of
data being analysed. Ideally, analysts would like to solve for ρ, the population coefficient of
correlation. However, because analysts virtually always deal with sample data, this unit introduces a
widely used sample coefficient of correlation, r
The term r is a measure of the linear correlation of two variables, it ranges between −1 and +1,
representing the strength of the relationship between the variables. For r-value of +1 denotes a
perfect positive relationship between two variables. For r-value of −1 denotes a perfect negative
correlation, indicating an inverse relationship between two variables: as one variable gets increases,
the other decreases. For r-value of 0 means, no linear relationship is present between the two
variables. It measures the strength of the linear association between X and Y. Pearson’s Correlation
Coefficient (r) measures the correlation between two ratio-scaled random variables.
Formula
𝑛 ∑ 𝑥𝑦 − ∑ 𝑥 ∑ 𝑦
r=
√[𝑛 ∑ 𝑥 2 − (∑ 𝑥)2 ] × [𝑛 ∑ 𝑦2 − (∑ 𝑦)2 ]
To the economist, the cost of using something in a particular way is the benefit foregone by not using
it in the best alternative way. This is called opportunity cost. Whereas accountants and businesspeople
consider only the actual expenses incurred to produce a product, the economist measures the cost of
production as the best alternative sacrificed (or foregone) by choosing to produce a particular product.
70
Quantitative Techniques – STUDY GUIDE Damelin©
Revision Questions
1. What is regression analysis? What is correlation analysis?
5.8.6 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
71
Quantitative Techniques – STUDY GUIDE Damelin©
Time It will take you 12 hours to make your way through this study week.
Reading
5.9.1 Introduction
Data collected on a given phenomenon over a period of time at systematic intervals is known as
time-series data. Time-series forecasting methods endeavours to account for changes over time by
studying patterns, trends, or cycle, or making use of information about previous time periods to
predict the outcome for a future time period. Time-series methods include naïve methods,
averaging, smoothing, regression trend analysis, and the decomposition of the possible time-series
factors. Most data used in the statistical analysis is known as cross-sectional data, meaning that it is
gathered from sample surveys at one point in time. Conversely, data can also be collected over time.
For instance, when a business collects its daily, weekly or monthly gross revenue; or when a
household records their daily or monthly electricity usage, they are gathering a time series of data.
72
Quantitative Techniques – STUDY GUIDE Damelin©
Bond yield data portrayed in Figure 9.1. The general trend appears to move downward and comprises
two cycles. Each of the cycles passes through approximately 5 to 8 years. It is possible, although not
presented here, that seasonal periods of highs and lows within each year result in seasonal bond
yields. In addition, irregular daily variations of bond yield rates may occur but are unexplainable. Time-
series data that comprise of no trend, cyclical or seasonal effects are thought to be stationary.
Approaches used to forecast stationary data analyse only the irregular fluctuation effects.
Figure 9.2, which shows the effects of these time-series elements on data over a period of 13 years.
The long-term general direction of data is referred to as a trend. Notice that even though the data
move through upward and downward periods, the general direction or trend is increasing Cycles are
73
Quantitative Techniques – STUDY GUIDE Damelin©
patterns of highs and lows through which data move over time periods usually of more than a year.
Notice that the data in Figure 9.2 apparently move through two periods or cycles of highs and lows
over a 13-year period. Time-series data that do not extend over a long period of time may not have
enough “history” to show cyclical effects. Seasonal effects, on the other hand, are shorter cycles,
which generally occur in time periods of less than one year. Every so often seasonal effects are
measured by the month, but they may occur by a quarter or may be measured in as small a time frame
as a week or even a day. Note the seasonal effects shown in Figure 9.2 as up and down cycles, many
of which occur during a 1-year period. Irregular variations are rapid changes or “bleeps” in the data,
which ensue in even shorter time frames than seasonal effects. Irregular fluctuations can happen as
often as day to day. They are subject to momentary change and are often unexplained. Note the
irregular fluxes in the data of Figure 9.2.
y=T×C×S×I
Where:
T =trend
C =cycles
S =seasonal effects
I =irregular fluctuations
In this section of the unit, we examine statistical approaches to quantify trend and seasonal variations
only. These two components account for the most significant proportion of an actual value in a time
series. By isolating them, most of an actual time series value will be explained.
The moving average time series is a smoother series than the original time series values. It has
removed the effect of short-term fluctuations (i.e. seasonal and irregular fluctuations) from the
original observations, y, by averaging over these short-term fluctuations. The moving average value
can be seen as reflecting mainly the combined trend and cyclical movements.
74
Quantitative Techniques – STUDY GUIDE Damelin©
Moving average = T × C × S × I S × I = T × C
The moving average technique is an average that is updated or recomputed for every new time period
being considered. The most recent information is utilized in each new moving average. This advantage
is offset by the disadvantages that:
1. it is difficult to choose the optimal length of time for which to compute the moving average,
and
2. moving averages do not usually adjust for such time-series effects as trend, cycles, or
seasonality.
To determine the more optimal lengths for which to compute the moving averages, we would need
to forecast with several different average lengths and compare the errors produced by them.
5.9.7 Self-Assessment
Let us see what you have learned so far by taking this short self-assessment.
Be sure to complete the self-assessment quiz before you move on to the next
section!
75
Quantitative Techniques – STUDY GUIDE Damelin©
6 REFERENCES
Bergquist, T., Jones, S. and Freed, N (2013) Understanding Business Statistics. John Wiley & Sons
Black K (2013). Business Statistics: For Contemporary Decision Making, 7th Edition. John Wiley & Sons.
Wegner, T, (2016) Applied Business Statistics: Methods and Excel-based Applications, 4th ed. Juta.
Cape Town South Africa
76