Answer N6
Answer N6
1 INTRODUCTION
We are living in the age of technology. This h
entering the business world. First, technology
of data. Retailers collect point-of-sale data on
action occurs; credit agencies have all sorts o
obtain credit; investment companies have a li
terns of stocks, bonds, and other securities; a
nomic trends, the environment, social welfa
everything else we can imagine. It has becom
result, data are plentiful. However, as many org
is quite a challenge to analyze and make s
A second important implication of technolog
power and responsibility to analyze data and
analysis. Those entering the business world can
sis to the “quant jocks,” the technical speciali
crunching. The vast majority of employees now
2 Chapter 1 Introduction to Data Analysis and Decision Making
disposal, they have access to relevant data, and they have been trained in easy-to-use soft-
ware, particularly spreadsheet and database software. For these employees, statistics and
other quantitative methods are no longer forgotten topics they once learned in college.
Quantitative analysis is now an integral part of their daily jobs.
A large amount of data already exists and will only increase in the future. Many com-
panies already complain of swimming in a sea of data. However, enlightened companies
are seeing this expansion as a source of competitive advantage. By using quantitative
methods to uncover the information in the data and then acting on this information—again
guided by quantitative analysis—they are able to gain advantages that their less enlight-
ened competitors are not able to gain. Several pertinent examples of this follow.
■ Direct marketers analyze enormous customer databases to see which customers are
likely to respond to various products and types of promotions. Marketers can then
target different classes of customers in different ways to maximize profits—and give
their customers what the customers want.
■ Hotels and airlines also analyze enormous customer databases to see what their cus-
tomers want and are willing to pay for. By doing this, they have been able to devise
very clever pricing strategies, where not everyone pays the same price for the same
accommodations. For example, a business traveler typically makes a plane reserva-
tion closer to the time of travel than a vacationer. The airlines know this. Therefore,
they reserve seats for these business travelers and charge them a higher price (for the
same seats). The airlines profit, and the customers are happy.
■ Financial planning services have a virtually unlimited supply of data about security
prices, and they have customers with widely differing preferences for various types
of investments. Trying to find a match of investments to customers is a very chal-
lenging problem. However, customers can easily take their business elsewhere if
good decisions are not made on their behalf. Therefore, financial planners are under
extreme competitive pressure to analyze masses of data so that they can make
informed decisions for their customers.
■ We all know about the pressures U.S. manufacturing companies have faced from foreign
competition in the past couple of decades. The automobile companies, for example,
have had to change the way they produce and market automobiles to stay in business.
They have had to improve quality and cut costs by orders of magnitude. Although the
struggle continues, much of the success they have had can be attributed to data analysis
and wise decision making. Starting on the shop floor and moving up through the organi-
zation, these companies now measure almost everything they do, analyze these measure-
ments, and then act on the information from these measurements.
We talk about companies analyzing data and making decisions. However, companies don’t
really do this; people do it. And who will these people be in the future? They will be you! We
know from experience that students in all areas of business, at both the undergraduate and
graduate level, will soon be required to describe large complex data sets, run regression
analyses, make quantitative forecasts, create optimization models, and run simulations. You
are the person who will soon be analyzing data and making important decisions to help gain
your company a competitive advantage. And if you are not willing or able to do so, there will
be plenty of other technically trained people who will be more than happy to replace you.
Our goal in this book is to teach you how to use a variety of quantitative methods to
analyze data and make decisions. We plan to do so in a very hands-on way. We discuss a
number of quantitative methods and illustrate their use in a large variety of realistic busi-
ness problems. As you will see, this book includes many examples from finance, market-
ing, operations, accounting, and other areas of business. To analyze these examples, we
take advantage of the Microsoft Excel spreadsheet package, together with a number of
powerful Excel add-ins. In each example we will provide step-by-step details of the
method and its implementation in Excel.
This is not a “theory” book. It is also not a book where you can lean comfortably back
in your chair, prop your legs up on a table, and read about how other people use quantita-
tive methods. It is a “get your hands dirty” book, where you will learn best by actively fol-
lowing the examples throughout the book at your own PC. In short, you will learn by
doing. By the time you have finished, you will have acquired some very useful skills for
today’s business world.
1The fact that the uncertainty theme did not find its way into the title of this book does not detract from its impor-
tance. We just wanted to keep the title reasonably short!
Figure 1.1 shows where you will find these themes and subthemes in the remaining chap-
ters of this book. In the next few paragraphs we discuss the book’s contents in more detail.
Themes Subthemes
Figure 1.1
Themes and
Subthemes
We begin in Chapters 2 and 3 by illustrating a number of ways to summarize the infor-
mation in data sets. These include graphical and tabular summaries, as well as numerical
summary measures such as means, medians, and standard deviations. The material in these
two chapters is elementary from a mathematical point of view, but it is extremely impor-
tant. As we stated at the beginning of this chapter, organizations are now able to collect
huge amounts of raw data. The question then becomes, What does it all mean? Although
there are very sophisticated methods for analyzing data sets, some of which we cover in
later chapters, the “simple” methods in Chapters 2 and 3 are crucial for obtaining an initial
understanding of the data. Fortunately, Excel and available add-ins now make what was
once a very tedious task quite easy. For example, Excel’s pivot table tool for “slicing and
dicing” data is an analyst’s dream come true. You will be amazed at the complex analysis it
enables you to perform—with almost no effort!
After the analysis in Chapters 2 and 3, we step back for a moment in Chapter 4 to see
how we get the data we need in the first place. We know from experience that many students
and businesspeople are able to perform appropriate statistical analysis once they have the
data in a suitable form. Often the most difficult part, however, is getting the right data, in the
right form, into a software package for analysis. Therefore, in Chapter 4 we present a num-
ber of extremely useful methods for doing this within Excel. Specifically, we discuss meth-
ods for using Excel’s built-in filtering tools to perform queries on Excel data sets, for using
Microsoft Query (part of Microsoft Office) to perform queries on external databases (such
as Access) and bring the resulting data into Excel, for importing data directly into Excel
from Web sites, and for “cleansing” data sets (getting rid of “bad” data values). This chapter
provides tools that many analysts need but are usually not even aware of.
Uncertainty is a key aspect of most business problems. To deal with uncertainty, we need
a basic understanding of probability. We provide this understanding in Chapters 5 and 6.
Chapter 5 covers basic rules of probability and
cept of probability distributions. Chapter 6 foll
the most important probability distributions, th
briefly discusses the Poisson and exponential d
probability
We have found that one of the best ways to
and easier to understand is by using computer
mon theme that runs through this book, begi
chapters of the book are devoted entirely to si
tion early and often to illustrate
In Chapter 7 we apply our knowledge of pro
tainty. These types of problems—faced by all
terized by the need to make a decision now, e
demand for a product or returns from investme
rial in Chapter 7 provides a rational basis fo
illustrate do not guarantee perfect outcomes—
ently than we had expected—but they do enab
of the given circumstances. Additionally, the s
allows us, with very little extra work, to see
inputs. This is crucial because the inputs to ma
guesses. Finally, we examine the role of risk a
In Chapters 8, 9, and 10 we discuss samplin
problem is to estimate one or more characteris
time consuming to learn about the entire popu
a random sample from the population and the
the characteristics of the population. We see t
the results of various polls. We also see it in m
typically sample only a fraction of a company’
of the entire population of records from the re
company has been following acc
In Chapters 11 and 12 we discuss the extrem
which is used to study relationships between v
its generality. Every part of a business has va
regression can often be used to estimate possib
managerial accounting, regression is used to es
labor hours and production volume. In market
volume depends on advertising and other ma
used to estimate how the return of a stock de
studies, regression is used to estimate how t
assessed valuation of the house and character
square footage. Regression analysis finds per
any method in
From regression, we move to times series an
topic is particularly important for providing i
example, manufacturing companies must fore
sible decisions about quantities to order from
rants must forecast customer arrivals, sometim
so that they can staff their r
There are many approaches to forecasting,
involve regression-based methods, in which o
forecast the variable of interest, whereas othe
extrapolation method the historical patterns
6
6 Chapter 1 Introduction to Data Analysis and Decision Making
demand or customer arrivals, are studied carefully and are then “extrapolated” into the
future to obtain forecasts. A number of extrapolation methods are available. In Chapter 13
we study both regression and extrapolation methods for forecasting.
Chapters 14 and 15 are devoted to spreadsheet optimization, with emphasis on linear
programming. We assume a company must make several decisions, and there are con-
straints that limit the possible decisions. The job of the decision maker is to choose the
decisions such that all of the constraints are satisfied and an objective, such as total profit
or total cost, is optimized. The solution process consists of two steps. First, we build a
spreadsheet model that relates the decision variables to other relevant quantities by means
of logical formulas. In this first step there is no attempt to find the optimal solution; all we
want to do is relate all relevant quantities in a logical way. The second step is then to find
the optimal solution. Fortunately, Excel contains a Solver add-in that performs this step.
All we need to do is specify the objective, the decision variables, and the constraints;
Solver then uses powerful algorithms to find the optimal solution. As with regression, the
power of this approach is its generality. An enormous variety of problems can be solved by
spreadsheet optimization.
Finally, Chapters 16 and 17 illustrate a number of computer simulation models. This is
not our first exposure to simulation—it is used in a number of previous chapters to illustrate
statistical concepts—but here it is studied in its own right. As we discussed previously, most
business problems have some degree of uncertainty. The demand for a product is unknown,
future interest rates are unknown, the delivery lead time from a supplier is unknown, and so
on. Simulation allows us to build this uncertainty explicitly into spreadsheet models.
Essentially, some cells in the model contain random values with given probability distribu-
tions. Every time the spreadsheet recalculates, these random values change, which causes
“bottom-line” output cells to change as well. The trick then is to force the spreadsheet to
recalculate many times and keep track of interesting outputs. In this way we can see which
output values are most likely, and we can see best-case and worst-case results.
Spreadsheet simulations can be performed entirely with Excel’s built-in tools. However,
this can be quite tedious. Therefore, we use a spreadsheet add-in to streamline the process. In
particular, we learn how the @RISK add-in can be used to run replications of a simulation,
keep track of outputs, create useful charts, and perform sensitivity analyses. With the inher-
ent power of spreadsheets and the ease-of-use of such add-ins as @RISK, spreadsheet simu-
lation is becoming one of the most popular quantitative tools in the business world.
StatTools Add-in
Much of this book discusses basic statistical an
decision as we developed the book. A number o
the market, including Minitab, SPSS, SAS, StatG
now user-friendly Windows versions of these p
have found through our own experience that st
ages, regardless of their inherent quality, so we w
(We briefly discuss SPSS and SAS in Chapter
book.) Unfortunately, Excel’s built-in statistic
ToolPak (developed by a third party) that sh
Therefore, we developed an add-in called
StatTools is powerful, easy to use, and capab
interpretable form. We do not believe you sho
to produce some statistical output. This might
but after that it acts as a strong incentive not to
should be able to generate output quickly and
the output, and it also allows you to
A good illustration involves the construction
graphs, discussed in Chapter 2. All of these extre
forward way with Excel’s built-in tools. But by th
“dress up” the charts exactly as you want them, y
process again. StatTools does it all quickly and
2Users of the previous edition of the book will note the change from StatPro to StatTools. Palisade Corporation
has redeveloped StatPro as a commercial package under the name StatTools. The user interface has changed con-
siderably (for the better), but the statistical functionality is virtually the same.
resulting charts, but that’s up to you.) Therefore, if we advise you in a later chapter, say, to look
at several scatterplots as a prelude to a regression analysis, you can do so in a matter of seconds.
SolverTable Add-in
An important theme throughout this book is sensitivity analysis: How do outputs change
when inputs change? Typically these changes are made in spreadsheets with a data table, a
built-in Excel tool. However, data tables don’t work in optimization models, where we
would like to see how the optimal solution changes when certain inputs change. Therefore,
we include an Excel add-in called SolverTable to perform this type of sensitivity analysis.
It works almost exactly like Excel’s data tables, and it is included with this book. In
Chapters 14 and 15 we explain how to use SolverTable.
@RISK
The simulation add-in @RISK enables us to run as many replications of a spreadsheet sim-
ulation as we like. As the simulation runs, @RISK automatically keeps track of the outputs
we select, and it then displays the results in a number of tabular and graphical forms.
@RISK also enables us to perform a sensitivity analysis, so that we can see which inputs
have the most effect on the outputs. Finally, @RISK provides a number of spreadsheet func-
tions that enable us to generate random numbers from a variety of probability distributions.
PrecisionTree
The PrecisionTree add-in is used in Chapter 7 to analyze decision problems with uncer-
tainty. The primary method for performing this type of analysis is to draw a decision tree.
Decision trees are inherently graphical, and they have always been difficult to implement
in spreadsheets, which are based on rows and columns. However, PrecisionTree does this
in a very clever and intuitive way. Equally important, once the basic decision tree has been
built, it is easy to use PrecisionTree to perform a sensitivity analysis on the model inputs.
TopRank
Although we will not use the other Palisade add-ins as extensively as @RISK and
PrecisionTree, they are all worth investigating. TopRank is the most general of them. It
starts with any spreadsheet model, where a set of inputs are used, along with a number of
spreadsheet formulas, to produce an output. TopRank then performs a sensitivity analysis
to see which inputs have the largest effect on the output. For example, it might tell us
which input affects after-tax profit the most: the tax rate, the risk-free rate for investing, the
inflation rate, or the price charged by a competitor. Unlike @RISK, TopRank is used when
uncertainty is not explicitly built into a spreadsheet model. However, it considers uncer-
tainty implicitly by performing sensitivity analysis on the important model inputs.
RISKOptimizer
RISKOptimizer combines optimization with simulation. There are often times when we
want to use simulation to model some business problem, but we also want to optimize a
summary measure, such as a mean, of an output distribution. This optimization can be
performed in a trial-and-error fashion, where we try a few values of the decision vari-
able(s) and see which provides the best solution. However, RISKOptimizer provides a
more automatic (and time-intensive) optimization procedure.
BestFit
BestFit is used to determine the most appropriate probability distribution for a spreadsheet
model when we have data on some uncertain quantity. For example, a simulation might
model each week’s demand for a product as a random variable. What probability distribu-
tion should we use for weekly demand: the well-known normal distribution or possibly
some skewed distribution? If we have historical data on weekly demands for the product,
we can feed them into BestFit and let it recommend the distribution that best fits the data.
This is a very useful tool in real applications. Instead of guessing a distribution that we
think might be relevant, we can let BestFit point us to a distribution that fits historical data
well. We discuss BestFit briefly in Chapter 6.
RISKview
Finally, RISKview is a drawing tool that complements @RISK. A number of probability dis-
tributions are available in @RISK and can be used in simulations. Each has an associated
@RISK function, such as RiskNormal, RiskBinomial, and so on. Before selecting any of
these distributions, however, it is useful (especially for beginners) to see what these distribu-
tions look like. RISKview performs this task easily. For any selected probability distribution
(and any selected parameters of this distribution), it creates a graph of the distribution, and it
allows us to find probabilities for the distribution in a completely intuitive, graphical manner.
We use RISKview in Chapter 16 to help learn about potential input probability distributions
for simulation models.
Palisade Corporation Finally, RISKview is a drawing tool that complements @RISK. A number of probability dis-
originally marketed tributions are available in @RISK and can be used in simulations. Each has an associated
BestFit and RISKview @RISK function, such as RiskNormal, RiskBinomial, and so on. Before selecting any of
as separate products. these distributions, however, it is useful (especially for beginners) to see what these distribu-
Although they still exist tions look like. RISKview performs this task easily. For any selected probability distribution
as separate products, (and any selected parameters of this distribution), it creates a graph of the distribution, and it
their functionality is now allows us to find probabilities for the distribution in a completely intuitive, graphical manner.
included in @RISK. We use RISKview in Chapter 16 to help learn about potential input probability distributions
for simulation models.
Software Guide
Figure 1.2 provides a guide to where these various add-ins appear throughout the book. We
don’t show Excel explicitly in this figure for the simple reason that Excel is used exten-
sively in all chapters.
With Excel and the add-ins included in this book, you have a wealth of software at
your disposal. The examples and step-by-step instructions throughout this book will help
you to become a power user of this software. Admittedly, this takes plenty of practice and
a willingness to experiment, but it is certainly within your grasp. When you are finished,
we will not be surprised if you rate “improved software skills” as the most valuable thing
you have learned from this book.
Developer Add-In
Figure 1.2
Software Guide StatTools
@RISK
RISKview
1 A SAMPLING OF EXAMPLES
Perhaps the best way to illustrate what you will be learning in this book is to preview a few
examples from later chapters. Our intention here is not to teach you any methods; that will
come later. We only want to indicate the types of problems you will learn how to solve.
Each example below is numbered as in the chapter where it appears.
EXAMPLE 4
he Spring Mills Company produces and distributes a wide variety of manufactured
goods. Because of its variety, it has a large number of customers. The company classi-
fies these customers as small, medium, and large, depending on the volume of business
each does with Spring Mills. Recently, Spring Mills has noticed a problem with its
accounts receivable. It is not getting paid back by its customers in as timely a manner as it
would like. This obviously costs Spring Mills money. If a customer delays a payment of
$300 for 20 days, say, then the company loses potential interest on this amount. The com-
pany has gathered data on 280 customer accounts. For each of these accounts, the data set
lists three variables: Size, the size of the customer (coded 1 for small, 2 for medium, 3 for
large); Days, the number of days since the customer was billed; and Amount, the amount
the customer owes. What information can we obtain from these data?
Solution
It is always a good idea to get a rough sense of the data first. We do this by calculating sev-
eral summary measures for Days and Amount, a histogram of Amount, and a scatterplot of
Amount versus Days. The next logical step is to see whether the different customer sizes
have any effect on either Days, Amount, or the relationship between Days and Amount.
There is obviously a lot going on here. We point out the following: (1) there are far fewer
large customers than small or medium customers; (2) the large customers tend to owe
considerably more than small or medium customers; (3) the small customers do not tend
to be as long overdue as the medium or large customers; and (4) there is no relationship
between Days and Amount for the small customers, but there is a definite positive relation-
ship between these variables for the medium and large customers. If Spring Mills really
wants to decrease its receivables, it might want to target the medium-size customer group,
from which it is losing the most interest. Or it could target the large customers because
they owe the most on average. The most appropriate action depends on the cost and effec-
tiveness of targeting any particular customer group. However, the analysis presented here
gives the company a much better picture of what’s currently going on.
This example from Chapter 3 is a typical example of trying to make sense out of a large
data set. Spring Mills has 280 observations on each of three variables. By realistic stan-
dards, this is not a large data set, but it still presents a challenge. We examine the data from
a number of angles and present several tables and charts. For example, the scatterplots in
Figures 1.3 through 1.5 clearly indicate that there is a positive relationship between the
amount owed and the number of days since billing for the medium- and large-size cus-
tomers, but that no such relationship exists for the small-size customers. As we will see,
graphs such as these are very easy to construct in Excel, regardless of the size of the data set.
EXAMPLE 7
ciTools Incorporated, a company that specializes in scientific instruments, has been
invited to make a bid on a government contract. The contract calls for a specific num-
ber of these instruments to be delivered during the coming year. The bids must be sealed
(so that no company knows what the others are bidding), and the low bid wins the contract.
SciTools estimates that it will cost $5000 to prepare a bid and $95,000 to supply the instru-
ments if it wins the contract. On the basis of past contracts of this type, SciTools believes
that the possible low bids from the competition, if there is any competition, and the associ-
ated probabilities are those shown in Table 1.1. In addition, SciTools believes there is a
30% chance that there will be no competing bids.
Solution
This is a typical example of decision making
SciTools has to make decisions now (whether
knowing what the competition is going to do.
outcome, but it can make a rational decision in light of the uncertainty it faces. We will see
how decision trees, produced easily with the PrecisionTree add-in to Excel, not only lay
out all of the elements of the problem in a logical manner but also indicate the best solu-
tion. The completed tree for this problem is in Figure 1.6, which indicates that SciTools
should indeed prepare a bid, for the amount $115,000.
EXAMPLE 10
n auditor wants to determine the proportion of invoices that contain price errors—that
is, prices that do not agree with those on an authorized price list. He checks 93 ran-
domly sampled invoices and finds that two of them include price errors. What can he con-
clude, in terms of a 95% one-sided confidence interval, about the proportion of all invoices
with price errors?
Solution
This is an important application of statistical in
try to determine what is true about a populatio
by examining a relatively small sample from
limit so that he is 95% confident that the ove
Figure 1.7
Analysis of Auditing
Example
EXAMPLE 11
he Bendrix Company manufactures various types o
ager of the factory wants to get a better understand
head costs include supervision, indirect labor, supp
depreciation, and a number of miscellaneous items
tion, insurance, utilities, and janitorial and mainten
costs are “fixed” in the sense that they do not vary
being done, whereas others are “variable” and do v
The fixed overhead costs tend to come from the su
neous categories, whereas the variable overhead co
supplies, payroll taxes, and overtime premiums ca
a clear line between the fixed and variable overhea
The Bendrix manager has tracked total overhead c
help “explain” these, he has also collected data on
amount of work done at the factory. These variable
■ MachHrs:
■ ProdR
The first of these is a direct measure of the amoun
second, we note that Bendrix manufactures parts in
sponds to a production run. Once a production run
for the next production run. During this setup there
machinery is reconfigured for the part type schedu
Therefore, the manager believes both of these vari
ways) for variations in overhead costs. Do scatterp
Solution
This is a typical regression example, in a cost-acco
see what type of relationship, if any, there is betwe
tory variables: number of machine hours and numb
requested appear in Figures 1.8 and 1.9. They do i
tionship between overhead and the two explanator
EXAMPLE 13
he file PCDevices.xls contains quarterly sales data (in millions of dollars) for a chip-
manufacturing firm from the beginning of 1990 through the end of 2004. Are the com-
pany’s sales growing exponentially through this entire period?
Solution
This example illustrates a regression-based trend curve, one of several possible forecasting
techniques for a time series variable. A time series graph of the company’s quarterly sales
appears in Figure 1.11. It indicates that sales have been increasing steadily at an increasing
rate. This is basically what an exponential trend curve implies. To estimate this curve we
use regression analysis to obtain the following equation for predicted quarterly sales as a
function of time:
Predicted Sales 5 6
This equation implies that the company’s sales are increasing by approximately 6.6%
per quarter during this period, which translates to an annual percentage increase of about
29%! As we see in Chapter 13, this is the typical approach used in forecasting. We look at
a time series graph to discover trends or other patterns in historical data and then use one
of a variety of techniques to fit the observed patterns and extrapolate them into the future.
Figure 1.11
Time Series Graph
of Quarterly Sales
at a PC Chip
Manufacturer
EXAMPLE 16
t the present time, the beginning of year 1, the Barney-Jones Investment Corporation
has $100,000 to invest for the next 4 years. There are five possible investments,
labeled A through E. The timing of cash outflows and cash inflows for these investments is
somewhat irregular. For example, to take part in investment A, cash must be invested at the
beginning of year 1, and for every dollar invested, there are returns of $0.50 and $1.00 at
the beginnings of years 2 and 3. Similar information for the other investments are as fol-
lows, where all returns are per-dollar invested:
■ Investment B: Invest at the beginning of year 2, receive returns of $0.50 and $1.00 a
the beginnings of years 3 and 4.
■ Investment C: Invest at the beginning of year 1, receive return of $1.20 at the begin-
ning of year 2.
■ Investment D: Invest at the beginning of year 4, receive return of $1.90 at the begin
ning of year 5.
■ Investment E: Invest at the beginning of year 3, receive return of $1.50 at the begin-
ning of year 4.
We assume that any amounts can be invested in these strategies and that the returns are the
same for each dollar invested. However, to create a diversified portfolio, Barney-Jones
decides to limit the amount put into any investment to $75,000. The company wants an
investment strategy that maximizes the amount of cash on hand at the beginning of year 5.
At the beginning of any year, it can invest only cash on hand, which includes returns from
previous investments. Any cash not invested in any year can be put in a short-term money
market account that earns 3% annually.
Solution
This is one of many optimization examples w
situation is that a company such as Barney-Jo
certain constraints, that optimize some objectiv
the amounts to invest, subject to some constrain
now. Our job is to formulate a spreadsheet model, similar to the one shown in Figure 1.12,
that relates the various elements of the problem.
The investment amounts in row 26 are the decision variables, called “changing cells”
in Excel’s terminology. When we formulate the model, we can enter any values in these
changing cells; we do not need to guess “good” values. Then we turn it over to Excel’s
Solver add-in. The Solver uses a powerful algorithm to find the optimal values in the
changing cells—that is, the values that optimize the objective while satisfying the con-
straints. The values shown in Figure 1.12 are actually the optimal values. They imply that
Barney-Jones can end with final cash of $286,792 by investing as indicated in row 26.
Solution
This is a typical example of computer simulation. We make a number of assumptions,
build a spreadsheet model around these assumptions, explicitly incorporate uncer-
tainty into some of the cells, and see how this uncertainty affects “bottom-line” out-
puts. The simulation model appears in Figure 1.13. Several cells in this model are
random, including all of the numerical values in rows 21 through 24. Therefore, the
numbers you see in this figure represent just one possible scenario of how market
shares might evolve through time. By generating new random values, we see different
scenarios.
Our job is to build the logic and randomness into the spreadsheet model. Then we can
use Excel’s built-in tools or an add-in such as @RISK to replicate the model and keep track
of selected outputs. A typical result from @RISK appears in Figure 1.14. It shows a time
series graph of how our market share might evolve, given a certain strategy of blitzing by
our company and theirs.
Figure 1.14 Summary Chart of Our Market Share for One Set of Strategies
1 MODELING AND MODELS
We have already used the term model several times in this chapter. In fact, we have shown
several spreadsheet models in the previous section. Models and the modeling process are
key elements throughout this book, so we explain them in more detail in this section.3
A model is an abstraction of a real problem. A model tries to capture the essence and
key features of the problem without getting bogged down in relatively unimportant details.
There are different types of models, and, depending on an analyst’s preferences and skills,
each can be a valuable aid in solving a real problem. We describe three types of models
here: (1) graphical models, (2) algebraic models, and (3) spreadsheet models.
Figure 1.15
Influence Diagram
for Souvenir
Example
3Management scientists tend to use the terms model and modeling more than statisticians. However, many tradi-
tional statistics topics such as regression analysis and forecasting are clearly applications of modeling.
This particular influence diagram is for a company that is trying to decide how many
souvenirs to order for the upcoming Olympics. The essence of the problem is that the com-
pany will order a certain supply, customers will request a certain demand, and the combi-
nation of supply and demand will yield a certain payoff for the company. The diagram
indicates fairly intuitively what affects what. As it stands, the diagram does not provide
enough quantitative details to enable us to “solve” the company’s problem. But this is usu-
ally not the purpose of a graphical model. Instead, its purpose is usually to show the impor-
tant elements of a problem and how they are related. For complex problems this can be
very helpful and enlightening information for management.
max ^pjxj
j51
0 # xj # uj,
Here xj is the amount of product j produced, uj is an upper limit on the amount of product j
that can be produced, pj is the unit profit margin for product j, aij is the amount of resource i
consumed by each unit of product j, bi is the amount of resource i available, n is the number
of products, and m is the number of scarce resources. This algebraic model states very con-
cisely that we should maximize total profit [expression (1.1)], subject to consuming no
more of the resources than is available [inequalities (1.2)], and all production quantities
should be between 0 and the upper limits [inequalities (1.3)].
Algebraic models such as this appeal to mathematically trained analysts. They are
concise, they spell out exactly which data are required (we would need to estimate the uj’s,
the pj’s, the aij’s, and the bi’s from company data), they scale well (a problem with
500 products and 100 resource constraints is just as easy to state as one with only 5 prod-
ucts and 3 resource constraints), and many software packages accept algebraic models in
essentially the same form as shown here, so that no “translation” is required. Indeed, alge-
braic models were the preferred type of model for years—and still are by many analysts.
Their main drawback is that they require an ability to work with abstract mathematical
symbols. Some people have this ability, but many perfectly intelligent people do not.
spreadsheets. If you enter a formula incorrectly, it is often immediately obvious (from error
messages or unrealistic numbers) that you have made an error, which you can then go back
and fix. Algebraic models provide no such immediate feedback.
A specific comparison might help at this point. We already saw a general algebraic
model of the product mix problem. Figure 1.16, taken from Chapter 14, illustrates a
spreadsheet model for a specific example of this problem. The spreadsheet model should
be fairly self-explanatory. All quantities in shaded cells are inputs to the model, the quanti-
ties in row 16 are the decision variables (they correspond to the xj’s in the algebraic model),
and all other quantities are created through appropriate Excel formulas. To indicate con-
straints, we enter inequality signs in appropriate cells.
2. Collect and summarize data. This crucial step in the process is often the most
tedious. All organizations keep track of various data on their operations, but these d
are often not in the form an analyst requires. They are also typically scattered in dif-
ferent places throughout the organization, in all kinds of different formats. Therefor
one of the first jobs of an analyst is to gather exactly the right data and summarize th
data appropriately—as we discuss in detail in Chapters 2 and 3—for use in the model.
Collecting the data typically requires asking questions of key people (such as the
accountants) throughout the organization, studying existing organizational databases,
and performing time-consuming observational studies of the organization’s processes.
In short, it entails a lot of leg work.
3. Formulate a model. This is the step we emphasize, especially in the latter chapters
of the book. The form of the model varies from one situation to another. It could be a
graphical model, an algebraic model, or a spreadsheet model. The key is that the
model should capture the key elements of the business problem in such a way that it
is understandable by all parties involved. This latter requirement is why we favor
spreadsheet models, especially when they are well designed and well documented.
4. Verify the model. Here the analyst tries to determine whether the model developed
in the previous step is an accurate representation of reality. A first step in determining
how well the model fits reality is to check whether the model is valid for the current
situation. This verification can take several forms. For example, the analyst could use
the model with the company’s current values of the input parameters. If the model’s
outputs are then in line with the outputs currently observed by the company, the ana-
lyst has at least shown that the model can duplicate the current situation.
A second way to verify a model is to enter a number of input parameters (even if
they are not the company’s current inputs) and see whether the outputs from the
model are reasonable. One common approach is to use extreme values of the inputs
to see whether the outputs behave as they should. If they do, then we have another
piece of evidence that the model is reasonable.
If certain inputs are entered in the model, and the model’s outputs are not as expected,
there could be two causes. First, the model could simply be a poor representation of
reality. In this case it is up to the analyst to refine the model until it provides reason-
ably accurate predictions. The second possible cause is that the model is fine but our
intuition is not very good. In this case the fault lies with us, not the model.
A typical example of faulty intuition occurs with random sequences of 0’s and 1’s,
such as might occur with successive flips of a fair coin. Most people expect that heads
and tails will alternate and that there will be very few sequences of, say, four or more
heads (or tails) in a row. However, a perfectly accurate simulation model of these flips
will show, contrary to what most people expect, that fairly long runs of heads or tails
are not at all uncommon. In fact, one or two long runs should be expected if there are
enough flips.
The fact that outcomes sometimes defy intuition is an important reason why models
are important. These models prove that our ability to predict outcomes in complex
environments is often not very good.
5. Select one or more suitable decisions. Many, but not all, models are decision mod-
els. For any specific decisions, the model indicates the amount of profit obtained, the
amount of cost incurred, the level of risk, and so on. If we believe the model is work-
ing correctly, as discussed in step 4, then we can use the model to see which deci-
sions produce the best outputs.
6. Present the results to the organization. In a classroom setting you are typically fin-
ished when you have developed a model that correctly solves a particular problem. In
the business world a correct model, even a useful one, is not always enough. An ana-
lyst typically has to “sell” the model to management. Unfortunately, the people in
management are sometimes not as well trained in quantitative methods as the analyst,
so they are not always inclined to trust complex models.
In addition, the model will probably have to be updated over time, either because of
changing conditions or because the company sees more potential uses for the model
as it gains experience using it. This presents one of the greatest challenges for a
model developer, namely, the ability to develop a model that can be modified as the
need arises. Keep this in mind as you develop models throughout this book. Always
try to make them as general as possible.
2 CONCLUSION
In this chapter we tried to convince you that the skills in this book are important for you to
know as you enter the business world. The methods we discuss are no longer the sole
province of the “quant jocks.” By having a PC on your desk that is loaded with powerful
software, you incur a responsibility to use this software to solve business problems. We
have described the types of problems you will learn to solve in this book, along with the
software you will use to solve them. We also discussed the modeling process, a theme that
runs throughout this book. Now it’s time for you to get started!
the age of technology. This has two important implications for everyone
ness world. First, technology has made it possible to collect huge amounts
s collect point-of-sale data on products and customers every time a trans-
redit agencies have all sorts of data on people who have or would like to
vestment companies have a limitless supply of data on the historical pat-
bonds, and other securities; and government agencies have data on eco-
the environment, social welfare, consumer product safety, and virtually
e we can imagine. It has become relatively easy to collect the data. As a
entiful. However, as many organizations are now beginning to discover, it
allenge to analyze and make sense of all the data they have collected.
tant implication of technology is that it has given many more people the
onsibility to analyze data and make decisions on the basis of quantitative
ntering the business world can no longer pass all of the quantitative analy-
jocks,” the technical specialists who have traditionally done the number
ast majority of employees now have a desktop or laptop computer at their
However, companies don’t
ture? They will be you! We
oth the undergraduate and
data sets, run regression
, and run simulations. You
rtant decisions to help gain
g or able to do so, there will
han happy to replace you.
quantitative methods to
nds-on way. We discuss a
e variety of realistic busi-
les from finance, market-
lyze these examples, we
gether with a number of
1 Introduction 3
ked with quantitative methods and examples, probably more than can be
single course. Therefore, we purposely intend to keep this introductory
hat you can get on with the analysis. Nevertheless, it is useful to introduce
will be learning and the tools you will be using. In this section we provide
e methods covered in this book and the software that is used to implement
e next section we preview some of the examples we cover in much more
apters. Finally, we present a brief discussion of models and the modeling
mary purpose at this point is to stimulate your interest in what is to follow.
parate fields: statistics and
analysis, whereas man-
d decision making. In the
d, sometimes widely.
However, from a user’s
in accomplishing what the
anagement science”
on of useful quantitative
s decisions. In addition,
sing a single package,
ds of statistics and man-
mportantly, their combina-
usiness problems.
in the title: data analysis
ach of these themes has sub-
and the search for
echniques for problems with
and structured sensitivity
nty and modelinguncertainty
these themes and subthemes.
s in data, we must deal with
we must deal with uncer-
we must deal with uncertainty,
nt in Chapter 4 to see
xperience that many students
alysis once they have the
getting the right data, in the
hapter 4 we present a num-
ecifically, we discuss meth-
n Excel data sets, for using
n external databases (such
data directly into Excel
d” data values). This chapter
aware of.
with uncertainty, we need
ding in Chapters 5 and 6.
features in more
ool that enables you to
l has many useful tools, but
d powerful of all. We
roduce Excel’s RAND
n all spreadsheet simula-
alth of software at
ghout this book will help
kes plenty of practice and
When you are finished,
s the most valuable thing
6, 16–17
1 A Sampling of Examples 11
omers
Customers
tomers
1 A Sampling of Examples
an equation relating
output such as that shown
edicted overhead as a
1 A Sampling of Examples 17
■
estment Corporation
le investments,
ws for these investments is
ash must be invested at the
ns of $0.50 and $1.00 at
r investments are as fol-
1 A Sampling of Examples 19
of Strategies
■
a general algebraic
pter 14, illustrates a
preadsheet model should
ts to the model, the quanti-
xj’s in the algebraic model),
mulas. To indicate con-
E F G H I
4
0
<=
4 Totals
$0 $50,200
$0 $32,000
$0 $3,000
$0 $6,000
$0 $9,200
del such as the one in
ts algebraic counterpart,
ously, they must be correct.
ct syntax, the correct cell
quite a challenge.
o be used in the busi-
d. Otherwise, no one other
d) will be able to under-
readsheets is their
Modeling Process
overall modeling process
ep process, as discussed
mple, the analysis of sur-
ision making), without the