An Introduction To Economic Modelling
An Introduction To Economic Modelling
SIMULATION
Introduction
Economists often seek to understand the nature and functioning of various types of economic
systems. In doing so, they rely on the relationships that exist among various economic variables,
and use such relationships to make predictions or forecasts of how the system will change, if
This process, however, must be guided by economic theory in order to establish such
relationships (models), to test their predictability or forecasting power, and if possible, revise
such models in the light of new information. This is the concept of economic modeling, which
simply put, is the process of creating models to study some phenomena of interest.
through creation of appropriate models, being guided by economic theory, and utilising relevant
data and the appropriate analytical requirement, i.e., skills and tools.’
Economic modeling can be done either at the macro- level that involves modeling an aspect for a
larger aggregate, such as models for a whole country or region, or they could also be done at the
micro- level that involves modeling a particular aspect of interest relevant to the analysts.
Today, economic modeling has evolved and has found a significant part in decision making, both
in the public and private sectors, academia and even in international organizations. Ministries,
organizations such as the IMF, African Development Bank, the World Bank and many other
software have been developed in this regard, for commercial purposes, including EVIEWS,
GAMS, DEA, GRETL, among many others. However, some software are specialized for
particular modeling exercises only, while others are specialized for simulation. Computers will
remain an invaluable part to the applied econometrician or modeler, both in their training and
professional work. Economic modeling and computer simulation is therefore at the heart of
Objectives
In not more than 500 words summarize the features of a model. Post your contribution in the
provided journal.
Learning Activity 1.3: Assignment
The journal in activity 1.2. and Assignment in activity 1.3 will be graded
Topic Resources
URL link
https://www.youtube.com/watch?v=qiQhAyuH2P4
Please go through the video link above to attempt the assignment above.
Topic 1 Notes
Economic Models
Definition of a model
A model refers to anything that is used to represent reality, or the working of a particular system
The reason why we abstract from reality is due to the fact that the real world has a lot of
complexities, interactions and surprises, and thus, we cannot claim to be able to study the
Instead, we seek to capture only the salient or most important features of the system (the
essentials) of the real situation which it represents that are of particular interest to us, and then
make an assumption that the other relevant factors will remain constant for the period under
Another reason why we abstract from reality is due to the recognition that not all the inter-
relationships in the variables of interest are of equal importance in the understanding of the
particular economic phenomena under study. Thus, we only focus attention on those
Therefore, models do not describe the true economic world, but instead, they abstract from the
truth or reality. A model can be constructed at different levels of aggregation and sophistication.
Hence, the degree of abstraction depends on the PURPOSE for which the model is constructed,
the EXPERIENCE of the MODELER in terms of their knowledge of available techniques and
their expertise in constructing and implementing models, and the NATURE of the problem that
However, abstraction from reality does not imply unrealism! It implies the simplification of
reality and the beginning of the understanding of the great complexity of the real world.
Any model has to satisfy two key features, i.e., simplicity and realism.
1. SIMPLICITY
- Thus, a model should be similar to, but simpler than the real system it represents, i.e., a
2. REALISM
- Despite the fact a model simplifies a real system, this does not justify unrealism.
- Thus, the modeler still has to incorporate some reasoning and judgement in developing
and implementing models, and not view the model as a perfect substitute for reasoning.
- Realism thus implies that a model may be relevant one situation (region or time), and be
logical abstraction.
Since both simplicity and realism are vital features for a model, the recommended practice in
modeling is to increase the complexity of a model iteratively, based on the need, the situation
(region or time), and the purpose; all aimed at enhancing model realism.
(i) The model is relatively simple and easy to understand and use,
(ii) The model is guided by the relevant economic theory in its development,
(iii) The model aspires to as much as possible represent the system – and this can be done
(iv) The assumptions behind the model are specified, and should be valid as well,
(v) The model outlines its structure, i.e., variables used and types of equations.
(vi) Any model is time- and location- bound – any model can only be true at a particular
time and location since model features may change due to change in fundamentals
Model validity is an important issue in modeling. It involves comparing model output with
system output. Thus, the ease and closeness with which a model reflects the real system is a key
By model validity, were interested in the significance of the variables used in the model, the
validity of the model itself to effectively address the problem being examined, and also the
validity of the estimation techniques, whether Ordinary Least Squares (OLS), Generalized Least
(iii) The extent of information it provides, - a valid model should provide as much
information as possible, but this depends on the purpose or use to which the model is
put.
(iv) Its generality – that is, the number or range of cases to which it applies,
(v) Its simplicity. A model that is simple can be replicated elsewhere and can have a long
There is no general agreement regarding which of the above criteria is more important.
- Milton Friedman proposes that predictive performance of a model is its most important
- Paul Samuelson is of the view that realism of assumptions and power of the model in
explaining the behavior of the system or agents in the model is the most important
attribute to a model.
Realism of assumptions and explanatory power are important attributes of a model if the
- Most economists have the dominant position that the generality and simplicity of the
model are first-hand criteria: i.e., that a model’s validity is determined by the purpose or
use to which it is put, should provide as much information as possible, and should at least
Although models are diverse, and therefore have varied elements, any typical economic model
i. Endogenous variable
v. Closure rule
- The endogenous (dependent) variable is the one whose value is determined within the
model. Thus, the value of the endogenous variable is determined by the other variables in
- For example, the famous Keynesian national income model is defined as:
Y = C + I + G + NX …………………………………. (1)
is income lagged by one period and r is the real interest rate. Also
In equation (2), C is determined by the values taken by Y, C t-1 and r. Thus, C is the
endogenous variable.
- A model could have one or several endogenous variables, and this determines the level of
sophistication or detail that a model wants to have. Thus, a model is either a single-
equation model or a multiple-equation model, but this will depend on the purpose for
which the model is constructed. Thus, any economic model must clearly specify its
endogenous variable(s)
- An exogenous variable is one whose value is determined from outside the model, and
hence their values are defined elsewhere in other models, and are thus taken as given.
equation (1) above, C, I, G and NX are the exogenous variables. However, whereas C is
was endogenous in the first equation, it is now exogenous in the second equation, and so
on.
(iii) PARAMETERS
- A parameter is a scalar or constant which is either attached to a variable or it could be
stand-alone. Thus, there are two types of parameters: coefficient and intercept.
- Coefficient: - this is a scalar attached to a variable and it measures the partial derivative
of the endogenous variable with respect to the respective explanatory variables, holding
For example, in the equation: C = b0 + b1Y + b2Ct-1 + b3r, the scalars b0, b1, b2 and b3 are
the model coefficients. Thus, b1 measures the partial derivative of the dependent variable
C with respect to income Y, b2 measures the partial derivative of the dependent variable
C with respect to income Ct-1, and so on. Partial derivative is also referred to as Marginal
effect.
- Parameters are generally useful in a model since they define for us elasticities and
the function is a log-log function, then the coefficients are readily defined as elasticities.
- Intercept, on the other hand, is stand-alone. It measures the expected (mean) value of the
endogenous variable if all other factors are held constant, or the point where the function
crosses the vertical intercept. Thus, b0 in equation 2 is the intercept, and it measures the
level of consumption when all other factors are held constant – i.e., autonomous
consumption.
(iv) ASSUMPTIONS
- Assumptions are statements or expressions that must be true or valid for a model to be
reliable and predictable. However, most importantly, it is not the assumption, but the
law to be valid or true. In the consumption function, the assumption made is that the
value of the marginal propensity to consume (MPC) lies between 0 and 1, i.e., 0<MPC<1.
- Recall that Paul Samuelson holds in esteem the consistency and realism of assumptions
- Models are created not just for their own sake, but rather to help in forecasting and policy
analysis. To this end, any model must be capable of yielding solution values for the
- On the other hand, a closure rule helps to examine how equilibrium relationships are
obtained, or how a model will reach equilibrium, after a shock to the system.
- For most models, the closure rule usually consists of an identity, such as the national
income identity.
Characteristics of a Model
Models are diverse in terms of their features and specifications, but we can identify three key
For example, the production function describes the relationship between output and inputs, and it
could be linear, cob-Douglas, Leontief, constant-elasticity of substitution (CES), and so on. The
tax function is also a technical function since it shows the relationship between tax revenue and
the determinants of tax, including some institutional factors, such as reforms in administration,
will behave in response to changes in various explanatory variables or factors. For example, the
consumption function shows how consumer expenditure on consumption of goods and services
function as it shows how the level of investment relates to the real interest rate.
IDENTITY is an equation stating the equality that holds true by virtue of the definition of the
variables involved. For example, the national income identity states as follows: Y º C + I + G +
X – M. thus, the left hand side has exactly the same meaning as the right hand side, or simply
put: Income is equal to expenditure. An identity helps to determine the closure rule of the
equilibrium position.
Economic models can be classified based on: the time period, the number of variables or
equations, the purpose of the model, the type of data used in the model, and the methodology or
Based upon these criteria, we can identify the following types of models: (i) Graphical or visual
models, (ii) Mathematical Models, (iii) Empirical Models, and (iv) Simulation Models.
(i) Graphical or visual models
- These are models that depict an economy or economic system by making use of graphs or
networks.
- Consider the following data on price (P), aggregate demand before (AD 0) , aggregate
demand after (AD1), and aggregate supply (AS) of a particular good in a given economy:
AD0 95 85 75 65 55 45 35 25
AD1 105 95 85 75 65 55 45 35
Qs 15 25 35 45 55 65 75 85
P 1 2 3 4 5 6 7 8
- Enter this information into excel spreadsheet but place it column wise.
- Now, highlight AD0, AD1 and AS then go to menu, and select: INSERT/ LINE and click
on the first graph that is indicated. You will notice the graph for AD-AS Model is
- The AD-AS model illustrates the effects of changes in either aggregate demand or
aggregate supply on the equilibrium price and quantity. The diagram above demonstrates
- Apart from the AD-AS model, other examples of graphical models are: the IS-LM model,
- Although graphical models do not require heavy knowledge of mathematics, they are
models, but they suffer from the 2- or 3- dimensional problem and thus may not capture
optimization, and other concepts of mathematics. As a result, such models are also known
as algebraic models.
- A classic example of models under this type is the Keynesian national income model for
Y=C+I+G+X–M (1)
C = c0 + c1(Y – T) (2)
T = t0 + t1Y (3)
M = m0 + m1Y (4)
I = I0 (5)
G = G0 (6)
X = X0 (7)
- The solution of the Keynesian national model is obtained by substituting the behavioral
Y=C+I+G+X–M
Equilibrium Y =
- The Keynesian national income model therefore, helps us to estimate the level of
equilibrium national income that an economy should expect, if all the assumptions hold
true.
- The challenge with the Keynesian model is the accurate determination or measurement of
the actual level of national income, and hence it has not been used in measuring income.
Instead, many countries have used the value-added approach in measuring their levels of
national income.
If we assume specific values for the parameters such as: c 0 = 200, c1 = 0.8, t0 = 100, t1 =
0.5, m0 = 60, m1 = 0.2, I0 = 50, G0 = 40, and X0 = 80, then we can obtain the specific value
(i) They allow us to have more than 2 variables in a model, i.e., they allow for an n-
variable case.
method of describing how a system will behave if one or more variables change.
A 10% increase in c0 = 200 means new c0 becomes 110/100 * 200. Thus, new c0 =
This exercise would best have been done on EXCEL SPREADSHEET as follows
A B C D E F G H I J K L
1 Scenario c0 c1 t0 t1 m0 m1 I0 G0 X0 Y %D
20%
Assignment 1: perform the following scenario analysis (using both your calculator
C2*E2 + G2), then press ENTER. The value reported in that cell should be 287.5
In cell B3, enter the following formula: = 110 / 100 * B2, then press ENTER. The value
To get the value for cell K3, place your cursor in cell K2, then drag-and-drop into cell
In cell L3, enter the following formula: = (K3 – K2) / K2 * 100, then press ENTER. The
Now, proceed that way to get the values in cells: C4, K4, L4, G5, K5 and L5. What
- Empirical models or econometric models are similar to mathematical models except for
the fact that empirical models heavily rely on data. Therefore, before formulating an
- Thereafter, using appropriate statistical techniques, we can then estimate the model from
the data, therefore empirical models calls upon the researcher to have adequate
P 1 2 3 4 5 6 7 8
Formulate an empirical model from this data. We shall use the regression analysis to
- From regression analysis: the formula for obtaining the Ordinary Least Squares (OLS)
- Enter the data on Y and X on excel spreadsheet column-wise in columns A and B of the
spreadsheet.
- In column C of the spreadsheet type y. Below this cell, which is in cell C2, enter the
following formula: = A2 – 60 and press enter. Now, drag from cell C2 to cell C9.
- In column D of the spreadsheet type x. Below this cell, which is in cell D2, enter the
following formula: = B2 – 4.5 and press enter. Now, drag from cell D2 to cell D9.
- In column E of the spreadsheet type xy. Below this cell, which is in cell E2, enter the
following formula: = C2 * D2 and press enter. Now, drag from cell E2 to cell E9.
- In column F of the spreadsheet type x2. Below this cell, which is in cell F2, enter the
following formula: = D2 * D2 and press enter. Now, drag from cell F2 to cell F9.
- In cell A10, enter the formula: = SUM (A2:A9) and press enter. Now drag across to cell
F10. The values in cells A10 to F10 are the totals or summations.
- In cell A11, enter the formula: = AVERAGE (A2:A9) and press enter. Now drag across
to cell F11. The values in cells A11 to F11 are the means or averages.
- The results are now shown in the table as follows:
Y X Y X xy x2
- We can also perform a regression equation using this data on excel spreadsheet. First,
type in the data given column-wise. Go to MENU, click DATA, and then select DATA
ANALYSIS. In the dialogue box that appears, select REGRESSION and click OK. On
input Y range, highlight the data on Y. On input X range, highlight the data on X. Finally
click OK.
- Empirical models can either be linear or non-linear and the specification is usually
informed by economic theory and other econometric tests of model specification such as
or prediction of certain economic variables. For example, we can thus predict what will
happen to quantity demanded when price is 9, and so on. Such predictions are however
conclusions can be drawn from the simulation exercise, as what would actually happen in
the real world. This therefore means that, before doing any establishment or development
of an existing system which in most cases are irreversible, simulations help to mimic how
the system would otherwise be and this is less costly and can be easily reversed.
- In a simulation exercise, the model usually begins with an initial equation with values
assigned by the programmer (the baseline scenario), then certain variables are changed –
depending on the issue being examined or the issue of interest or new information
becomes available, and their outcomes or solutions are observed through computer
simulation.
- Thus, simulation models are applicable with graphical, mathematical or empirical models
- Simulation programs therefore can be done on specialized computer software meant for
that exercise.
ADVANTAGES OF SIMULATION:
(vi) Changing a real system may be costly and irreversible, but simulation helps to
(vii) With the development of modern computers and specialized software, simulation
exercises have become greatly enhanced, and though expensive in most cases,
cheaper software exists for simulation programming. For instance, excel has features
such as “what-if” and “goal seek” analysis in the toolbar, which are useful for
DISADVANTAGES OF SIMULATION
(i) Simulation does not guarantee optimality. It is not easy to tell whether the optimal
solution has been reached, or whether it is in fact the best, under the given conditions,
simulation exercises,
(iii) The process of simulation requires a lot of expertise to build and maintain, while most
software are also expensive to acquire. However, other cheaper software can be
(iv) Simulation does not always yield absolute numbers, and this limits the usefulness of
Simulation models have wide applications in day to day life, but they are especially utilized in
the following areas: Manufacturing, health, transport, defense and business process engineering.
I. MANUFACTURING
- Simulation also allows a company to compare various designs and alternative policies on
- By doing this, simulation helps to reduce cost and risk associated with errors. Simulation
- Health care systems aim at providing health care services efficiently and effectively at
minimum cost. To achieve this, health care system test different policies and drugs
- Hospitals and other health care institutions have applied simulation to determine the best
drug prescription for various ailments; to ensure the movement of patients, staff and
equipment move about in the facility without inconveniencing them or putting life at
danger.
- Therefore, simulation models aim at examining various networks or designs and policies
- Simulation models in transport systems have led to the construction of by-passes and
IV. DEFENCE
- Militaries aim at protecting their jurisdictions and repulsing any threats of the enemy as
- They therefore make use of simulation plans to understand various strategies to achieve
these objectives, and counter possible strategies that the enemy could employ.
entity either to improve service delivery, lower costs of operation, create more effective
management, or improve the customer experience. BPE also aims to ensure that the
- Examples of Business Process Engineering include: shift from line queuing to electronic
queuing; services in banking halls to agency and internet banking; paper work to
paperless transactions; online tax services (ITAX in Kenya); cash payments to credit card
provided by the computer, or those published in journal articles, without focusing on the validity
of those estimates.
However, a modeler should carefully think about and evaluate every aspect of the equations or
model, right from the underlying theory to the quality of data, before accepting the output as
valid. They also have some expectations of the results before the data collection and analysis.
Once the computer estimates have been produced, however, it is then time to evaluate the results.
The following is a checklist of some of the questions that should be asked in evaluating the
(1) Is the equation or model specification supported by sound theory i.e., - has the most
(2) How well does the estimated model fit the data i.e., – to what extent is the model output
(4) Is the estimation technique used (e.g., Ordinary Least Squares - OLS) the best estimator
(5) How well do the estimated coefficients correspond to the prior expectations by the
researcher?
(6) Are all the important variables included in the model, or are there omitted variables that
(7) Does the regression appear to be free from major econometric problems that may render
Although there are no hard and fast rules for conducting econometric research, most
The choice of dependent variable is determined by the purpose of the research, and will be
THEORETICAL MODEL
Many econometric decisions ranging from which variable to use to which functional form to
employ, are determined by the underlying theoretical model; and these will be informed by
Having identified the theoretical model, the modeler may modify the model to derive the
empirical model that will be used. The specification of the model involves choosing the
following components:
- The distribution of the stochastic error term – whether normally distributed or not
Any error in the specification is most disastrous to the validity of the estimated model.
Based on economic theory, we can expect the signs of the coefficients in the model, whether
positive or negative.
We can then compare the regression estimates actual signs against expectations, and this will
Having specified the model, the analyst then determines the sample size and then collects the
data. If the data is time series, it should be of the same frequency, e.g., monthly, quarterly,
annually, and so on. The amount of data collected should be sufficiently large to ensure enough
To inspect the data means checking if the data is consistent. For example, a scatter plot of the
data can help to reveal if there are any outliers in the data. Also, summary statistics help us to
examine mean, minimum, maximum, standard deviation, etc all of which will define the data.
Cleaning data involves replacing an incorrect number with a correct one; or where possible, drop
and having estimated the model; evaluate its quality as outlined in the checklist above.
Economic models have diverse uses depending on the specific aspects that the modeler is
interested in. in general however; economic models are utilized in the following areas:
i. Models are used to analyze behavior or investigate facts. For example, given a
consumption function, we can then investigate how a change in say tax rate or disposable
income will affect aggregate consumption. Also, how a change in bank rates will affect
ii. Models can also be used to prove or falsify economic theory. If a theory is tested by
use of a model, and is proven to be false, then such a theory can either be revised in
iii. Models are useful in evaluating and analyzing policies. Government economic
policies include fiscal policy, monetary policy, trade policy and exchange rate policy.
These policies usually produce various effects, some desirable while other
undesirable. Thus, models can help to examine various policies and to choose among
alternative policies.
iv. Models are very invaluable in measuring and forecasting economic variables, such as
economic growth, and how shocks to the system can affect the variables. This refers
have tax-revenue models, inflation models, exchange rate models, trade models, etc.
Forecasting is meant to guide any policy making unit to achieve its long term goals.
v. Models are also used in structural analysis. This refers to the use of an estimated
coefficients of the structural equations, reduced form equations, and models with
vi. Models are also useful for analyzing strategic behavior among competing economic
game theory models help to analyze such interactions and access the behavior of
agents affected by other agents or the regulator. For example, how do oil companies
compete for large market share? Or how do oil companies behave following a
vii. With the heavy toll that environmental degradation and climate change take on the
world, most models are trying to incorporate modeling for climate change and its
effects on the economy (green economy models). A famous example of this model is
viii. Models have also been constructed for national planning and stabilization purposes.
For example, the World Bank Revised Minimum Standard model (RMSM) and the
IMF’s Polak Model have for long time been used for economic planning in most
African countries. Today however, the IMF’s Financial Programming model has
taken centre stage in planning and economic stabilization in many developing
Models of the macroeconomy have gotten quite sophisticated, thanks to decades of development
and advances in computing power. Such models have also become indispensable tools for
monetary policymakers, useful both for forecasting and comparing different policy options.
Their failure to predict the recent financial crisis does not negate their use, it only points to some
areas that can be improved.
Forecasting plays a vital role in the conduct of monetary policy. Policymakers need to predict the
future direction of the economy before they can decide which policy to adopt. While, strictly
speaking, they do not necessarily need an economic model to discuss where the economy is
heading, the use of a model’s forecast has the benefit of elevating that discussion to a scientific
and systematic level. Models can be used to test different theories, for example, and they require
forecasters to clearly spell out their underlying hypotheses.
But policymakers need forecasting tools that do more than project the likely path of important
economic indicators like inflation, output, or unemployment. They need tools that can provide
them with policy guidance tools that help them determine the economic implications of
monetary-policy changes. For example, what will the economy look like under the original
monetary policy, and what will it look like after the change? For this reason, there has been an
effort over the past 40 to 50 years to develop empirical forecasting models that are able to
provide policymakers with this kind of guidance. Three broad categories of macroeconomic
models have arisen during this time, each with its own strengths and weaknesses: structural, non-
structural, and large-scale models.
Structural models are built using the fundamental principles of economic theory, often at the
expense of the model’s ability to predict key macroeconomic variables like GDP, prices, or
employment. In other words, economists who build structural models believe that they learn
more about economic processes from exploring the intricacies of economic theory than from
closely matching incoming data.
Non-structural models are primarily statistical time-series models—that is, they represent
correlations of historical data. They incorporate very little economic structure, and this fact gives
them enough flexibility to capture the force of history in the forecasts they generate. They
intentionally “fudge” theory in an effort to more closely match economic data. The lack of
economic structure makes them less useful in terms of interpreting the forecast, but at the same
time, it makes them valuable in producing unconditional forecasts. That means that they generate
the expected future paths of economic variables without imposing a path on any particular
variable. These unconditional forecasts are typically accurate if the overall monetary policy
regime does not change. Since policy regimes change infrequently, most forecasts from non-
structural models are useful.
The third category, large-scale models, is a kind of middle ground between the structural and
nonstructural models. Such models are a hybrid; they are like nonstructural models in that they
are built from many equations which describe relationships derived from empirical data. They
are like structural models in that they also use economic theory, namely to limit the complexity
of the equations. They are large, and their size brings pros and cons. One advantage is that
relationships can be selected from a huge variety of data series, making it possible to provide a
thorough description of the economic condition of interest. For instance, structural models rarely
feature variables such as “car sales,” while large-scale models often do. The main disadvantage
is their complexity, which poses some limitations to their understanding and use.
System Dynamics Modelling (SDM) is a methodology for studying and managing complex
feedback systems. It is typically used when formal analytical models do not exist, but where
system simulation can be developed by linking a number of feedback mechanisms.
The main goal is to help people make better decisions when confronted with complex, dynamic
systems. The approach provides methods and tools to model and analyzes dynamic systems.
Model results can be used to communicate essential findings to help everyone understand the
system's behavior.
If properly performed, Dynamic Modelling can reveal design flaws that may not show up readily
during the prototyping and testing phases of the product design cycle.
Econodynamics
Dynamism is a general name for a group of philosophical views concerning the nature of
matter. However different they may be in other respects, all these views agree in making matter
consist essentially of simple and indivisible units, substances, or forces.
Dynamic economics has an important place in economics because many economic theories are
based on it. For example, saving and investment theory, theory of interest, effect of time
element in price determination, etc. are based on dynamic economics.
The competition and job-to-job mobility fostered by dynamism helps power wage gains for
workers. In a churning labor market where demand for workers is high and competition is
intense, workers gain leverage and have ample opportunities to seek out better jobs and higher
wages.
Each economy functions based on a unique set of conditions and assumptions. Economic
systems can be categorized into four main types: traditional economies, command economies,
mixed economies, and market economies
Decision theory
Decision theory deals with methods for determining the optimal course of action when a number
of alternatives are available and their consequences cannot be forecast with certainty. It is
difficult to imagine a situation which does not involve such decision problems, but we shall
restrict ourselves primarily to problems occurring in business, with consequences that can be
described in dollars of profit or revenue, cost or loss. For these problems, it may be reasonable to
consider as the best alternative that which results in the highest profit or revenue, or lowest cost
or loss, on the average, in the long run. This criterion of optimality is not without shortcomings,
but it should serve as a useful guide to action in repetitive situations where the consequences are
not critical. (Another criterion of optimality, is the maximization of expected utility, provides a
and the associated probabilities for each alternative, calculating the expected monetary values of
all alternatives, and selecting the alternative with the highest expected monetary value. The
determine of the optimal alternative becomes a little more complicated when the alternatives
information about an uncertain variable. This additional information is rarely entirely accurate.
amount one would be willing to pay to acquire it should depend on the difference between the
best one expects to do with the help of this information and the best one expects to do without it.
1. Identify Decisions.
Identify the decisions that are the focus of the Next-Best-Action platform.
o From a marketing perspective these decisions usually fall into three classes:
eligibility, relevancy and prioritisation
What are the KPIs (Key Performance Indicators) surrounding that decision? Now is the
time to consider how a decision will drive a given KPI such as churn, risk, NPS, etc.
2. Describe Decisions.
Describe the decisions and document how improving them will impact the business
objectives and metrics of the business. The activities performed in this stage would guide
how simulations are performed on the strategy. For example, what would be the likely
impact of using a suitability rule which excludes propositions related to already owned
products? What would be the impact of using a different prioritisation algorithm which
considers NPS (Net Promotor Score) for service-related communications?
Specify the detailed information and knowledge required to make the decisions. What are
the inputs (human and computer) required to make the decision? Decision requirements
will now be used to express this. For example, what knowledge would be required to
define suitability rules for a cross-sell mobile international call discount proposition? The
cross-sell marketing leader would provide input to determine who should (or should not)
be suitable for the international call discount proposition.
Refine the requirements for these decisions using the graphical notation of Decision
Requirements Diagrams. In addition, identify additional decisions that need to be
described and specified.
For example, eligibility decisions could be decomposed by applicability and suitability.
Applicability defines rules which ask “Can I show this proposition?” Suitability defines
the rules which ask “Should I show this proposition?”
The cycle is repeated. We repeat the process by returning to step 1.
Our starting point is one of the oldest and most important ideas in economics, going back at least
to Adam Smith, namely the idea that a decentralized economic system is self-organizing. It is
capable of “spontaneous order,” in the sense that a globally coherent pattern of transactions can
result from purely local interactions, without the intervention of a central coordinator. Indeed,
like an anthill, a free market economy can organize transactions into patterns that are beyond the
comprehension of any of its individual participants. We would like to understand how this self-
organization takes place. Specifically, what is the process that coordinates the exchange activities
of millions of independent transactors in a decentralized economy?
The reason why these questions are critical for understanding macroeconomic policy is that an
economy’s coordination mechanism works better at some times than others. Even Smith and
Hayek recognized that the automatic workings of the decentralized economy could sometimes be
improved by collective intervention. Consider, for example, the increase in unemployment that
takes place during a deep recession. Unemployed workers who used to be employed are just as
willing and able to work as before, the fall in aggregate output that accompanies recession has
enhanced the scarcity value of the output they could potentially produce if employed, and yet the
market for their services has somehow disappeared. The coordination mechanism that had
previously allowed them and those with a taste for their output to realize their potential gains
from mutually advantageous exchange is no longer allowing them to do this, even though those
gains are if anything larger than before. Macroeconomic policy to deal with unemployment thus
amounts to fixing a mechanism that has malfunctioned, and a highly complex mechanism at that.
And attempts to fix this broken mechanism without first understanding how it is supposed to
work normally are likely to be as successful as medieval medicine was in treating bacterial
infections.
The main premise of research has been that the role of coordinating transactions in a
decentralized economy is performed, for better or worse, by a self-organizing network of
business firms that seek profit from creating and operating the markets through which others
transact. To use a phrase that Clower once coined, business firms are the visible fingers of the
invisible hand. Economics has a long tradition of regarding exchange as a do-it-yourself affair, in
which people with goods and services to sell trade directly with the ultimate demanders of those
goods and services. But a little reflection on the experience of daily life is enough to persuade
most people that exchange in a market economy is not a do-it-yourself affair. People are not like
the actors in a typical monetary search model, who when hungry go wandering aimlessly in
hopes of randomly encountering someone with surplus food. They go to a grocery store or to a
restaurant. When in search of clothing they visit a tailor or a clothing store. They lend surplus
funds through the intermediation of a bank, arrange for long-distance travel by using facilities
provided by a travel agent, and so on and so forth. Most of us also sell our labor services to an
economic entity, either a private business or a government agency (the latter of which would not
exist in a purely decentralized economy) whose primary purpose is to purchase various such
services, organize them into production, and sell the resulting output. So to understand how, and
how well, exchange activities are coordinated in a decentralized economy, the first place to look
is to this self-organizing network of firms that constitutes the institutional structure through
which we all conduct our daily business.
Now some would argue that a good economic theory involves abstraction, and that if we want to
model the transactions process of a modern economy we should maybe abstract from business
firms, by assuming that people who work for businesses trade their output directly with those
having a taste for the workers’ output. This is the stance taken by search theoretic models of
money, which typically assume that trade takes place directly between ultimate suppliers and
ultimate demanders, who meet in random non-repeated encounters without the aid of any
intermediary. Presumably the rationale for this way of looking at the transactions process is the
same as the rationale for abstracting from money in the theory of value. On the surface, what we
see is people trading goods and services for money, but the deeper underlying reality that we see
once we pierce the veil of money is that people are ultimately trading goods and services for
other goods and services, with money acting only as a device for executing these ultimate
exchanges.
The analogy between money and firms is a useful one. But as John Stuart Mill once observed,
there is nothing more insignificant than money, except when it goes wrong. By the same token,
the fact that people find it convenient to trade through shops rather than directly with one another
is perhaps of little significance for understanding the long-run structure of relative prices. But
when something goes wrong with the network of firms that people normally rely upon, then
abstracting from the existence of such firms is as unhelpful as it would be to ignore money when
trying to understand inflation.
Moreover, a good case can be made that, by recognizing that production takes place in firms but
not recognizing their role in coordinating transactions by creating and operating markets, we are
ignoring their most important activity, as measured by the value of resources they use. The value
added in Finance, Insurance, and Real Estate, for example, is typically much more than that of
the entire manufacturing sector, as is the value added in Retail and Wholesale Trade. Moreover,
much of the input to the manufacturing sector is best construed as being used up in the
production of transactions that help people realize gains from trade rather than being used up in
transforming inanimate objects. We have in mind the inputs of lawyers, sales people, those
engaged in personnel, marketing, and advertising, and so forth, all of whom are undertaking
activities whose main purpose is to facilitate transactions.
Of course, there are many serious criticisms one might make of DSGE, and many of them have
been made in the literature. The criticism we consider most important for present purposes is that
existing DSGE models, even those with imperfectly flexible prices, are built on a conceptual
foundation that pays little or no attention to the way in which economic transactions are
organized. To borrow a phrase from Jevons, they ignore the institutions and processes that make
up the mechanism of exchange.
When we examine DSGE models, looking for what else might go wrong with the market
mechanisms that coordinate economic transactions, we find that in most of them there is no such
mechanism. In models with perfect competition, the setting of prices is left to a mysterious
outside agent called the auctioneer, whose behavior is left largely unexplained. But that is just
the tip of the iceberg. There is no description of how trades are arranged. Even if we accept that
the auctioneer can provide everyone with a price vector such that the sum of desired demands
equals total supply for each tradable object, there is no account of how buyers and sellers are
matched up with one another and how the trades that people have planned will be executed.
When demand and supply are not equal, the theory offers no guidance as to who gets to trade
how much and with whom, no indication of how people learn about trading opportunities, about
who creates and maintains the shops and other facilities through which they trade, about how
bids and offers are transmitted, and so on and so forth.
The canonical model of Woodford (2003), which forms the basis of the estimated New
Keynesian DSGE models, now used in central banks around the world, makes less use of the
mysterious auctioneer, inasmuch as many prices are set by a given set of monopolistically
competitive firms who are explicitly motivated to maximize their shareholders’ wealth. So far, so
good. But, there is no account of where these price-setting monopolists come from, how they
maintain their monopolies against the threat of entry, how people decide to trade with one set of
firms rather than another, how firms manage to coordinate with their suppliers and customers,
what happens when one of them goes out of business in a recession, and so forth. Instead, all
transactors are in continuous touch with each other through the intermediation of these firms,
whose existence is merely assumed, and who take care of enough details of the transactions
process that the other people in the model are connected only through the market prices that they
take as given from the firms. As a result, there is nothing that can go wrong in the transactions
process other than some mistake in price-setting.
In essence, these New Keynesian DSGE models are providing the same diagnosis that
economists have given for generations; unemployment rises because wages and prices are slow
to adjust to shifts in demand and supply. This is the answer provided by classical economists
from Hume through Marshall. It is still the answer offered by modern Keynesian economics.
Indeed, it is now even the answer that has been finally accepted by most proponents of the real-
business-cycle school of macroeconomics, who admit the need for wage-and-price stickiness to
account for various features of the business cycle.
The problem with this time-honored tradition of blaming wage-and-price stickiness is not that
the assumption of stickiness is factually incorrect. On the contrary, the stickiness of wages and
prices is one of the most well-documented facts of macroeconomics. Instead, as Leijonhufvud
(1968) forcefully pointed out, the problem is that, first, the experience of the Great Depression in
the United States shows clearly that the downturn that started in 1929 did not come to an end
until wages and prices started to rise, that is, until the reflation that was clearly a deliberate
policy move on the part of the Roosevelt administration started to take place. If lack of wage and
price flexibility had caused the downturn, then it would have taken deflation rather than reflation
to cure the unemployment problem. Second, as Keynes argued in Chapter 19 of the General
Theory, and as Fisher had already argued in Debt Deflation Theory of Depressions, there are
many reasons for believing that wage and price flexibility would actually make fluctuations in
unemployment larger rather than smaller.
So when unemployment rises in a recession, something has gone wrong with the process by
which economic transactions become organized, something that goes beyond the mere stickiness
of wages and prices, something that we think can only be discovered by investigating simple
stylized models of economies in which trading activities take place, in and out of equilibrium, in
a self-organizing network of markets that are created and operated by profit-seeking business
firms, and by asking how, and how well, those activities are coordinated in various
circumstances. What we would like to do in this paper is to give the reader an idea of what kind
of model that research has led us to construct, and why we think this class of models provides a
more solid framework for analyzing certain policy questions than does any DSGE model
currently in use.
It turns out that this research agenda is one for which Agent-Based Computational Economics
(ACE) is particularly well suited for two main reasons. First, by endowing each agent with a set
of relatively simple adaptive behavioral rules that allow the agent to operate intelligently in an
unknown environment, an ACE model gives the system a chance to achieve some semblance of
order without giving anyone the kind of systemic knowledge that would allow him to act as a
central coordinator. A rational expectations equilibrium might or might not emerge from the
interaction of these rules. If it does, then we have discovered something about at least one
possible mechanism that produces that kind of spontaneous order whereas, if the system fails to
approximate a rational expectations equilibrium, we will have discovered something about the
conditions under which a spontaneous order is likely to require some kind of collective
intervention. The second reason for using ACE is that models of spillovers between multiple
markets that are not in supply-demand equilibrium are notoriously difficult to analyze. The
attempts by Barro and Grossman, Benassy, Malinvaud, and others in the 1970s to understand
what some called “general disequilibrium analysis,” building on the original contributions of
Clower and Patinkin, made little progress largely because the problem quickly became
analytically intractable. ACE can deal with this kind of intractability by substituting simulation
and Monte-Carlo results for unattainable analytical results.
(a) What the model is all about and uses of the model
References/Reading List
1. Neradilová, H., & Fedorko, G. (2016). The use of computer simulation methods to reach
data for economic analysis of automated logistic systems. Open Engineering.
2. Brown, C. (2021). Quantitative modelling and computer simulation. In The Routledge
Handbook of Landscape Ecology. Routledge.
3. Bossel, H. (2018). Modeling and simulation. AK Peters/CRC Press.
Supplementary Reading List
1. Albright, Christian, Christopher J. Zappe and Wayne L. Winston (2011). Data Analysis,
Optimization and Simulation Modelling: Australia, Cengage learning
2. Rajib Mall (2010). Fundamentals of software engineering: New Delhi, PHI learning
3. Aluja, J. G. (2012). Towards an advanced modelling of complex economic phenomena.
Berlin, Heidelberg: Springer Berlin Heidelberg.
4. Taudes, A. (2005). Adaptive Information Systems and Modelling in Economics and
Management Science. Vienna: Springer Vienna