Stochastic Modeling of Manufacturing Systems PDF
Stochastic Modeling of Manufacturing Systems PDF
Stochastic Modeling
of Manufacturing Systems
Advances in Design,
Performance Evaluation,
and Control Issues
123
George Liberopoulos J. M. Smith
Department of Mechanical Department of Mechanical
and Industrial Engineering and Industrial Engineering
University of Thessaly University of Massachusetts
38334 Volos Amherst, Massachusetts 01003
Greece USA
E-mail: [email protected] E-mail: [email protected]
Barış Tan
Graduate School of Business
Koç University
80910 Sariyer, Istanbul
Turkey
E-mail: [email protected]
Parts of the papers of this volume have been published in the journal OR Spectrum.
Abstract. The problems of factory design are notorious for their complexity. It
is argued in this paper that factory design problems represent a class of problems
for which there are crucial dilemmas and correspondingly deep-seated underlying
paradoxes. These paradoxes, however, give rise to novel paradigms which can bring
about fresh approaches as well as insights into their solution.
1 Introduction
The purpose of this paper is to develop a new paradigm for factory design that
integrates much of the theoretical underpinnings of the problems and processes
encountered in the author’s experiences with factory design. As a side benefit to this
paper, many of the ideas discussed within point towards a new direction for which
manufacturing and industrial engineering professionals might re-align themselves,
since the paradigms which have guided these fields are in need of a new vision and
repair.
1.1 Motivation
The origins of this paper stem from an invitation to give a keynote address at a
conference on the Analysis of Manufacturing Systems 1 where the idea of the
I would like to thank the referees for their insights and suggestions and pointing out some
problems in earlier drafts. My approach to factory design has evolved over the years, and is
still evolving, and it is largely due to the influence of Professor Horst Rittel, my professor
at the University of California at Berkeley during my formative undergraduate days, who
instilled much of the basis of this philosophy.
1
4th Aegean Conference on: “The Analysis of Manufacturing Systems”, Samos Island
Greece, July 1st-July 4th, 2003
4 J. MacGregor Smith
address was to recount the author’s philosophy about manufacturing systems design
and in particular an approach to factory design problems.
Concurrently with the conference there appeared a related conundrum on the
email listserv: [email protected] of the Industrial Engineering fac-
ulty about an “identity” crisis within the industrial engineering community and the
direction of the profession and more practically speaking what fundamental courses
should be taught students of industrial engineering. It is not the first time this iden-
tity crisis has arisen in IE, nor is the crisis one exclusive to industrial engineers, as it
commonly occurs throughout most professions from time-to-time. Paradoxically,
all professions have a vested interest in their clients, but cannot be trusted to act in
their clients best interests, “a conspiracy against the laity.”[21, 17].
Since, the Factory Design Problem (FDP) is a very important aspect within
manufacturing and industrial engineering, it became obvious that the subject matter
of the keynote address and the crisis in industrial engineering education are two
closely related matters. So while not attempting to be presumptuous, the resulting
paper was a response partly to this crisis and also more importantly to demonstrate
the author’s philosophy about factory design. The viewpoint and conclusions in the
paper may also apply to the problems of factory planning and control, but the focus
for the present paper is on the FDP problem.
2 Background
λ11 θ1
θ2
λ21
λjk
Σ θn
Fig. 1. Factory flow design paradigm
Dilemmas in factory design: paradox and paradigm 5
useful caricature of the flow paradigm. The Σ represents the mathematical model
of the queueing network underlying the people, resources, products and their flow
relationships.
The professionals (especially the academics) would like to know the set of
underlying equations Σ (no questions asked) which would allow them to design the
factory to maximize the overall throughput (Θ) of the products and also minimize
the work-in-process (WIP) inside the plant.
The desire to find all these equations, or laws [9] as some people would like
to characterize them, is largely attributed to the scientific foundation of Industrial
Engineering education with a strong physics, chemistry, and mathematics back-
ground. A sterling example of one of these laws is Little’s Law L = λW which is
an extremely robust, effective tool to calculate numbers of machines, throughput,
and waiting times in queueing processes[9]. What will be shown in the following
is that this scientific approach is deficient. The problems of factory design cannot
be answered with just a scientific background, but need to be augmented with other
knowledge-based skills. The scientific background is necessary but not sufficient
to solve the problem.
In order to realize this factory flow paradigm, most IE professionals system-
atically define the multiple products (there can be hundreds) and their input rates
and raw material requirements, the constraint relationships with the machines, peo-
ple, resources, and materials handling equipment, and the functional equations for
achieving the WIP and throughput objectives, utilization, cycle time, lateness, etc.
This factory flow paradigm is often realized as a series of well-defined steps or
phases similar to the following top-down approach (see Fig. 2).
This top-down approach is also a hallmark of an operations research (OR)
paradigm typically argued for in OR textbooks found in the Industrial Engineering
curriculum. While this top-down (“waterfall”) [3] paradigm has its merits, mainly
for project management, it will be argued in this paper that other paradigms are
warranted, ones more realistically appropriate for treating FDPs. A key criticism
of the top-down approach is that no feedback loops occur at the detailed stages,
which is clearly unrealistic. A bottom-up approach, on the other hand, is really not
much better, since one has no real overall knowledge of what is being constructed.
One needs a paradigm that is paradoxically top-down and bottom up at the same
time. Unfortunately, very few individuals are capable of this prescient feat, thus
necessitating development of new external aids.
It will also be argued later on in this paper, that the recommended paradigm has
strong implications for changes in the profession and in the education of manufac-
turing and industrial engineers.
2.1 Definitions
Before we proceed too far along, it would be good to posit some of the key definitions
and notation utilized throughout the paper [6].
Dilemma: (Late Greek) dilEmmat, dilEmmatos- an argument presenting two or
more conclusive alternatives against an opponent; a problem involving
a difficult choice; a perplexing predicament.
6 J. MacGregor Smith
Optimal TND
Step 5.0
Alternatives
Stochastic Flow
Step 6.0
Matrices
Evaluation of
Step 7.0
Alternatives
Factory Plan
Step 8.0
Synthesis
Sensitivity
Step 9.0
Analysis
Step 10.0
Factory Plan
Implementation
The notion of a dilemma in Factory Design is that we are often faced with diffi-
cult issues of what to do, and, occasionally, we must select between two alternatives
that are not necessarily desirable.
The notion of paradox is important because it helps frame the seemingly con-
tradictory elements which are contrary to common sense.
Dilemmas give rise to paradoxes which in turn underly paradigms for solution.
Paradigm is a particularly appropriate word when one thinks of it as a “pattern”,
since this is often what we employ in resolving design problems because of its
modular structure.
All three of these concepts are crucial underpinnings to what is to follow and
they form the basis of the general design “philosophy” purported in this paper.
The fact that these three concepts are derived from the Greek philosophers is an
indication of their importance.
2.2 Notation
– ∆:= Dilemma
– χ:= Paradox
– δi := Deontic issue
– i := Causal or explanatory issue
– ιi := Instrumental issue
– φi := Factual issue
– πi := Planning Issue
– FDP:= Factory Design Problem
– WP:= Wicked Problem
– TP:= Tame Problem
– IBIS:= Issue Based Information System
– NI:= Non-Inferior set of solutions
Figure 3 illustrates the initial layout of the plant that formed the basis of the layout
and systems model about to be discussed. One can see the 4 × 4 gaylords spread
throughout the facility in Figure 3.
As one can see in the plant, there is little room for expansion and there is a
restricted material handling system where the forklift traffic coming and going
must traverse the same aisles.
8 J. MacGregor Smith
The notion of the dilemmas in factory design stems from a seminal paper of Horst
Rittel and Mel Webber [17] on wicked problems. They outline the characteristics
of wicked problems and go on to recount how many planning problems are actu-
ally wicked problems. In fact they argue that there are essentially two classes of
problems:
– Tame Problems (TPs)
– Wicked Problems (WPs)
Tame problems are like puzzles: precisely described, with a finite (or count-
ably infinite) set of solutions, although perhaps extremely difficult to solve.
Problems solved via numerical and combinatorial algorithms can be grouped
in this category. The relationship of Computational Complexity and its classes
P, N P, N P−Complete, and N P−Hard are very appropriate characterizations
for tame problems. Also, more recently, designing large scale interacting systems
has been shown to be N P- complete [5].
It will be argued that the N P Complexity classification is a useful way of char-
acterizing TPs. On the other hand, Wicked problems are the exact opposite of tame
problems, and while not “evil” in themselves, present particulary nasty character-
istics which Rittel and Webber feel justly to deserve the approbation. Their wicked
Dilemmas in factory design: paradox and paradigm 9
WP
WP
NP
N P−Hard
N P−Complete
Fig. 4. Wicked problem tame problem dichotomy
problem framework is useful for characterizing the FDP, since the characteristics of
FDPs as shall be argued are similar. Not all IEs or manufacturing engineers might
agree with the equivalence statement, but the equivalence framework, as we shall
argue, will become the basis for the new paradigm.
Very often, IEs utilize algorithmic approaches to solve FDPs, so they become
integral parts of the solution process of factory design problems, but a key question
here is: Can we utilize systematic procedures to solve FDPs?
While no formal classification of WPs has been developed so far, other than what
is depicted in Figure 4, it appears that the distinction between one type of wicked
problem and another can be based on the following three measurable dimensions:
The dilemma argues that factory design problems cannot be written down on a
sheet of paper (like a quadratic equation), given to someone, where they then can
go off into a corner and work out the solution. Students are continually drilled with
textbook problems (the author is guilty of this himself), but these are not the real
problems. Recent research on the modularization of design problems has shown that
modularization avoids trade-offs in decision making and often ignores important
interactions between decision choices [5].
If someone states the problem as: “build a new plant” or “remodel the existing
facility”, or “add another storey”, then, i.e. the solution and problem are one and
the same! This is antithetical to the scientific paradigm. In fact, the entire edifice
of NP-Completeness problems (i.e. Tame Problems) is critically structured around
the precise problem definition e.g. 3-satisfiability.
For FDPs, it is important whom you talk with and their worldview because
in the ensuing dialog the solution to the problem and the problem definition will
emerge.
In the case of the polymer recycling plant, when the facility was first examined,
their receiving and shipping areas were co-located in the same area of the plant, see
the lower left hand corner of Figure 3 which resulted in severe material handling
conflicts with forklift truck movements, accidents, and space utilization problems.
It was obvious that separate receiving and shipping areas were desirable– thus, the
problem was the same as the solution: “re-layout the plant and separate receiving
and shipping.”
Thus, we have the first formal paradox: χ1 := Every formulation of a problem
corresponds to a statement of its solution and vice versa[14].
This first dilemma of factory design is a most difficult one. One cannot know
a priori the problems inherent in factory design, independent of the client and the
context around which the problem occurs. In essence, the factory design process is
essentially information deficient.
Many “experts” in manufacturing and IE purport to know the answers, yet one
must talk with the owners, the plant manager, the line staff, and many others involved
with the facility, before the problems and their solutions can be identified. As the
paper proceeds, we will postulate the underlying principles of the new paradigm
as Propositions. In fact, the principle underlying the paradigm associated with this
first dilemma and paradox is:
Proposition 1. The FDP design system ≡ Knowledge/Information System.
What is meant here by an knowledge/information system? The knowl-
edge/information system here is a special type of information system, not just
a sophisticated data base system, where one collects data for the sake of collecting
data, but data is collected to resolve the planning issues. The planning issues are
the fundamental units within the information system [13]. A related information
system approach based on the first proposition is that of Peter Checkland’s work
[1], however, the information system and resulting paradigm discussed in this paper
is based upon different concepts and is directly related to the FDP.
Dilemmas in factory design: paradox and paradigm 11
φi ι1
πi i
δi ι2
Questions Issues
Arguments Heard
Decisions Reached
Knowledge Gained
Fig. 6. Planning issues resolution process
– Explanatory Issue (i ):= There is not clear separation between the forklift
trucks and the plant personnel within the receiving and shipping area.
Instrumental Issue (ι1 ):= If space is available, separate receiving and shipping
and design the material handling systems in the plant in a U-shape layout.
Instrumental Issue (ι2 ):= If space is unavailable, clearly demarcate the
receiving and shipping areas and the paths of the vehicles and pedestrians.
The reason the above are stated as issues is that evidence for their support
must be brought forth to support or refute each issue. People must be convinced
of the case being made. Some issues are easily resolved as questions, while others
may not be so easily resolved. Not everyone might agree with what we mean by
“excessive” traffic in the receiving and shipping area of φi so some supporting
data may be necessary. Likewise, even the instrumental issues will likely need
supporting evidence such as is possible with sophisticated simulation and queueing
models to estimate expected (maximum) volume of forklift traffic, # number of
expected gaylords in the shipping and receiving areas, etc. Why a U-shape layout?
is certainly arguable. Figure 6 is suggestive of the issue resolution process.
While this approach to problem formulation through the planning issues
paradigm can be seen as well-structured, there can be many planning issues in
factory design, which, unfortunately, leads to the next dilemma.
Dilemmas in factory design: paradox and paradigm 13
C1 C2 · · · Cj · · · Cn−1 Cn
The second dilemma underscores the fact that there are many problems nested
together, there is not simply one isolated problem to be solved. The paradox sur-
rounding the second dilemma is that: χ2 :=Tackling the problem as formulated
may lead to curing the symptoms of the problem rather than the real problem-you
are never sure you are tackling the right problem at the right level.
One needs to tackle the problems on as high a level as possible. In the poly-
mer recycling project, issues of scheduling, resource configuration and utilization,
quality control, and many others became functionally related to the plant layout
problem. As will be shown, these other issues emerged as critical to the plant lay-
out. The principle needed in the paradigm in response to the paradox of dilemma
#2 is:
Each Cj represents a stage of the DP paradigm and each state has a set of
alternative ways of resolving each planning issue πij labelled as alternative k within
each planning issue xijk Transitions between states in adjacent stages would have
an associated cost for transitioning or linking adjacent states. One possible recursive
cost function for an additive or separable resource constrained problem could be
[8]:
∗
fj (πi , xijk ) = cπxk + fj+1 (xijk )
In general, the recursive cost function need not be additive, yet the additive
situation would be quite appropriate in many resource constrained IBIS scenarios.
The general recursive cost function relationship would more likely be:
One can consider the overall cost of resolving a set of planning issues as a
path/tree through the stages and states of the IBIS problem. Each such path repre-
sents a morphological plan solution.
πi πn
cij
π
cjk ckm
cmn
πj πm
Figure 8 illustrates another IBIS network that was utilized by the author to
approach a resource planning problem at the University of Massachusetts [20]. In
this study there were five categories (stages) of planning issues (22 issues total):
When one plays chess, there are only a finite number of moves to start the game.
In linear programming, one needs a starting feasible solution to begin the process.
In factory design, there is no one single place to start the problem formulation and
solution process.
For the polymer recycling project, we could have visited other polymer pro-
cessing plants, travelled to other locations besides Western Massachusetts, read all
the literature on polymer re-processing, carried out a mail survey, talked with all the
employees, and so on. We should have done all the above, but alas, it was not prac-
tical nor cost-effective. This dilemma is founded on the following paradox: χ3 :=If
one is rational, one should consider the consequences of their actions; however,
one should also consider the consequences of considering the consequences, i.e. if
there is nowhere to start to be rational, one should somehow start earlier [15].
The paradox indicates that a great deal of knowledge about the system under
study is needed to assist the client and the engineers in making decisions about the
FDP. Of course, a logical response to this paradox is the following principle:
Proposition 4. Construct a system representation Σ (analytical or simulation) of
the manufacturing system within which the FDP is situated.
This principle is very useful one but obviously can be expensive in time to con-
struct. It makes eminent sense in the supply-chain business environment currently
popular, so the more one understands the logistics and the manufacturing systems
and processes, the better. At this point, the system model Σ becomes an integral
part of the new paradigm.
A discrete-event digital simulation model of the polymer recycling plant was
constructed in order to better understand the manufacturing processes and the sys-
tem as well as the logistics of the product shipments to and from the plant. This
16 J. MacGregor Smith
was felt to be crucial before simply re-laying out the plant and will be shown to be
an extremely fortunate decision.
Figure 9 illustrates the layout plan arrived at with a u-shaped circulation flow
to eliminate the forklift conflicts from the previous scheme (Fig. 3). Unfortunately,
this was not the end of the story.
Thus, for the Manufacturing and IE professional, system models such as supply-
chain networks, simulation and queueing network models are critically important
to frame the context of the problem. The “systems approach” is still sage advice.
Related to ∆3 is:
In chess, you either win, lose, or draw– game over! In linear programming, either
you find the optimal solution, an unbounded one, or find out that the problem is
infeasible. In factory design, you can always make improvements to the system. As
we saw above, simply arriving at the layout design in not enough. Thus, we have
the following paradox: χ4 :=If one is rational, one should also know that every
consequence has a consequence, so once one starts to be rational, one cannot stop-
one can always do better [15].
Dilemmas in factory design: paradox and paradigm 17
As one can argue, there are many explanations for each planning issue, and thus,
there are many potential solutions, not just one. Refer to Figure 11 for an illustration
of this process.
18 J. MacGregor Smith
ι11
φi
i1 ι12
πi i2 ι21
δi
ι22
ι11
φi ι12
i1
ι21
i2
πi
ι22
δi i3
ι31
ι32
The paradox surrounding this dilemma: χ5 :=People need to choose one so-
lution as a “best” solution; but, unfortunately, there are many potential solutions,
with correspondingly difficult tradeoffs.
In response to this situation, one needs much help to generate innovative so-
lutions to the underlying FDP problems. Layout planning algorithms were used
in the polymer processing plant to help come to a solution to the layout problem
and also were seen as vehicles to resolve issues in the layout problem, not as ends
in themselves. Besides using combinatorial optimization algorithms, one needs to
generate a special set of solutions, in fact, the paradigmatic principle which ∆5 is
based upon is closely related to the next dilemma both in spirit and in practice.
In most TPs, there are objective functions which clearly demarcate feasible from
optimal solutions. The gap between linear and nonlinear programming TPs can be
quite huge. In wicked problems, there are a multiple number of objective func-
tions, not only linear and nonlinear ones. Paradoxically, in factory design we have:
χ6 :=Solutions are either good or bad not right or wrong (true or false). There are
multiple criteria embedded within each planning issue.
Dilemmas in factory design: paradox and paradigm 19
∆5 and ∆6 are closely related since one of the reasons why there are so many
solutions is that there are multiple objectives in FDP. Thus, we need to generate a
Non-Inferior(NI) set of solutions, and the notion of optimality becomes spurious
because it only makes sense in a single objective environment. It is very rare that an
FDP has only one objective. In another project we worked on, the project manager
gave out the following daunting list of objectives before we started our project:
Optimize product flow in order to:
– Reduce project costs;
– Reduce WIP investment;
– Increase Inventory turns;
– Reduce scrap and rework;
– Quicker response to customer needs;
– Improve response time to quality problems;
– Improve housekeeping;
– Better utilize floor space;
– Improve safety;
– Eliminate fork lift trucks.
Thus, we have the following:
Proposition 5. Generate a Non-inferior (NI) set of Solutions based upon the mul-
tiple objectives/criteria {f1 (x), f2 (x), . . . , fp (x)} involved in the FDP.
The implications of ∆5 and ∆6 for the manufacturing and IE profession and
curriculum are that multi-criteria and multi-objective programming are essential
methodological concepts and algorithmic tools in manufacturing systems and IE.
MCDM concepts and methodologies have slowly been introduced into IE curricu-
lums which is a very positive sign. Related to the last dilemma is the fact that:
In factory design problems, one doesn’t get a second chance. One can play chess
or solitaire many times over. Solving mathematical programming programs on
one computer or a distributed computer network is routine. Markovian queueing
networks can be run forwards or backwards in time and this affords their decom-
posability. The paradox is that χ8 :=FDPs are not time reversible. There is no
20 J. MacGregor Smith
trial and error with factory design problems, no experimentation you cannot build
a plant, tear it down, and rebuild it without significant consequences.
This dilemma and paradox are very troubling because once the factory design
project goes to the construction phase, there is no turning back. In many scientific
disciplines, repeated experimentation to test an hypothesis is routine and accepted
practice because the costs and consequences are justified. The principle relating
∆7 and ∆8 is:
Proposition 6. Dynamic Models Σ(t) are needed for FDPs.
For the manufacturing and IE profession, simulation modelling is accepted
practice and with good reason. Analytical system models with queueing networks
are also becoming more important and many of these analytical tools are often used
in addition to simulation.
The polymer recycling project is most appropriate as an illustration of these
dilemmas at this stage. In order to test our final factory design layout, we ran the
simulation model and calculated the number of gaylords in the warehouse as a
function of variations in the input demand λi , ∀ i, from 0% − 20%. Figure 12
illustrates the results of the simulation runs for the total number of raw material
gaylords possible on the y-axis vs. the input demand on the x-axis.
The first 3 columns of Figure 12 illustrate the number of raw material gaylords
as a function of input demand. Thus, as one can see the initial design of the plant
was fairly robust.
However, Figure 13 revealed that as the input volume ramped up in the plant to
120% (3rd column), a serious problem arose with one of the key resources because
at 120% of input demand the minimum raw material input volume went negative
by 670 gaylords. Essentially, the plant input-processing of raw materials basically
shut-down. We needed to find out which resource was the bottleneck.
After a detailed analysis of the simulation model outputs, it was revealed in the
third column of Figure 14 where it is shown that the auger blender was operating
at 100% capacity and could not handle any more input. The auger blender was the
bottleneck. Thus, if the input demand was to be greater than 20% of the current
demand, it became obvious that a minimum of 2 auger blenders were needed as
opposed to only one.
This is also very challenging for professionals as well as academics, because the
principles of scientific research can be compromised. Science can accept or refute
an hypothesis, mathematicians can disprove conjectures, but running a business
cannot accept failure. Compromise is essential. The cynical remarks by George
Bernard Shaw [21] mentioned at the beginning of this paper underly the moral
dilemma captured by this last dilemma. The paradox surrounding this last dilemma
is that: χ10 :=Design cannot be carried out in solitary confinement, the FDP design
process is democratic. The final principle summarizes our overall approach to FDP:
Dilemmas in factory design: paradox and paradigm 23
The elemental implications for the manufacturing and IE profession are proba-
bly best described in a summary implication diagram centered around the dilemmas
∆1
∆2
∆3
Σ ∆4
∆5 ∆6 IBIS
∆7
Σ(t) ∆8
∆9
∆10
Fig. 16. Final IBIS paradigm
24 J. MacGregor Smith
and the IBIS which must integrate them and the models necessary to resolve the
issues (see Fig. 16).
The IBIS is necessary to frame ∆1 and to interrelate the different issues and
problems spawned by ∆2 . A systems model Σ is necessary for ∆3 , ∆4 . Generating
ideas and evaluating as captured by ∆5 and ∆6 must rely on effective algorithmic
tools but these must be tempered with a cognizance of the multiple objectives and
criteria involved so that effective tradeoffs can be made. A Stochastic/Dynamic
model Σt is necessary to address the variability, prediction, and control issues
surrounding ∆7 , ∆8 , ∆9 . Indeed, the degree of uncertainty in most FDPs makes
this last stage very challenging. Finally, the IBIS needs to be an open and democratic
system that links all aspects of the FDP process.
Perhaps the weakest element in most manufacturing and IE curriculums, at least
from the perspectives argued in this paper, is adequate exposure to FDPs as Wicked
Problems.
Design problems within academia with real clients are most desirable, whereas,
if this is not possible, projects derived from a real world setting with realistic con-
straints and expectations should be pursued. In a very positive sense, many schools
have semester or year-long senior design projects which can capture this aspect
of the FDP problem. An interesting development in Engineering education is the
Conceiving-Designing-Implementing-Operating real-world systems and products
(CDIO) collaborative http://www.cdio.org/ which underscores much of what has
been argued here in the is paper. It is oriented to all of Engineering rather than just
Industrial and Manufacturing Engineering, but its philosophy is similar. However,
it does not appear to rely on an IBIS approach, which as argued for in this paper, is
very critical to success in resolving real-world problems.
Problem formulation and structuring for WPs are very difficult topics to treat
and teach, but the IBIS framework is something which has clear paradigmatic
and teachable elements. Of course, how these elements are put together into the
curriculum remains the real wicked problem.
References
1. Checkland P (1984) Rethinking a systems approach. In: Tomlnson R, Kiss I (eds) Re-
thinking the process of operational research and systems analysis, pp 43–65. Pergamon
Press, New York
2. Cook SA (1971) The complexity of theorem-proving procedures. Proc. of the Third
ACM Symposium on Theory of Computing, pp 4–18
3. DeGrace P, Stahl L (1990) Wicked problems, righteous solutions. Yourden Press Com-
puting Series, Upper Saddle River, NJ
4. Dixon JR, Poli C (1995) Engineering design and design for manufacturing. Field Stone,
Conway, MA
5. Ethiraj SK, Levinthal D (2004) Modularity and innovation in complex systems. Man-
agement Science 50(2): 159–173
6. Funk, Wagnalis (1968) Standard college dictionary. Harcourt, Brace and World, New
York
7. Garey MR, Johnson DS (1979) Computers and intractability: a guide to the theory of
NP-completeness. Freeman, San Francisco
8. Hillier F, Lieberman G (2001) Introduction to operations research. McGraw-Hill, New
York
9. Hopp W, Spearman M (1996) Factory physics. McGraw-Hill, New York
10. Karp RM (1972) Reducibility among combinatorial problems. In: Miller RE, Thatcher
JW (eds) Complexity of computer computations. Plenum Press, New York
11. Karp RM (1975) On the computational complexity of combinatorial problems. Net-
works 5: 45–68
12. Krishnamoorthy MS, Deo N (1979) Complexity of minimum-dummy-activities prob-
lem in a PERT network. Networks 9: 189–194
13. Kunz W, Rittel H (1970) Issues as elements of information systems. Institute for Urban
and Regional Development, Univeristy of California, Berkeley, Working Paper No. 131
14. Rittel H (1968) Lecture notes for arch, vol 130. Unpublished, Department of Architec-
ture, University of California, Berkeley, CA
15. Rittel H (1972a) On the planning crisis: systems analysis of the ‘first and second
generations’. Bedrifts Okonomen 8: 390–396
16. Rittel H (1972b) Structure and usefulness of planning information systems. Bedrifts
Okonomen (8): 398–401
17. Rittel H, Webber M (1973) Dilemmas in a general theory of planning. Policy Sciences
4: 155–167
18. Rittel H (1975) On the planning crisis: systems analysis of the first and second gen-
erations. Reprinted from: Bedrifts ØKonmen, No. 8, October 1972; Reprint No. 107,
Berkeley, Institute of Urban and Regional Development
19. Roberts N (2001) Coping with wicked problems: the case of Afghanistan. Learning from
International Public Management Forum, vol 11B, pp 353–375. Elsevier, Amsterdam
20. Robinson D, et al. (1999) A change proposal for small-scale renovation and construction
projects. Project Team Report. Internal University of Massachusetts Report. Campus
Committee for Organization Restructuring, University of Massachusetts, Amherst, MA
21. Shaw GB (1946) The doctors dilemma. Penguin, New York
22. Smith J MacGregor, Larson RJ, MacGilvary DF (1976) Trial court facility. National
Clearinghouse for Criminal Justice Planning and Architecture, Monograph B5. In:
Guidelines for the planning and design of state court programs and facilities. University
of Illinois, Champaign, IL
Section II: Unreliable Production Lines
Lean buffering in serial production lines
with non-exponential machines
Emre Enginarlar1 , Jingshan Li2 , and Semyon M. Meerkov3
1
Decision Applications Division, Los Alamos National Laboratory,
Los Alamos, NM 87545, USA
2
Manufacturing Systems Research Laboratory, GM Research and Development Center,
Warren, MI 48090-9055, USA
3
Department of Electrical Engineering and Computer Science, University of Michigan,
Ann Arbor, MI 48109-2122, USA (e-mail: [email protected])
Abstract. In this paper, lean buffering (i.e., the smallest level of buffering neces-
sary and sufficient to ensure the desired production rate of a manufacturing system)
is analyzed for the case of serial lines with machines having Weibull, gamma, and
log-normal distributions of up- and downtime. The results obtained show that: (1)
the lean level of buffering is not very sensitive to the type of up- and downtime dis-
tributions and depends mainly on their coefficients of variation, CVup and CVdown ;
(2) the lean level of buffering is more sensitive to CVdown than to CVup but the
difference in sensitivities is not too large (typically, within 20%). Based on these
observations, an empirical law for calculating the lean level of buffering as a func-
tion of machine efficiency, line efficiency, the number of machines in the system,
and CVup and CVdown is introduced. It leads to a reduction of lean buffering by a
factor of up to 4, as compared with that calculated using the exponential assump-
tion. It is conjectured that this empirical law holds for any unimodal distribution of
up- and downtime, provided that CVup and CVdown are less than 1.
1 Introduction
The smallest buffer capacity, which is necessary and sufficient to achieve the desired
throughput of a production system, is referred to as lean buffering. In (Enginarlar
et al., 2002, 2003a), the problem of lean buffering was analyzed for the case of
Correspondence to: S.M. Meerkov
30 E. Enginarlar et al.
serial production lines with exponential machines, i.e., the machines having up-
and downtime distributed exponentially. The development was carried out in terms
of normalized buffer capacity and production system efficiency. The normalized
buffer capacity was introduced as
N
k= , (1)
Tdown
where N denoted the capacity of each buffer and Tdown the average downtime of
each machine in units of cycle time (i.e., the time necessary to process one part
by a machine). Parameter k was referred to as the Level of Buffering (LB). The
production line efficiency was quantified as
P Rk
E= , (2)
P R∞
where P Rk and P R∞ represented the production rate of the line (i.e., the average
number of parts produced by the last machine per cycle time) with LB equal to k
and infinity, respectively. The smallest k, which ensured the desired E, was denoted
as kE and referred to as the Lean Level of Buffering (LLB).
Using parameterizations (1) and (2), Enginarlar et al., (2002, 2003a) derived
closed formulas for kE as a function of system characteristics. For instance, in the
case of two-machines lines, it was shown that (Enginarlar et al., 2002)
⎧
⎪
⎪ 2e(E − e)
⎪
⎪ , if e < E,
⎨ 1−E
exp
kE = (3)
⎪
⎪
⎪
⎪
⎩ 0, otherwise.
Here the superscript exp indicates that the machines have exponentially distributed
up- and downtime, and e denotes machine efficiency in isolation, i.e.,
Tup
e= , (4)
Tup + Tdown
where Tup is the average uptime in units of cycle time. For the case of M > 2-
machine serial lines, the following formula had been derived (Enginarlar et al.,
2003a):
⎧
⎪ e(1 − Q)(eQ + 1 − e)(eQ + 2 − 2e)(2 − Q)
⎪
⎪ ×
⎪
⎪ Q(2e − 2eQ + eQ2 + Q − 2)
⎪
⎪
⎨ E−eE+eEQ−1+e−2eQ+eQ2 +Q
⎪
1
exp
kE (M ≥3)= ln , if e<E M −1 , (5)
⎪
⎪ (1 − e − Q + eQ)(E − 1)
⎪
⎪
⎪
⎪
⎪
⎪
⎩
0, otherwise,
where
1
−3 M/4
1 −3 M/4
1+( M 1+( M M −2
M −1 ) M −1 )
Q = 1−E2 + E2 − E M −1
1
E M −1 − e
× exp − . (6)
1−E
Lean buffering in serial production lines with non-exponential machines 31
Table 1. Difficulties of the non-exponential case as compared with the exponential one
2.1 Model
The block diagram of the production system considered in this work is shown
in Figure 1, where the circles represent the machines and the rectangles are the
buffers. Assumptions on the machines and buffers, described below, are similar to
those of (Enginarlar et al., 2003a) with the only difference that up- and downtime
distributions are not exponential. Specifically, these assumptions are:
(i) Each machine mi , i = 1, . . . , M , has two states: up and down. When up, the
machine is capable of processing one part per cycle time; when down, no production
takes place. The cycle times of all machines are the same.
34 E. Enginarlar et al.
(ii) The up- and downtime of each machine are random variables measured in units
of the cycle time. In other words, uptime (respectively, downtime) of length t ≥ 0
implies that the machine is up (respectively, down) during t cycle times. The up-
and downtime are distributed according to one of the following probability density
functions, referred to as reliability models:
(a) Weibull, i.e.,
W
(t) = pP e−(pt) P tP −1 ,
P
fup
W
(t) = rR e−(rt) RtR−1 ,
R
fdown (8)
W W
where fup (t)and fdown (t)
are the probability density functions of up- and
downtime, respectively and (p, P ) and (r, R) are their parameters. (Here, and
in the subsequent distributions, the parameters are positive real numbers). These
distributions are denoted as W (p, P ) and W (r, R), respectively.
(b) Gamma, i.e.,
g (pt)P −1
fup (t) = pe−pt ,
Γ (P )
g (rt)R−1
fdown (t) = re−rt , (9)
Γ (R)
∞
where Γ (x) is the gamma function, Γ (x) = 0 sx−1 e−s ds. These distribu-
tions are denoted as g(p, P ) and g(r, R), respectively.
(c) Log-normal, i.e.,
1 (ln(t)−p)2
LN
fup (t) = √ e− 2P 2 ,
2πP t
1 (ln(t)−r)2
LN
fdown (t) = √ e− 2R2 . (10)
2πRt
We denote these distributions as LN (p, P ) and LN (r, R), respectively.
The expected values, variances, and coefficients of variation of distributions
(8)–(10) are given in Table 2.
(iii) The parameters of distributions (8)–(10) are selected so that the machine effi-
ciencies, i.e.,
Tup
e= , (11)
Tup + Tdown
and, moreover, Tup , Tdown , CVup , and CVdown of all machines are identical for
all reliability models, i.e.,
−1 1
Tup = p Γ 1 + (Weibull)
P
Lean buffering in serial production lines with non-exponential machines 35
Table 2. Expected value, variance, and coefficient of variation of up- and downtime distri-
butions considered
P
= (gamma)
p
2
= ep+P /2
(log-normal);
−1
Tdown = r Γ (1 + 1/R) (Weibull)
R
= (gamma)
r
2
= er+R /2 (log-normal);
Γ (1 + 2/P ) − Γ 2 (1 + 1/P )
CVup = (Weibull)
Γ (1 + 1/P )
1
= √ (gamma)
P
= eP 2 − 1 (log-normal);
Γ (1 + 2/R) − Γ 2 (1 + 1/R)
CVdown = (Weibull)
Γ (1 + 1/R)
1
= √ (gamma)
R
= eR2 − 1 (log-normal).
Remark 1.
– Assumptions (i)–(iii) imply that all machines are identical from all points of
view except, perhaps, for the nature of up- and downtime distributions. The
buffers are also assumed to be of equal capacity (see (iv)). We make these
assumptions in order to provide a compact characterization of lean buffering.
– Assumption (ii) implies, in particular, that time-dependent, rather than
operation-dependent failures, are considered. This failure mode simplifies the
analysis and results in just a small difference in comparison with operation-
dependent failures.
2.2 Notations
Using the parameterizations (1), (2), the model (i)–(vi), and the notations (12)–(16),
this paper is intended to
– develop a method for calculating Lean Level of Buffering in production lines
(13)–(15) under the assumption that the coefficients of variation of up- and
downtime, CVup and CVdown , are identical, i.e., CVup = CVdown = CV ;
– develop a method of calculating LLB in production lines (13)–(15) for the case
of CVup =/ CVdown ;
– extend the results obtained to production lines (16).
Solutions of these problems are presented in Sections 4–6 while Section 3
describes the approach used in this work.
Lean buffering in serial production lines with non-exponential machines 37
3 Approach
Since LLB depends on line efficiency E, the calculation of kE requires the knowl-
edge of the production rate, P R, of the system. Unfortunately, as it was mentioned
earlier, no analytical methods exist for evaluating P R in serial lines with either
Weibull, or gamma, or log-normal reliability characteristics. Approximation meth-
ods are also hardly applicable since, in our experiences, even 1%-2% errors in the
production rate evaluation (due to the approximate nature of the techniques) often
lead to much larger errors (up to 20%) in lean buffering characterization. There-
fore, the only method available is the Monte Carlo approach based on numerical
simulations. To implement this approach, a MATLAB code was constructed, which
simulated the operation of the production line defined by assumptions (i)–(vi) of
Section 2. Then, a set of representative distributions of up- and downtime was se-
lected and, finally, for each member of this set, P R and LLB were evaluated with
guaranteed statistical characteristics. Each of these steps is described below in more
detail.
0.06
g, Tdown = 20
g, Tdown = 100
W, Tdown = 20
W, T = 100
0.05 down
LN, T = 20
down
LN, T = 100
down
0.04
f(t)
0.03
0.02
0.01
0
0 50 100 150 200 250 300 350
t
Remark 2. Both CVup and CVdown considered are less than 1 because, according
to the empirical evidence of (Inman, 1999), the equipment on the factory floor often
satisfies this condition. In addition, it has been shown by Li and Meerkov (2005)
that CVup and CVdown are less than 1 if the breakdown and repair rates of the
machines are increasing functions of time, which often takes place in reality.
Lean buffering in serial production lines with non-exponential machines 39
To evaluate the production rate in systems (17) and (18), using the MATLAB code
and the up- and downtime distributions discussed in Sections 3.1–3.3, zero initial
40 E. Enginarlar et al.
conditions of all buffers have been assumed and the states of all machines at the
initial time moment have been selected “up”. The first 100,000 cycle times were
considered as warm-up period. The subsequent 1,000,000 cycle times were used
for statistical evaluation of P R. Each simulation was repeated 10 times, which
resulted in 95% confidence intervals of less than 0.0005.
The lean buffering, kE , necessary and sufficient to ensure line efficiency E, was
evaluated using the following procedure:
For each model of serial line (13)–(15), the production rate was evaluated first
for N = 0, then for N = 1, and so on, until the production rate P R = E ·P R∞ was
achieved. Then kE was determined by dividing the resulting NE by the machine
average downtime (in units of the cycle time).
Remark 3. Although, as it is well known (Hillier and So, 1991b), the optimal
allocation of a fixed total buffer capacity is non-uniform, to simplify the analysis we
consider only uniform allocations. Since the optimal (i.e., inverted bowl) allocation
typically results in just 1 − 2% throughput improvement in comparison with the
uniform allocation, for the sake of simplicity we consider only the latter case.
Figures 3 and 5 present the simulation results for production lines (17) for all
distributions of Table 3. These figures are arranged as matrices where the rows
and columns correspond to e ∈ {0.85, 0.9, 0.95} and E ∈ {0.85, 0.9, 0.95}, re-
spectively. Since, due to space considerations, the graphs in Figures 3 and 5 are
congested and may be difficult to read, one of them is shown in Figure 4 in a larger
scale. (The dashed lines in Figs. 3–5 will be discussed in Sect. 4.3.) Examining
these data, the following may be concluded:
exp
– As expected, kE for non-exponential machines is smaller than kE . Moreover,
kE is a monotonically increasing function of CV . In addition, kE (CV ) is
convex, which implies that reducing larger CV ’s leads to larger reduction of
kE than reducing smaller CV ’s.
– Function kE (CV ) seems to be polynomial in nature. In fact, each curve of
Figures 3 and 5 can be approximated by a polynomial of an appropriate order.
However, since these approximations are “parameter-dependent” (i.e., different
polynomials must be used for different e and E), they are of small practical
importance, and, therefore, are not reported here.
– Since for every pair (E, e), corresponding curves of Figures 3 and 5 are identical,
it is concluded that kE is not dependent of Tup and Tdown explicitly but only
through the ratio e. In other words, the situation here is the same as in lines with
exponential machines (see (5), (6)).
Lean buffering in serial production lines with non-exponential machines 41
10
Gamma
Weibull
8 log−normal
empirical law
6
kE
0
0 0.2 0.4 0.6 0.8 1
CV
Fig. 4. LLB versus CV for system {(D(p, P ), D(r, R))i , i = 1, . . . , 10} with Tdown =
20, e = 0.9, E = 0.9
Fig. 6. Sensitivity of LLB to the nature of up- and downtime distributions for systems (17)
Lean buffering in serial production lines with non-exponential machines 43
Function 1 (CV ) is illustrated in Figure 6. As one can see, in most cases it takes
values within 10%. Thus, it is possible to conclude that for all practical purposes
kE depends on the coefficients of variation of up- and downtime, rather than
on actual distribution of these random variables.
Figures 7 and 8 present the simulation results for lines (18), while Figure 9 char-
acterizes the sensitivity of kE to up- and downtime distributions. This sensitivity
is calculated according to (19) with the only difference that the max is taken over
A, B ∈ {(18)}. Based on these data, we affirm that the conclusions formulated in
Section 4.1 hold for production lines of the type (13) as well.
Fig. 9. Sensitivity of LLB to the nature of up- and downtime distributions for systems (18)
Fig. 11. Verification: LLB versus CV for system {(D(p, P ), D(r, R))i , i = 1, . . . , 5}
with Tdown = 10
4.3.2 Verification
To verify the empirical law (20), production lines (17) and (18) were simulated with
parameters M , E, and e other than those considered in Sections 4.1 and 4.2. Specif-
ically, the following parameters have been selected: M = 5, E ∈ {0.7, 0.8, 0.9},
e ∈ {0.6, 0.7, 0.8}, Tdown = 10. (In lines (18), the first 5 machines were selected.)
The results are shown in Figure 11. As one can see, the upper bound given by (20)
still holds.
The case of CVup = / CVdown is complicated by the fact that CVup and CVdown may
have different effects on kE . If this difference is significant, it would be difficult
Lean buffering in serial production lines with non-exponential machines 47
to expect that the empirical law (20) could be extended to the case of unequal
coefficients of variation. On the other hand, if CVup and CVdown affect kE in a
somewhat similar manner, it would seem likely that (20) might be extended to the
case under consideration. Therefore, analysis of effects of CVup and CVdown on
kE is of importance. This section is devoted to such an analysis.
To investigate this issue, introduce two functions:
and
where
when α > β, it must be concluded that CVdown has a larger effect on kE than CVup .
If the inequality is reversed, CVup has a stronger effect. Finally, if (25) holds for
some α and β from (24) and does not hold for others, the conclusion would be that,
in general, neither has a dominant effect.
To investigate which of these situations takes place, we evaluated functions
(22) and (23) using the approach described in Section 3. Some of the results for
Weibull distribution are shown in Figure 12 (where the broken lines and CVef f will
be defined in Sect. 5.2). Similar results were obtained for gamma and log-normal
distributions as well (see Enginarlar et al., 2003b for details). From these results,
the following can be concluded:
– For all α and β, such that α > β, inequality (25) takes place. Thus, CVdown
has a larger effect on kE than CVup .
– However, since each pair of curves (22), (23) corresponding to the same α are
close to each other, the difference in the effects of CVup and CVdown is not too
dramatic. To analyze this difference, introduce the function
A
3 (CV |CVup = CVdown = α)
A A
kE (CVup =CV |CVdown = α)−kE (CVdown =CV |CVup =α)
= A
·100 , (26)
kE (CVup =CV |CVdown =α)
where A ∈ {W, g, LN }. The behavior of this function for Weibull distribution
is shown in Figure 13 (see Enginarlar et al., 2003b for gamma and log-normal
distributions). Thus, the effects of CVup and CVdown on kE are not dramatically
different (typically within 20% and no more than 40%).
48 E. Enginarlar et al.
To investigate the validity of the empirical law (27), consider the following
function:
upper bound
where kE is the right-hand-side of (29), i.e.,
upper bound exp
kE (M, E, e, CVef f ) = CVef f kE (M, E, e).
If for all values of its arguments, function ∆(M, E, e) is positive, the right-hand-
side of inequality (27) is an upper bound. The values of ∆(10, E, e) for E ∈
{0.85, 0.9, 0.95} and e ∈ {0.85, 0.9, 0.95} are shown in Table 4. As one can see,
function ∆(10, E, e) indeed takes positive values. Thus, the empirical law (27)
takes place for all distributions and parameters analyzed.
50 E. Enginarlar et al.
Figure 14 illustrates the behavior of this function. Comparing this with Figure 10,
we conclude that the tightness of bound (27) appears to be similar to that of (20).
5.2.2 Verification
To evaluate the validity of the upper bound (27), serial production lines with M = 5,
E ∈ {0.7, 0.8, 0.9}, e ∈ {0.6, 0.7, 0.8}, and Tup = 10 were simulated. For each
of these parameters, systems (17) and (18) have been considered. (For system (18),
the first 5 machines were selected.) Typical results are shown in Figure 15 (see
Enginarlar et al., 2003b for more details). The validity of empirical law (27) for
these cases is analyzed using function ∆(M, E, e), defined in (30) with the only
difference that the first min is taken over A ∈ {(17), (18)}. Since the values of
this function, shown in Table 5, are positive, we conclude that empirical law (27) is
indeed verified for all values of M , E, e, and all distributions of up- and downtime
considered.
Lean buffering in serial production lines with non-exponential machines 51
So far, serial production lines with Weibull, gamma, and log-normal reliability
models have been analyzed. It is of interests to extend this analysis to general
probability density functions. Based on the results obtained above, the following
conjecture is formulated:
The empirical laws (20) and (27) hold for serial production lines satisfying
assumptions (i), (iii)–(vi) with up- and downtime having arbitrary unimodal prob-
ability density functions.
The verification of this conjecture is a topic for future research.
52 E. Enginarlar et al.
7 Conclusions
Results described in this paper suggest the following procedure for designing lean
buffering in serial production lines defined by assumptions (i)–(vi):
1. Identify the average value and the variance of the up- and downtime, Tup ,
2 2
Tdown , σup , and σdown , for all machines in the system (in units of machine
cycle time). This may be accomplished by measuring the duration of the up-
and downtimes of each machine during a shift or a week of operation (depending
on the frequency of occurrence). If the production line is at the design stage,
this information may be obtained from the equipment manufacturer (however,
typically with a lower level of certainty).
2. Using (5), (6), and Tup , Tdown , determine the level of buffering, necessary
and sufficient to obtain the desired efficiency, E, of the production line, if the
exp
downtime of all machines were distributed exponentially, i.e., kE .
σup σdown
3. Finally, if CVup = Tup ≤ 1 and CVdown = Tdown ≤ 1, evaluate the level of
buffering for the line with machines under consideration using the empirical
law
max{0.25, CVup } + max{0.25, CVdown } exp
kE ≤ · kE .
2
As it is shown in this paper, this procedure leads to a reduction of lean buffering
by a factor of up to 4, as compared with that based on the exponential assumption.
References
Altiok T (1985) Production lines with phase-type operation and repair times and finite buffers.
International Journal of Production Research 23: 489–498
Altiok T (1989) Approximate analysis of queues. In: Series with phase-type service times
and blocking. Operations Research 37: 601–610
Altiok T, Stidham SS (1983) The allocation of interstage buffer capacities in production
lines. IIE Transactions 15: 292–299
Altiok T, Ranjan R (1989) Analysis of production lines with general service times and
finite buffers: a two-node decomposition approach. Engineering Costs and Production
Economics 17: 155–165
Buzacott JA (1967) Automatic transfer lines with buffer stocks. International Journal of
Production Research 5: 183–200
Caramanis M (1987) Production line design: a discrete event dynamic system and general-
ized benders decomposition approach. International Journal of Production Research 25:
1223–1234
Chow W-M (1987) Buffer capacity analysis for sequential production lines with variable
processing times. International Journal of Production Research 25: 1183–1196
Conway R, Maxwell W, McClain JO, Thomas LJ (1988) The role of work-in-process inven-
tory in serial production lines. Operations Research 36: 229–241
Enginarlar E, Li J, Meerkov SM, Zhang RQ (2002) Buffer capacity to accommodating ma-
chine downtime in serial production lines. International Journal of Production Research
40: 601–624
Lean buffering in serial production lines with non-exponential machines 53
Enginarlar E, Li J, Meerkov SM (2003a) How lean can lean buffers be? Control Group Report
CGR 03-10, Deptartment of EECS, University of Michigan, Ann Arbor, MI; accepted
for publication in IIE Transactions on Design and Manufacturing (2005)
Enginarlar E, Li J, Meerkov SM (2003b) Lean buffering in serial production lines with
non-exponential machines. Control Group Report CGR 03-13, Deptartment of EECS,
University of Michigan, Ann Arbor, MI
Gershwin SB, Schor JE (2000) Efficient algorithms for buffer space allocation. Annals of
Operations Research 93: 117–144
Harris JH, Powell SG (1999) An algorithm for optimal buffer placement in reliable serial
lines. IIE Transactions 31: 287–302
Hillier FS, So KC (1991a) The effect of the coefficient of variation of operation times on the
allocation of storage space in production line systems. IIE Transactions 23: 198–206
Hillier FS, So KC (1991b) The effect of machine breakdowns and internal storage on the
performance of production line systems. International Journal of Production Research
29: 2043–2055
Inman RR (1999) Empirical evaluation of exponential and independence assumptions in
queueing models of manufacturing systems. Production and Operation Management 8:
409–432
Jafari MA, Shanthikumar JG (1989) Determination of optimal buffer storage capacity and
optimal allocation in multistage automatic transfer lines. IIE Transactions 21: 130–134
Li J, Meerkov SM (2005) On the coefficients of variation of up- and downtime in manufac-
turing equipment. Mathematical Problems in Engineering 2005: 1–6
Park T (1993) A two-phase heuristic algorithm for determining buffer sizes in production
lines. International Journal of Production Research 31: 613–631
Powell SG (1994) Buffer allocation in unbalanced three-station lines. International Journal
of Production Research 32: 2201–2217
Powell SG, Pyke DF (1998) Buffering unbalanced assembly systems. IIE Transactions 30:
55–65
Seong D, Change SY, Hong Y (1995) Heuristic algorithm for buffer allocation in a production
line with unreliable machines. International Journal of Production Research 33: 1989–
2005
Smith JM, Daskalaki S (1988) Buffer space allocation in automated assembly lines. Opera-
tions Research 36: 343–357
Tempelmeier H (2003) Practical considerations in the optimization of flow production sys-
tems. International Journal of Production Research 41: 149–170
Yamashita H, Altiok T (1988) Buffer capacity allocation for a desired throughput of produc-
tion lines. IIE Transactions 30: 883–891
Analysis of flow lines
with Cox-2-distributed processing times
and limited buffer capacity
Stefan Helber
University of Hannover, Department for Production Management, Königsworther Platz 1,
30167 Hannover, Germany (e-mail: [email protected])
1 Introduction
which has already been analyzed by Buzacott and Kostelski [8] using the Matrix
geometric method. Our analysis of the two-machine system, however, follows the
approach for two-machine systems which is thoroughly explained by Gershwin
[21]. The decomposition algorithm is briefly described in Section 4. In Section 5
we present some preliminary numerical results by comparing our results to those
obtained from the multistage flow line analysis with the stopped arrival queue model
as proposed by Buzacott et al. [9].
2 The model
We assume that the flow line consists of M machines or stages. The processing
times at machine Mi follow a Cox-2 distribution. Each buffer Bi between machines
Mi and Mi+1 has the capacity to hold up to Ci workpieces which flow from the
leftmost to the rightmost machine. An example of such a flow line is depicted in
Figure 1.
a1 a2 a3
µ11 µ12 µ21 µ22 µ31 µ32
The rates of the two phases of stage i are µi1 and µi2 respectively. The second
phase of stage i will be required after completion of phase one with probability ai .
Therefore, a workpiece completes its service at stage i with probability 1 − ai after
the completion of the first phase and with probability ai after the completion of the
second phase. Note that these states of a machine or stage do not represent servers:
No more than one workpiece can be at a machine at any moment in time, and if it
is there, it is in one out of the two phases of the respective machine. Each machine
Mi except for the first and the last can be either idle (starved) or blocked or it can
be processing a part in phase one or two. The state of machine Mi is denoted as
αi (t). The possible machine states are αi (t) ∈ {1, 2, B, S}, representing phase
one, phase two, blocking and starvation. The buffer level n(i) is defined such that it
includes the parts between machines Mi and Mi+1 , the one part residing at machine
Mi+1 (if this machine is not starved) and the one part waiting at machine Mi if this
machine is blocked because the buffer between Mi and Mi+1 is full.
In order to analyze larger systems with more than two machines, we first study a
two-machine line. The state of this two-machine line is given by the state of the first
machine, the state of the second machine, and the buffer level. In the analysis to
follow, we define the buffer level include all parts that are currently being processed
at the second machine, that are waiting in the physical buffer between the first and
58 S. Helber
the second machine, and those parts that have been processed at the first machine
but cannot leave it because the physical buffer between the machines is full so that
the first machine is blocked. That is, we follow the blockage convention which
is described in [21, p. 95]. The (total or extended) buffer capacity is therefore
Ni = Ci + 2. In order to describe the state space, we use use the triple (n, α1 , α2 )
where n denotes the buffer level. The probability of the system being in this state is
p(n, α1 , α2 ). Machine M1 can either be in the first phase (α1 = 1), in the second
phase (α1 = 2) or it can be blocked (α1 = B). The downstream machine M2 can
either be in the first phase (α2 = 1), in the second phase (α2 = 2) or it can be
starved (α2 = S).
This leads to the following transition equations which differ for states with an
empty or almost empty buffer, states for a full a almost full buffer, and the states
with a buffer level that is in between:
Lower boundary states:
µ11 p(0, 1, S) = (1 − a2 )µ21 p(1, 1, 1) + µ22 p(1, 1, 2) (1)
µ12 p(0, 2, S) = a1 µ11 p(0, 1, S) + (1 − a2 )µ21 p(1, 2, 1) +
µ22 p(1, 2, 2) (2)
(µ11 + µ21 )p(1, 1, 1) = (1 − a1 )µ11 p(0, 1, S) + µ12 p(0, 2, S) +
(1 − a2 )µ21 p(2, 1, 1) + µ22 p(2, 1, 2) (3)
(µ11 + µ22 )p(1, 1, 2) = a2 µ21 p(1, 1, 1) (4)
Intermediate stages:
(µ11 + µ21 )p(n, 1, 1) = (1 − a1 )µ11 p(n − 1, 1, 1) + µ12 p(n − 1, 2, 1) +
(1 − a2 )µ21 p(n + 1, 1, 1) + µ22 p(n + 1, 1, 2)
(for 2 ≤ n ≤ N − 2) (5)
(µ11 + µ22 )p(n, 1, 2) = (1 − a1 )µ11 p(n − 1, 1, 2) + µ12 p(n − 1, 2, 2) +
a2 µ21 p(n, 1, 1) (for 2 ≤ n ≤ N − 1) (6)
(µ12 + µ21 )p(n, 2, 1) = a1 µ11 p(n, 1, 1) + µ22 p(n + 1, 2, 2) +
(1 − a2 )µ21 p(n + 1, 2, 1) (for 1 ≤ n ≤ N − 2) (7)
(µ12 + µ22 )p(n, 2, 2) = a1 µ11 p(n, 1, 2) + a2 µ21 p(n, 2, 1)
(for 2 ≤ n ≤ N − 1) (8)
Upper boundary states:
µ21 p(N, B, 1) = (1 − a1 )µ11 p(N − 1, 1, 1) +
µ12 p(N − 1, 2, 1) (9)
µ22 p(N, B, 2) = a2 µ21 p(N, B, 1) +
(1 − a1 )µ11 p(N − 1, 1, 2) +
µ12 p(N − 1, 2, 2) (10)
Analysis of flow lines with Cox-2-distributed processing times 59
3.2 Identities
Conservation of flow. The rate at which parts leave machine M1 is the product
of the steady-state probabilities of all states where M1 is not blocked times the
respective rate for this state:
P R1 = µ11 (1 − a1 )p(0, 1, S) + µ12 p(0, 2, S) +
N
−1
2
(µ11 (1 − a1 )p(n, 1, α2 ) + µ12 p(n, 2, α2 )) (14)
n=1 α2 =1
P R1 = P R2 (16)
The reason is that the flow of material is linear and parts are neither created nor
destroyed at either machine.
60 S. Helber
Rate of changes from phase one to two equals rate of changes from phase two
to one. For each change of machine M1 from phase one to phase two there must
be a change from phase two to phase one
N −1
2
a1 µ11 p(0, 1, S) + p(n, 1, α2 )
n=1 α2 =1
N −1
2
= µ12 p(0, 2, S) + p(n, 2, α2 ) (17)
n=1 α2 =1
The reasoning for the expected processing time E[T2 ] at at the second (down-
stream) machine of a two-machine-line leads to an analogous result:
1 1 1
E[T2 ] = (1 − a2 ) + a2 + (20)
µ21 µ21 µ22
This leads to an equation for the probability of the machine being blocked:
1 1 1
pB = 1 − P R1 (1 − a1 ) + a1 + (22)
µ11 µ11 µ12
Analysis of flow lines with Cox-2-distributed processing times 61
1 − pS 1 − pS
P R2 = = (23)
E[T2 ] (1 − a2 ) µ21 + a2 µ121 +
1 1
µ22
is similar and it also leads to a similar equation for the probability of the downstream
machine not being starved:
1 1 1
pS = 1 − P R2 (1 − a2 ) + a2 + (24)
µ21 µ21 µ22
While equations (21) and (23) can be used to determine the production rate of
a two-machine system, the equations (22) and (24) will later be used in a decom-
position approach to analyze larger flow lines with more than two machines.
In this section, we derive a specialized solution procedure similar to the one given
in [22].
where cj , Xj , Y1j , and Y2j are parameters to be determined. The analysis below
is very similar to the one in [22] and [26, Sect. 3.2.4]. Replacing p(n, α1 , α2 ) by
α1 −1 α2 −1
Xjn Y1j Y2j in Equations (6), (7) and (8), we derive the following non-linear
set of equations:
(µ11 + µ22 )XY2 = a2 µ21 X + (1 − a1 )µ11 Y2 + µ12 Y1 Y2 (26)
(µ12 + µ21 )Y1 = a1 µ11 + (1 − a2 )µ21 XY1 + µ22 XY1 Y2 (27)
(µ12 + µ22 )Y1 Y2 = a2 µ21 Y1 + a1 µ11 Y2 (28)
Equations (26) and (27) are used to eliminate X. From the resulting equation
and (28) we can next eliminate Y2 . A considerable algebraic effort leads to the
following fourth degree equation in Y1
1
s= (µ11 µ12 − µ212 + a1 µ11 µ21 + a2 µ11 µ21 (31)
w
−a1 a2 µ11 µ21 − µ12 µ21 + µ11 µ22
−µ12 µ22 − µ21 µ22 )
1
t = (a1 µ11 (−µ11 + 2µ12 + µ21 + µ22 )) (32)
w
1
v = −(a21 µ211 ) (33)
w
From the first term on the left side of Equation (29) we see that one solution to
(29) is
a1 µ11
Y11 = (34)
µ12
Applying this result to Equation (28), we find
a2 µ21
Y21 = (35)
µ22
and from (34) and (35) in (26) or (27) we see that
X1 = 1. (36)
The remaining three solutions to (29) are1
a φ s
Y12 = 2 − cos − (37)
3 3 3
a φ 2φ s
Y13 = 2 − cos + − (38)
3 3 3 3
a φ 4φ s
Y14 = 2 − cos + − (39)
3 3 3 3
with auxiliary variables
1
a= 3t − s2 (40)
3
1 3
b= 2s − 9st + 27v) (41)
27 ⎛ ⎞
b
φ = arccos ⎝− ⎠ (42)
−a3
2 27
The corresponding values of Y22 , Y22 , and Y24 are again determined via (28).
The values of X2 , X3 , and X4 are next computed from (26) or (27).
Since we have found four solutions to equations (26), (27), and (28), the general
expression for the steady-state probabilities of the internal states is as follows
4
4
α1 −1 α2 −1
p(n, α1 , α2 ) = cj ξj (n, α1 , α2 ) = cj Xjn Y1j Y2j (43)
j=1 j=1
the equations for boundary states hold for each solution ξj (n, α1 , α2 ) of the equa-
tions for internal states. The equation (45) corresponding to state (1, 1, 2), for
example, leads to
4
4
(µ12 + µ22 ) j=1 cj ξj (1, 2, 2)
cj ξj (1, 1, 2) = −
j=1
a1 µ11
4
a2 µ21 j=1 cj ξj (1, 2, 1)
(49)
a1 µ11
64 S. Helber
Similar equations can be found to determine the terms ξj (n, α1 , α2 ) for the other
boundary state probabilities. The terms ξj (n, α1 , α2 ) corresponding to transient
states are all zero.
Now all steady-state probabilities have been related to equation (43). What
remains to be done is to find appropriate values of the coefficients cj in (43).
4
4
µ21 (1 − a2 ) cj ξj (N, B, 1) − µ22 cj ξj (N, B, 2) = 0 (50)
j=1 j=1
Rate of changes from phase one to 2 equals rate of changes from phase two to 1 at
Maschine M1
⎛ ⎞
4 N
−1
2 4
a1 µ11 ⎝ cj ξj (0, 1, S) + cj ξj (n, 1, α2 )⎠
j=1 n=1 α2 =1 j=1
⎛ ⎞
4 N
−1
2
4
−µ12 ⎝ cj ξj (0, 2, S) + cj ξj (n, 2, α2 )⎠ = 0 (51)
j=1 n=1 α2 =1 j=1
Rate of changes from phase one to 2 equals rate of changes from phase two to 1 at
Maschine M2
⎛ ⎞
4 N
−1
2 4
a2 µ21 ⎝ cj ξj (N, B, 1) + cj ξj (n, α1 , 1)⎠
j=1 n=1 α1 =1 j=1
⎛ ⎞
4 N
−1
2
4
−µ22 ⎝ cj ξj (N, B, 2) + cj ξj (n, α1 , 2)⎠ = 0 (52)
j=1 n=1 α1 =1 j=1
Analysis of flow lines with Cox-2-distributed processing times 65
4
4
cj ξj (N, B, 1)+ cj ξj (N, B, 2) = 1 (53)
j=1 j=1
Note that the right hand side of the three of the four equations is zero. For this
reason, it is relatively painless to solve this linear system of equations in the four
unknowns cj , j = 1..4 numerically.
very quickly. For this reason decomposition approaches are frequently used to
analyze larger systems. The basic idea is to decompose a system with K machines
and K − 1 buffers into K − 1 two-machine systems with virtual machines that
mimic to an observer in the buffer the flow of material in and out of this buffer
as it would be seen in the corresponding buffer of the real system. We followed
the ideas presented in great detail in [21] to develop an iterative decomposition
algorithm to analyze flow lines with more than two machines. However, some
modifications were necessary which we will now briefly outline. While the models
analyzed in [21] assumed unreliable machines and consequently lead to so-called
interruption-of-flow- and resumption-of-flow-equations, we are studying a flow line
with reliable machines which cannot fail. The machines in our system, however,
change their phases of operation as described in Section 2. For this reason, we
derived the following three types of decomposition equations:
– Phase-One-to-Two(P1t2)-Equation: This type of equation deals with the
probability of the transition of the virtual machine from operating in its first
phase to its second.
– Phase-Two-to-One(P2t1)-Equation: This type of equation deals with the
probability of the transition of the virtual machine from operating in its second
phase to its first.
– Flow-Rate-Idle-Time(FRIT)-Equation: This is a type of equation which re-
lates the flow of material through a machine to its isolated production rate and
its probability of being blocked and starved. This type of equation has also been
used by Gershwin et al.
In the following we will briefly discuss the derivation of the parameters of the
virtual machines.
The key to the derivation of the P1t2- and P2t1-equations is the definition of
virtual machine states. We study a virtual two-machine line L(i) which is related
to the buffer between machines Mi and Mi+1 . The virtual machines of line L(i)
are Mu (i) (upstream of the buffer) and Md (i) (downstream of the buffer). We
want to determine the parameters au (i), µu1 (i), and µu2 (i) of the virtual machine
Mu (i) as well as the parameters ad (i), µd1 (i), and µd2 (i) of the virtual machine
Md (i) in order to be able to use our two-machine model in Section 3 to determine
performance measures for the flow line.
The upstream machine of a two-machine line is never starved (and the down-
stream machine is never blocked). We therefore assume that the virtual machine
Mu (i) is in phase one if the real machine Mi is processing a workpiece in phase
one or when it is waiting for the next workpiece:
{αu (i, t) = 1} iff {αi (t) = 1} or {αi (t) = S} (57)
Machine Mu (i) is in phase two if Mi is in phase two
{αu (i, t) = 2} iff {αi (t) = 2} (58)
and it is blocked if Mi is blocked:
{αu (i, t) = B} iff {αi (t) = B} (59)
Analysis of flow lines with Cox-2-distributed processing times 67
The definition of virtual machine states for machine Md (i) is symmetric: Ma-
chine Md (i) is in phase one if the machine Mi+1 downstream of the buffer number
i is in phase one or blocked:
{αd (i, t) = 1} iff {αi+1 (t) = 1} or {αi+1 (t) = B} (60)
It is in phase two if machine Mi+1 in the real system is in phase two
{αd (i, t) = 2} iff {αi+1 (t) = 2} (61)
and starved if Mi+1 is starved:
{αd (i, t) = S} iff {αi+1 (t) = S} (62)
In this derivation, the equation (68) holds because a part must complete its phase
two (which happens with probability µi2 δt) in order for machine Mi to reach states
1, S, or B.
The reasoning for machine Md (i) is analogous:
µd2 (i)δt = Prob[{αd (i, t + δt) = 1} or {αd (i, t + δt) = S}|
{αd (i, t) = 2}] (70)
This leads to the following result:
The probability of machine Mi in the real system not being blocked or starved
is approximated as follows:
prob[{ni (t) > 0} and {ni+1 (t) < Ni+1 }] ≈
1 − prob[{ni (t) = 0}] − prob[{ni+1 (t) = Ni+1 }] (73)
This is only an approximation since Mi can both blocked and starved. Now
these probabilities are unknown. However, we can use equations (22) and (24)
from the two-machine model to approximate these quantities if we decompose the
real system into a set of two-machine lines. This leads to the following equation:
1 − pS (i) − pB (i + 1)
P Ri ≈
(1 − ai ) µ1i1 + ai µ1i1 + µ1i2
P R2 (i) (1 − ad (i)) µd 1(i) + ad (i) µd 1(i) + µd 1(i)
= 1
1
2
P R1 (i + 1) (1 − au (i + 1)) µu (i+1) 1
+ au (i + 1) µu1(i) + 1
µu2 (i)
+ 1
1
Equations (64), (69) and (74) can be solved simultaneously for the parameters
µu1 (i), µu2 (i), and au (i) of the upstream machine Mu (i) related to line L(i) of
the decomposition to find:
au (i) = ai µd1 (i − 1)µi1 µd2 (i − 1)µi2 +
70 S. Helber
Note that again all three parameters of the virtual downstream machine Md (i)
are expressed in terms of parameters of the real machine Mi+1 or parameters or
performance measures of line L(i + 1).
| P R(l) − P R(l + 1) |
< 0.000001, l = 1, . . . , M − 1 (83)
P R(l)
holds, terminate the algorithm because the conservation of flow condition
is met by the result of the decomposition. Otherwise, goto step 2a.
(The algorithm also terminates if no convergence should be reached after 50
iterations.)
72 S. Helber
Case 2 3 4 7
Since the analysis of the two-machine model is exact, the only interesting question
with respect the the two-machine algorithm is if it is numerically stable. The results
reported in [8] indicated that their method tended to have difficulties with larger
buffer sizes. We did not observe such instabilities for the buffer sizes they considered
(up to 100 buffer spaces).
In order to evaluate the accuracy of the decomposition algorithm, we com-
pared it to results given in [9]. In these cases, the expected value and the squared
coefficient of variation of the processing time for each machine was given. The
Cox-2-distribution, however, has three parameters, so that one degree of freedom
is left. We used the so-called “balanced-mean” two-phase Coxian distribution [10,
p. 542] to match the problem data given in [9, p. 450-451].
Table 2 gives parameters and results for a three-stage system with general
service times and one buffer space between adjacent machines. (In the paper by
Buzacott et al. [9], the buffer space includes the workspace at the downstream
machine. If there is just one buffer space between two adjacent machines, our
general approach to determine steady-state probabilities as described in Section 3.4
cannot be applied since there are no “internal states” in terms of our two-machine
model. However, such a system has only 10 states which are all either lower or
upper boundary states. It is trivial to determine the steady-state probabilities and
performance measures for such a tiny system and we therefore simply added the
required algorithm to our decomposition approach in order to be able to deal with
two-machine lines with just one buffer space between adjacent machines.) The
entries “BLS-a” and “BLS-b” are related to two approaches described in [9], “Sim”
denotes the simulation results and “CoxDC” the results from our approach. For the
system in Table 2 our approach gives comparable results.
Analysis of flow lines with Cox-2-distributed processing times 73
Case 1 2 3 4
Case 1 2 3
For the four-stage systems in Table 3 with exponential service times our ap-
proach is more accurate than the procedures proposed in [9]. There is again just
one buffer space between adjacent machines in these cases.
Table 4 presents results for systems given by Altiok as cited in [9]. In all cases
there are two buffer spaces between adjacent machines except for Case 3 with 9
buffer spaces between machines 2 and 3 (and two between machines 1 and 2). For
these systems our approach outperforms the other methods.
74 S. Helber
Case 1 2 3 4 5 6 7 8
Buffer sizes 1 10 1 10 1 10 1 10
References
22. Gershwin SB, Berman O (1981) Analysis of transfer lines consisting of two unreliable
machines with random processing times and finite storage buffers. AIIE Transactions
13(1): 2–11
23. Gershwin SB, Schick I (1980) Continuous model of an unreliable two-stage material
flow system with a finite interstage buffer. Technical Report LIDS-R-1039, Mas-
sachusetts Institute of Technology, Cambridge, MA
24. Gershwin SB, Schick I (1983) Modeling and analysis of three-stage transfer lines with
unreliable machines and finite buffers. Operations Research 31(2): 354–380
25. Helber S (1998) Decomposition of unreliable assembly/dissassembly networks with
limited buffer capacity and random processing times. European Journal of Operational
Research 109(1): 24–42
26. Helber S (1999) Performance analysis of flow lines with non-linear flow of material.
Springer, Berlin Heidelberg New York
27. Hillier F, Boling RW (1967) Finite queues in series with exponential or Erlang service
times – a numerical approach. Operations Research 16: 286–303
28. Koenigsberg E (1959) Production lines and internal storage – a review. Management
Science 5: 410–433
29. Okamura K, Yamashina H (1977) Analysis of the effect of buffer storage capacity in
transfer line systems. AIEE Transactions 9: 127–135
30. Papadopoulus HT, Heavey C, Browne J (1993) Queueing theory in manufacturing
systems analysis and design. Chapman & Hall, London
31. Sastry BLN, Awate PG (1988) Analysis of a two-station flow line with machine pro-
cessing subject to inspection and rework. Opsearch 25: 89–97
32. Sevast’yanov BA (1962) Influence of storage bin capacity on the average standstill
time of a production line. Theory of Probability and Its Applications 7: 429–438
33. Wijngaard J (1979) The effect of interstage buffer storage on the output of two unreliable
production units in series, with different production rates. AIIE Transactions 11(1):
42–47
34. Yeralan S, Muth EJ (1987) A general model of a production line with intermediate
buffer and station breakdown. IIE Transactions 19(2): 130–139
35. Zimmern B (1956) Etudes de la propagation des arrêts aleatoires dans les chaines de
production. Review Statististical Applications 4: 85–104
Performance evaluation of production lines with finite
buffer capacity producing two different products
M. Colledani, A. Matta, and T. Tolio
Politecnico di Milano, Dipartimento di Meccanica, via Bonardi 9, 20133 Milano, Italy
(e-mail: [email protected])
Abstract. This paper presents an approximate analytical method for the perfor-
mance evaluation of a production line with finite buffer capacity, multiple failure
modes and multiple part types. This paper presents a solution to a class of problems
where flexible machines take different parts to process from distinct dedicated input
buffers and deposit produced parts into distinct dedicated output buffers with finite
capacity. This paper considers the case of two part types processed on the line, but
the method can be extended to the case of n part types. Also, the solution is devel-
oped for deterministic processing times of the machines which are all identical and
are assumed to be scaled to unity. The approach however is amenable of extension
to the case of inhomogeneous deterministic processing times. The proposed method
is based on the approximate evaluation of the performance of the k-machine line
by the evaluation of 2(k-1) two-machine lines. An algorithm inspired by the DDX
algorithm has been developed and the validation of the method has been carried
out by means of testing and comparison with simulation.
1 Introduction
distinct dedicated input buffers and deposit produced parts into distinct dedicated
output buffers with finite capacity. By dedicated input and output buffers we mean
buffers that can store only one part type. The proposed solution is developed for
the case of two part types, however the approach is amenable to extension to the
multiple part type case. Also, the solution is developed for deterministic processing
times of the machines which are all identical and are assumed to be scaled to unity.
The approach however is amenable to extension to the case of inhomogeneous
deterministic processing times.
A typical system of the proposed class is represented in Figure 1. In this case
machines M1 , M2 , M3 , M6 and M7 are dedicated machines i.e. they can produce
only one part type. On the contrary, machines M4 and M5 are flexible machines
and can process both part types. The selection of which type of part to produce
depends on the state of the system and on a dispatching rule. If the upstream buffer
of one part type is empty or the downstream buffer is full, the machine will produce
the other part type. If both the part types are either blocked or starved, the machine
will not produce. If both the parts can be produced, than the machine will produce
part type A with probability αiA and part type B with probability αiB . In the paper,
systems formed only by flexible machines will be considered, but the proposed
method in principle can be extended to the case of systems in which both flexible
and dedicated machines are present, such that in Figure 1.
It is important to notice that the proposed system is quite different from assem-
bly/disassembly systems [2, 1]. Indeed in assembly/disassembly systems assembly
machines take contemporarily different parts from different input buffers to produce
a single subassembly while disassembly machines from one subassembly produce
contemporarily different components that are put into different buffers. In either
case there is no selection of components to work on but they are all contemporarily
involved in the process. On the contrary, in the described system a flexible machine
selects a single component to work on.
The proposed system is also different from fork and join networks [3, 4]. Indeed
in fork and join networks each machine can either take the input from different
buffers to produce an undifferentiated product or take in as input an undifferenti-
ated product and place it after processing in different buffers. On the contrary in the
described system flexible machines take in as input different parts from different
buffers and produce different products placed in the corresponding buffers. In other
words, the identity of the part is not lost within the machine.
The problem presented in this paper has been originally stated by S.B. Gershwin
and addressed by Nemec [6]. The original statement however considers a priority
rule between the parts and therefore when both part types can be produced, the part
Performance evaluation of two part type lines 79
type with the highest priority is selected. This would correspond in our statement
to the case of αA = 1, αB = 0. The solution approach adopted in Nemec is heavily
dependent on the original problem statement because parts are treated differently
depending on the priority. On the contrary, in the proposed approach all the parts
are considered in the same way.
It is interesting to consider the fact that the described problem, which has been
inspired by automated production system, is similar to other relevant problems
that can be addressed with the same methodology. In particular it is interesting
to consider the case of production networks where different enterprize cooperate
to produce complex products. In this case each enterprize of the network can be
modelled as a flexible machine while input and output storages can be modelled as
buffers.
– If for a given machine both the upstream buffers are not empty and both the
downstream buffers are not full the machine will produce a part of type A with
probability αiA and a part of type B with probability αiB (αiA + αiB = 1).
– Operation dependent failures are assumed, i.e. machines can only fail if they
are not down, not blocked or starved for both part type at the same time, and
not contemporarily starved for part type A(B) and blocked for part type B(A).
– A given machine Mi can fail in Fi different failure modes.
– An operational machine can fail in only one of its failure modes.
– Mean time to failure (MTTF) and mean time to repair (MTTR) are geometrically
distributed with average values of 1/pi,j and 1/ri,j respectively (i = 1, ..., K
; j = 1, ..., Fi ).
The method evaluates the performance measures of the systems described in the
previous section by using a generalization of the decomposition technique proposed
in [7]. The method can also be used in principle with the decomposition technique
proposed in [8]. The analyzed system is decomposed into 2(K − 1) sets of two-
machine lines that together represent the behavior of the system. Each two-machine
line (building block) models the flow of one of the two part types in the system
(Fig. 2). In other words the method creates a two-machine line for each buffer of
the original line; each building block is composed of two pseudo machines and
one intermediary buffer. The upstream pseudo machine represents the behavior
of the portion of the system that precedes, in the original line, the corresponding
buffer considered in the building block. In the same way, the downstream pseudo
machine represents the behavior of the portion of the system that follows, in the
original line, the corresponding buffer considered in the building block. The idea is
to analyze simple building blocks, easy to study with existing techniques, instead of
the complex original system. In such a way the complexity of the analysis is reduced
to study several two-machine lines instead of a long production line. However, the
different two-machine lines are not independent and have to be analyzed by means
of decomposition equations. To do this, the parameters of the pseudo machines are
calculated so that the flow of parts through the buffers of the decomposed systems
closely matches the flow through the corresponding buffers of the original line.
Therefore, for buffers BiA and BiB of the original line, two building blocks
(Fig. 2) are created. The first building block models the flow of type A parts and
is composed of the upstream pseudo-machine M U (A) (i), the downstream pseudo-
machine M D(A) (i) and the buffer B A (i). These two pseudo-machines together
with the buffer form the building block a(i). The second building block mod-
els the flow of type B parts and is composed of the upstream pseudo-machine
M U (B) (i), the downstream pseudo-machine M D(B) (i) and the buffer B B (i).
These two pseudo-machines together with the buffer form the building block b(i).
To model the interruptions of flow through the buffers of the original line, fail-
ure probabilities of different modes are associated to each pseudo-machine. In the
following we will consider the case of the upstream pseudo-machines. A similar
reasoning applies to the downstream pseudo-machines.
Performance evaluation of two part type lines 81
The Markov chain of the machine Mi of the original line is presented in Figure 4.
To simplify the picture, all the states of the same type are grouped into a unique
aggregate state, without considering different failure modes. Each aggregate state
is defined by two state indicators, one referred to part type A and the other referred
to part type B. Each state indicator can assume four values that are, for part type
A: working (W A ), down in local mode (RA ), starved (S A ), if the upstream buffer
dedicated to the storage of product A is empty, and blocked (B A ), if the downstream
buffer dedicated to the storage of part type A is full. The same can be written for
part type B. In total there are 16 possible aggregate states. For each aggregate state,
the probability is obtained by adding up the probabilities of all the states of the
same type. For example, the probability of the aggregate state named W A RB has
been obtained by adding up all the probabilities of the W A RiB states, one for each
Fi
failure mode of the machine, i.e. π(W A RB ) = π(W A RiB ).
i=1
Obviously, while in the picture we consider the aggregate states, in writing the
equations it is important to distinguish all the different failure modes, to correctly
evaluate the state probabilities. It must be noticed that machine Mi cannot be both
working a part type while being down in local mode for the other part type, therefore
84 M. Colledani et al.
the states of type W A RB , RA W B are not feasible and are not represented in the
picture. Also the aggregate state RA RB represents a situation where the machine
is down and therefore cannot produce either A or B. We call this aggregate state
pure local down state and we rename it R. Finally the state W A W B represents a
state where, for both part types, no local failures, starvation or blocking are present,
therefore the original machine can produce either A or B.
In the following, some key characteristics of the Markov chain of the machine
Mi (Fig. 4) are discussed:
As already mentioned, in Figure 4, to simplify the diagram, all the states of the same
type are grouped into a unique state without considering different failure modes.
The probability of these 14 aggregate states is therefore the sum of the probabilities
of the disaggregated states considering all the failure modes. It must be noticed
that, in the Markov chain of the original machine, the competition failure is not
considered, because this machine is able to produce both part types, as assumed in
the previous section.
It must also be noticed that in this Markov chain not all the transition prob-
S(A) S(B) B(A) B(B)
abilities are known. Indeed the values of pj,fj , pj,fj and pk,fk , pk,fk cannot
be derived directly from the original line and therefore they must be found using
appropriate equations. In the following the 14 sets of equations required to eval-
uate the probabilities of the various states, plus the equations required to evaluate
the unknown transition probabilities are provided. The approach used to derive the
following equations deals with the idea that in the decomposed lines, as they have
been defined, each building block mimics the flow of material through one buffer of
the original line. Therefore, the macro state probabilities of the machine Mi must
be coherent with probabilities of building blocks a(i − 1), b(i − 1), a(i) and b(i).
Performance evaluation of two part type lines 85
In the Markov chain, the sum of all the state probabilities must be equal to unity.
Therefore it is possible to write the normalization equation for the model:
All States = 1 (1)
Since buffers B A (i−1) and B B (i−1) of the decomposed lines are equal to buffers
A B
Bi−1 and Bi−1 of the original line, the probability of starvation of machine Mi
has to be equal to the one that derives from upstream building blocks a(i − 1) and
b(i − 1).
i−1 F
k +1 K F
k +1
A
π(Sj,f SB )
j k,fk
+ A
π(Sj,f j
B
Bk,f k
) (2)
k=1 fk =1 k=i+1 fk =1
Fi
+π(W B Sj,f
A
j
)+ B
π(Ri,f S A ) = P sA
i j,fj j,fj (i − 1)
fi =1
j = 1, ..., i − 1, fj = 1, ..., Fj + 1
i−1 F
j +1 K F
j +1
A
π(Sj,f SB )
j k,fk
+ A
π(Bj,f SB )
j k,fk
(3)
j=1 fj =1 j=i+1 fj =1
Fi
+π(W A Sk,f
B
k
)+ A
π(Ri,f S B ) = P sB
i k,fk k,fk (i − 1)
fi =1
k = 1...i − 1, fk = 1...Fk + 1
Where π(X) is the steady state probability of state X.
Since buffers B A (i) and B B (i) of the decomposed lines are equal to buffers BiA
and BiB of the original line, the probability of blocking of machine Mi has to be
equal to that of downstream building blocks a(i) and b(i).
i−1 F
k +1 K F
k +1
B A A B
π(Sk,f k
Bj,f j
)+ π(Bj,f j
Bk,f k
) (4)
k=1 fk =1 k=i+1 fk =1
Fi
B A
+ π(Ri,f i
Bj,f j
) + π(W B Bj,f
A
j
) = P bA
j,fj (i)
fi =1
j = i + 1...K; fj = 1...Fj + 1
i−1 F
j +1 K F
j +1
A B A B
π(Sj,f j
Bk,f k
)+ π(Bj,f j
Bk,f k
) (5)
j=1 fj =1 j=i+1 fj =1
Fi
A B
+ π(Ri,f i
Bk,f k
) + π(W A Bk,f
B
k
) = P bB
k,fk (i)
fi =1
k = i + 1...K; fk = 1...Fk + 1
Considering the states of type RA S B , RA B B , RB S A , RB B A , we can
write node equations balancing the probability of entering these states with the
probability of exiting the same states.
S(B)
π(W A Sj,f
B
j
)pi,fi (1 − rj,fj ) (6)
86 M. Colledani et al.
A S(B) S(B)
= π(Ri,f S B )rj,fj + π(Ri,f
i j,fj
A
S B )ri,fi (1 − rj,fj )
i j,fj
Fj
i−1
B
ri,fi (π(Ri,fi ) + π(Ri,f SA )
i j,fj
(14)
j=1 fj =1
Fk
i−1 K F
k +1
A
+ π(Ri,f SB )
i k,fk
+ A
π(Ri,f i
B
Bk,f k
)+
k=1 fk =1 k=i+1 fk =1
Performance evaluation of two part type lines 87
K F
j +1 i−1 F
j +1
B A A B
+ π(Ri,f i
Bj,f j
)) = pi,fi (π(W W ) + π(W B Sj,f
A
j
)+
j=i+1 fj =1 j=1 f1j =1
i−1 F
k +1 K F
j +1
+ π(W A Sk,f
B
k
)+ π(W B Bj,f
A
j
)
k=1 fk =1 j=i+1 fj =1
K F
k +1
+ π(W A Bk,f
B
k
)) fi = 1...Fi
k=i+1 fk =1
S(B)
P sB
j,fj (i − 1) S(B)
pj,fj = rj,fj
E B (i − 1)
S(A)
P sA
j,fj (i − 1) S(A)
pj,fj = r j = 1...i − 1; fj = 1...Fj + 1 (15)
E A (i − 1) j,fj
B(B) P bB
k,fk (i) B(B)
pk,fk = r
E B (i) k,fk
B(A) P bA
k,fk (i) B(A)
pk,fk = r k = i + 1...K; fk = 1...Fk + 1 (16)
E A (i) k,fk
where E A (i) and E B (i) are the average production rates of the building blocks
a(i) and b(i) respectively, i = 1, ..., K − 1.
Repair probabilities of these failures are supposed to be equal to those local failures
of the machines of the original line responsible for starvation and blocking.
Considering the expression of the efficiency in isolation ei of the flexible machine
Mi of the original line and the expression of the efficiency Ei of that machine
related to the presence of buffers, it is simple to demonstrate that from normalization
equation (1) it is possible to derive two conservation of flow equations, one for each
product type:
i−1 F
j +1
+ π(W A Bk,f
B
k
) = E A (i − 1) (17)
k=i+1 fk =1
i−1 F
j +1
+ π(W B Bk,f
A
k
) = E B (i − 1) (18)
k=i+1 fk =1
88 M. Colledani et al.
Once the probabilities of the various states of the flexible machine Mi of the orig-
inal line are obtained, it is possible to build two models (Fig. 5), one for each
pseudo-machine M U (A) (i) and M U (B) (i). This results in two five state models.
The probability of each state is obtained by adding up the values calculated in the
previous Markov chain.
Fig. 5. Five state models for pseudo-machines M U (A) (i) and M U (B) (i)
+ π(W A Bk,f
B
k
) (19)
k=i+1 fk =1
i−1 F
k +1
U (A) A
π(Sj,fj ) = π(Sj,f SB )
j k,fk
k=1 fk =1
K F
k +1 Fi
A B B A B
+ π(Sj,f j
Bk,f k
) + π(W Sj,f j
) + π(Ri,f S A (20)
i j,fj
)
k=i+1 fk =1 fi =1
j = 1, ..., i − 1 fj = 1, ..., Fj
Fk
i−1
U (A) A
π(Ri,fi ) = π(Ri,fi ) + π(Ri,f SB )
i k,fk
k=1 fk =1
K F
k +1
A B
+ π(Ri,f i
Bk,f k
) fi = 1, ..., Fi (21)
k=i+1 fk =1
i−1 F
k +1 K F
k +1
U (A) B A A B
π(Bj,fj ) = π(Sk,f k
Bj,f j
) + π(Bj,f j
Bk,f k
)
k=1 fk =1 k=i+1 fk =1
Performance evaluation of two part type lines 89
Fi
B A
+ π(Ri,f i
Bj,f j
) + π(W B Bj,f
A
j
) (22)
fi =1
j = i + 1, ..., K fj = 1, ..., Fj
B
π(W ) = π(W W )αiB
A B
(23)
U (A)
The state W represents the state in which the original machine Mi works
part type A, that is the up state for the pseudo-machine M U (A) (i). The state W B
represents the state in which the original machine Mi works part type B even if it
could work both part type, that is a down state for the pseudo-machine M U (A) (i).
The same can be written for pseudo machine M U (B) (i).
Having these two approximate models, one for each pseudo-machine, and knowing
all the state probabilities, it is possible to calculate new local failure probabilities
and remote failure probabilities for the two pseudo-machines. It must be noticed
that, for these machines, the probability of entering into local failure is higher than
the one of the corresponding machine in the original line because we take into
account the probability of failing while producing the other part type.
We can evaluate local failure parameters using balancing equation of nodes RU (A)
and RU (B) :
i−1 Fj
U (A) K Fj B(A)
Where pU (A) (i) = j=1 fj =1 pj,fj (i) + j=i+1 fj =1 pj,fj (i)
fi U (A)
+ i=1 pi,fi is the sum of all the other failure probabilities of the pseudo-
machine M U (A) (i). Now, using the equation (28), the probability of repair for the
competition failure mode can be evaluated as follows:
and
U (B) π(W U (B) ) U (B)
rFi +1 (i) = p (i) (31)
π(W A ) Fi +1
Once local, remote and competition failure probabilities are evaluated they can be
used within the building blocks a(i) and b(i).
5 Algorithm
U (A/B) U (A/B)
pj,fj (i) = λ rj,fj (i) = rj,fj j = 1, ..., i − 1; fj = 1, ..., Fj (34)
M D(A/B) (i − 1):
D(A/B) D(A/B)
pi,fi (i − 1) = pi,fi ri,fi (i − 1) = ri,fi (35)
i = 2, ..., k fi = 1, .., Fi
D(A) D(A)
pi,Fi +1 (i − 1) = λ ri,Fi +1 (i − 1) = αiA
D(B) D(B)
pi,Fi +1 (i − 1) = λ ri,Fi +1 (i − 1) = αiB (36)
B(A/B) B(A/B)
pj,fj (i − 1) = λ rj,fj (i − 1) = rj,fj (37)
j = i + 1, ..., K; fj = 1, ..., Fj
2. Step 1. For i = 1, ..., K − 1: failure parameters of machines M U (A) (i) and
M U (B) (i) are evaluated:
– Unknown transition probabilities are calculated using equations (15).
– Evaluation of all the state probabilities of the flexible machine Mi by using
the linear system formed by equations (1) to (14). Blocking probabilities
and transitions to blocking states are derived from previous iterations of
the algorithm and they are equal to remote failures of downstream pseudo-
machines M D(A) (i − 1) and M D(B) (i − 1), in case of i > 1, while for
i = 1 they can be evaluated using equations (16).
– Distribution of the calculated probabilities into the two pseudo-machine
models of Figure 4, using equations (19) to (23), for both part types.
– Evaluation of new local failures using equations (24), (25).
– Calculation of remote failures using equations (26), (27).
– Evaluation of competition failures using equations (29), (30), (31) and
(32).
– Insertion of calculated failure parameters into upstream pseudo-machines
of building blocks a(i) and b(i).
– Evaluation of average throughput, probabilities of blocking and probabil-
ities of starvation of blocks a(i) and b(i) using the building block solution
proposed in [5].
3. Step 2. For i = K, ..., 2: failure parameters of machines M D(A) (i − 1) and
M D(B) (i − 1) are evaluated:
– Unknown transition probabilities are calculated using equations (16).
– Evaluation of all the state probabilities of the flexible machine Mi by using
the linear system formed by equations (1) to (14). Starvation probabilities
and transitions to starvation states are derived from previous iterations of
the algorithm and they are equal to remote failures of upstream pseudo-
machines M U (A) (i) and M U (B) (i), in case of i < K, while for i = K
they can be evaluated using equations (15).
– Distribution of the calculated probabilities into the two pseudo-machine
models similar to those of Figure 5, using equations similar to (19) to (23).
– Evaluation of new local failures using equations similar to (24), (25).
– Calculation of remote failures using equations similar to (26), (27).
92 M. Colledani et al.
6 Numerical results
In order to show the accuracy of the new analytical method developed (method
CMT) a set of numerical tests has been carried out comparing the analytical re-
sults with those obtained running simulation experiments. More than 150 lines
producing two products have been analyzed using the proposed method with prob-
ability of failure and repair varying in the following ranges: 0 < pi,fi < 0.25 and
0 < ri,fi < 0.8. In particular, systems with three-machines/four-buffers (Table 1),
four machines/six buffers (Table 2), five machines/eight buffers (Table 3) and six
machines/ten buffers (Table 4) with one failure parameter for each machine have
been studied and a sampling of results are reported in the following tables. Also,
systems with machines characterized by multiple failure modes are studied and the
results are reported in Table 5. For each simulation experiment 10 replications have
been performed, with a warm-up period of 105 time units followed by simulation
period of 106 time units. Average throughput has been evaluated with a 95% half
confidence interval of 0.0009 as maximum value. Average buffer level has been
evaluated with a 95% half confidence interval of 0.08 as maximum value.
In all the tables, for each analyzed case, failure and repair probabilities of the
machines are reported on the left, together with buffer capacities and the αiA pa-
rameters, that are equal for all the machines in the line. The average production
rates of the line, calculated with the proposed method and with simulation, are
94 M. Colledani et al.
reported on the right and the error between the evaluations is estimated using the
following equations:
A
ESIM − ECM A
A T
∆%E = A
ESIM · 100
B
E − EB
∆%E B = SIM B CM T · 100
ESIM
As it can be seen by the results provided in this section and by the summary of
results in Table 6, the algorithm has proven to be reliable and accurate in all the
tested cases; indeed the maximum error in throughput evaluation is around 3% and
a high percentage of cases have an error in throughput evaluation lower than 1%.
In Table 7 the error in the evaluation of the average buffer level is reported
for the six machine cases of Table 4. The error between evaluations has been
Performance evaluation of two part type lines 95
Fig. 6. Throughput evaluation with αiA equal for all the machines of the line and variable
A
(ni )SIM − (nA
i )CM T
∆%(nA
i ) = NiA · 100
B
(ni )SIM − (nB
i )CM T
B
∆%(ni ) =
NiB · 100
96 M. Colledani et al.
estimates average throughput of the test line as it is shown in (Fig. 6) and (Table 8)
for a wide range of variability of parameter αiA .
As it can be seen by the results provided in this section, the algorithm has proven
to be reliable and accurate in all the tested cases with αiA and αiB parameters equal
for all the machines of the line.
7 Conclusions
References
Abstract. The paper addresses the problem of fully using buffer spaces in man-
ufacturing flow lines. The idea is to exploit recent technological devices to move
in reasonable times pieces from a machine to a common buffer area of the system
and vice versa. In such a way machines can avoid their blocking since they can
send pieces to the shared buffer area. The introduction of the buffer area shared
by all machines of the system leads to an increase of production rate as demon-
strated by simulation experiments. Also, a preliminary economic evaluation on a
real case has been carried out to estimate the profitability of the system comparing
the increase of production rate, obtained with the new system architecture, with the
related additional cost.
1 Introduction
can have different speeds and may be affected by random failures, the part flow
can be interrupted at a certain point of the system causing blocking and starvation
of machines. In particular, blocking in the line occurs when at least one machine
cannot move the parts just worked (BAS, Blocking After Service) or still to work
(BBS, Blocking Before Service) to the next station. In flow lines the blocking of a
machine can be caused only by a long processing time or a failure of a downstream
machine. Analogously, starvation occurs when one or more machines cannot be
operational because they have no input part to work; in this case the machine can-
not work and it is said to be starved. In flow lines the starvation of a machine can
be caused only by a long processing time or a failure of an upstream machine.
Therefore, in flow lines the state of a machine affects the rest of the system because
of blocking and starvation phenomena that propagate upstream and downstream
respectively the source of flow interruption in the line. If there is no area where to
store pieces between two adjacent machines, the behavior of machines is strongly
correlated.
In order to decrease blocking and starvation phenomena in flow lines, buffers be-
tween two adjacent machines are normally included to decouple the machines
behavior. Indeed, buffers allow to adsorb the impact of a failure or a long process-
ing time because (a) the presence of parts in buffers decreases the starvation of
machines and (b) the possibility of storing parts in buffers decreases the blocking
of machines. Therefore, production rate of flow lines is also a function of buffer
capacities; more precisely, production rate is a monotone positive function of the
total buffer capacity of the system. Refer to [5, 7] for a list of works focused on the
properties of production rate in flow lines as a function of the buffer size.
Traditionally, flow lines have been deeply investigated in literature. Re-
searchers’ efforts have been devoted to develop new models for evaluating the
performance of flow lines and for optimizing their design and management in shop
floors. Operations research techniques like simulation and analytical methods have
been widely used to estimate system performance parameters such as throughput
and work in process level. Performance evaluation models are currently used in
configuration algorithms for finding the optimal design of flow lines taking into
account the total investment cost, operative cost and production rate of the system.
In synthesis, academic innovation has been mainly focused on the development of
performance evaluation and optimization methods of flow lines without entering
into several mechanical details. See also the review [1] of Dallery and Gershwin
on a detailed view of performance evaluation models for flow lines and an updated
recent state of the art on optimization techniques applied in practice [8]. Indeed,
most of works is at system level as they deal with optimization of macro variables
such as number of machines in the line, buffer capacities and machines’ speed and
Automated flow lines with shared buffer 101
efficiency. On the other hand, engineers of firms have had to face the complex-
ity due to the fact that flow lines are designed in practice with all their mechanical
components. Innovation from builders of manufacturing flow lines has been mainly
dedicated to increase machines reliability and to reduce system costs by improv-
ing the design of specific mechanical components such as feed drives, spindles,
transporters, etc.
Therefore, advancements in flow line evolution do not regard the main philos-
ophy of the system. Parts are loaded into the system at the first machine and, after
having been processed, they are moved into the first buffer waiting for the avail-
ability of the second machine. Blocking phenomena is limited by buffers, larger
is their capacity and higher the throughput of the line is. However, buffers in flow
lines are dedicated to machines; this characteristic implies that a buffer can con-
tain only pieces worked by the immediately upstream machine. Therefore, when
a long failure occurs at a machine of the line, the portion of the system upstream
the failed machine is blocked but upstream machines continue to work until their
corresponding buffers are full. On the other hand, the portion of the system down-
stream the failed machine is starved because downstream machines cannot work
since they do not have any piece to work. In that case the buffer area downstream
the failed machine cannot be used to store parts worked by machines that are up-
stream the failed machine since empty buffers are dedicated and cannot be used for
pieces coming out from other machines. It appears that buffer spaces are not fully
exploited when needed. The problem of properly using all the available space in
flow lines represents the argument of this paper.
2.1 Motivation
The paper presents a new concept of manufacturing flow line characterized by two
different types of buffers: traditional dedicated buffers and a common buffer shared
by all the machines of the system. The common buffer allows to store pieces at
any point of the system thus increasing the buffer capacity of each machine (see
Fig. 2). The main advantage is related to the fact that wherever an interruption of
flow is in the system, the common shared buffer can be used by all machines. As
a consequence, blocking of machines should be lower than that of classical flow
lines thus allowing an increase of production rate at constant total buffer capacity.
However, profitability of the new system architecture depends on costs incurred
for the additional shared buffer. Traditionally the main goal in the design phase
of flow lines is to find the system configuration at minimum costs constrained
Bs , are full. The presence of shared buffer decreases blocking phenomena in the
flow line: if the dedicated buffer is full, pieces worked by machine Mi can be moved
to the shared buffer until the part flow resumes at machine Mi+1 and the level of
buffer Bi decreases. In more detail, a part which cannot be stored in a dedicated
buffer stays in the shared buffer until a place in the dedicated buffer becomes
available. The way in which parts in the shared buffer are positioned depends
on the technology used and the management rules adopted. If the shared buffer
consists of a simple conveyor on which parts flow until a new space is available at
dedicated buffers, the ordering of parts depends on the their entering sequence in
the conveyor. If the shared buffer consists of a series of racks, the ordering of parts
depends on the particular management rule adopted; in this case it is necessary to
have a resource like a robot or a carrier that takes parts from machines and put
them on racks. Tempelmeier and Kuhn describes different mechanisms in the case
of Flexible Manufacturing Systems (FMS) with central buffer [9].
A certain amount of time is necessary for physically moving parts from a ded-
icated buffer area to the shared buffer area and vice versa. Therefore, if blocking
decreases with the introduction of the shared buffer the starvation increases [9]. In
the remainder of the paper we call this time the travel time denoted with tt . The
profitability of the system depends on the value of travel time and its impact on
system performance. If the travel time is reasonably small, then the penalty time
incurred for using the shared buffer does not deeply decrease the system perfor-
mance since, after the resumption of flow, the time spent by parts for going from
the shared buffer area to the dedicated one is hidden, i.e. covered by the pieces
already present in the dedicated area and processed in the meanwhile by machine
Mi+1 . If the travel time is large, then the penalty time incurred for using the shared
buffer can strongly decrease the system performance since machines are frequently
starved. The increase of starvation as a consequence of the transport time from/to
the shared buffer has been previously described by Tempelmeier and Kuhn [9] in
the analysis of a special FMS configured as a flexible flow line. The next section
reports a numerical analysis for assessing the productivity of flow lines with shared
buffer in different situations.
3 Numerical evaluation
The objective of the section is to evaluate the gain in terms of productivity due to the
introduction of shared buffers in production lines. To do this, the experimentation
has been carried out by simulating flow lines on simple test cases, created ad hoc
to understand the system behavior in different situations (Sect. 3.1), and on a real
flow line (Sect. 3.2).
We consider a closed flow line composed of five machines, each one with a finite
buffer capacity immediately downstream. Machines are unreliable and character-
ized by the same type of failure. Failures are time dependent and do not depend
104 A. Matta et al.
After that, all assumptions required by the analysis of variance have been satisfied
and the main results are now presented. In particular the Anderson-Darling and
Bartlett tests at 95 percent confidence level have been used to test normality and
variance homogeneity of residuals respectively, the independence was assured by
the randomized execution of experiments and different seeds used for generating
pseudo-random numbers. Results from ANOVA are reported in Tables 2 and 3 where
significant factors and interactions are recognizable: a source is significant on the
system’s efficiency if the p-value in the last column is lower than the Bonferroni’s
alpha family (chosen equal to 0.05) divided by the number of executed statistical
tests (i.e. 15 in this experiment).
As far as the main effects are concerned, the number of pallets that circulate
in the system, the portion of dedicated buffers and the total buffer capacity are
significant for both the populations with different cycle times. The main effect of a
factor is the average change in the response due to moving the factor from its low
level to its high level [6]; this average is taken over all combinations of the factor
levels in the design. A first conclusion is that in the analyzed system a travel time
value equal to 0 or 30s is not relevant for the system’s efficiency due to the values
chosen for the levels. However increasing to a threshold value, greater than 30s,
the travel time leads to bad performance; we will see at the end of the paragraph
the threshold values after which the travel time becomes significant. The factor
α is significant for both populations and, furthermore, it is possible to conclude,
by comparing with the Tukey’s method the two levels, that the analyzed closed
flow line with shared buffer has statistically an efficiency superior than that of the
analyzed traditional flow line. Notice that the difference of efficiency in the two
levels is around 5% for tc = 5s and 1% for tc = 60s. The significance of the
number of pallets and the total buffer capacity is a well known result in literature
[2–4].
As far as the interactions effects are concerned, it is possible to state that two-
way interactions between P , α and NT OT are statistically relevant (see also Fig. 3
and 4). A two-way interaction is significant if the combined variation of the two
factors has a relevant effect on the response. The interaction between P and α
shows that the system performance decreases when the number of pallets is high
and all buffers are dedicated: in this case the blocking of machines is frequent
and it deeply affects the line efficiency. Notice that the system with shared buffer
has approximately the same efficiency value independently by how many pallet
circulate in the line. On the contrary, performance decreases in the traditional system
when the number of pallets is augmented. The interaction between NT OT and α
shows that system performance decreases when the total buffer capacity is low and
all buffers are dedicated: in this case the blocking of machines is frequent due to
the contemporary reduced and dedicated buffers capacity. The interaction between
NT OT and P is known in literature [3, 9] and we do not comment more. The triple
interaction among P , α and NT OT results to be significant only for cycle time
equal to 60 s.
The same system has been simulated also for a wider set of values to better
understand the effect of the shared buffer on the system performance. Figure 5
shows the average throughput for different sharing levels of the central buffer when
106 A. Matta et al.
the travel time is equal to 5 s. When the number of pallets is small the shared buffer is
never used and the different systems have the same performance. When the number
of pallets increases, the starvation decreases and the throughput increases; however
as the number of pallets increases the blocking occurs more frequently and the
systems with shared buffer perform better than the traditional one (i.e. α = 1). In
more detail higher the sharing percentage is and better is the performance. After a
certain value of pallets in the system the blocking penalizes the system performance
108 A. Matta et al.
Fig. 5. Test case: average production rate (±1.2 part/hour) vs P when NT OT = 50 and
tt = 5s for different values of α
Fig. 6. Test case: average production rate (±1.2 part/hour) vs tt when NT OT = 50 and
P = 30 for different values of α
and the throughput decreases [2, 3, 9]. This inversion value is larger in systems
with shared buffer than in traditional systems. In particular the inversion point
increases as the percentage sharing of buffers increases. Thus, in order to increase
the throughput the system’s user could move a portion of the buffer capacity from
dedicated to buffer and contemporary to increase the number of pallets.
Figure 6 shows the effect of the travel time on the average throughput. The
system performance stays stable for values of the parameter travel time inferior to
Automated flow lines with shared buffer 109
Fig. 7. Test case: average production rate (±1.2 part/hour) vs tt when NT OT = 50 and
α = 0.75 for different values of P
a threshold value and deteriorates after it. In this experiment the threshold value
is equal to 40 s, 25 s and 15 s for values of α equal to 0.75, 0.5 and 0.25 respec-
tively when the number of pallets is 30. The threshold value of the travel time
decreases as the percentage of shared buffer increases because there are less pallets
in the dedicated buffers and starvation occurs more frequently. This effect could be
compensated by increasing the number of pallets.
Figure 7 shows the effect of the travel time for the system with α = 0.5 and
different values of pallets. Notice that the loss of production after the threshold
value of the travel time is larger for small number of pallets. It is worthwhile to
notice that the results reported in this section are valid for closed flow lines. Open
flow lines are more difficult to manage due to the large number of parts which may
enter from the first machine. Indeed, if there is no limit to the number of parts
entering into the system and the first machine is very efficient, it happens that all
the parts just entered and processed by the first machine fills the shared buffer thus
limiting the possibility to the other machines of recurring to the shared buffer. Thus,
specific rules for managing the entering of parts should be designed.
during the production. Machines are unreliable and characterized by different types
of failures. In particular machines M1 , M2 and M5 can fail in three different ways
while M3 and M4 in only one way. MTTF and MTTR for each failure type are
exponentially distributed with values as reported in Table 4 and calculated by the
firm. As in the real system, a machine can fail only when it is occupied by a part. The
first failure type of each machine models mechanical and electronic failures in an
aggregated way, and the second and third failure types of M1 , M2 and M5 model the
emptying of component containers. Table 4 reports the failure parameters and the
deterministic processing rates of machines. The BAS control rule is considered,
that is machines may enter in a blocking state only after the completion of the
process. A physical constraint in the lay-out does not allow changes in the portion
of the system between M5 and M1 , i.e. the buffer B5 is dedicated and cannot be
modified in its capacity. The lay-out of the system is shown in Figure 8. The real
system already uses in the traditional way flexible transport modules for moving
parts through the line at a constant speed of 17.6 m/min.
Among a large set of feasible solutions, two alternative reasonable systems
with shared buffer are considered in the comparison with the real one. The first
alternative has a shared buffer, located at the center of the line, in addition to
the dedicated buffers of the real system. The increase of total buffer capacity is
around 31% corresponding to an increase of approximately 44 kEuro of the total
investment cost (this value has been estimated on the basis of additional conveyors,
sensors, engines and control system). The second alternative has been designed
Automated flow lines with shared buffer 111
with investment cost equal to that of the real line; the total buffer capacity is lower
than that of the real line because of the additional costs of sensors and engines.
Total buffer capacity is reported in Table 5 while the lay-outs of alternatives with
shared buffer are shown in Figures 9 and 10.
In the proposed alternatives each machine is blocked only if its dedicated buffer
and the common buffer are full. The mechanism of input/output into/from the
shared buffer is now described referring to Figure 11. Let us consider the portion
of the system between machines M1 and M2 . Before the machine M1 releases
a processed part, the system controls the availability of space in the portion of
conveyor between M1 and M2 , i.e. in the dedicated buffer with size N1 . If there
is space in the dedicated buffer the machine releases the part which then moves
towards machine M2 , otherwise the system controls if the part can be introduced in
the shared buffer. If there is space available in the shared buffer the machine releases
the part that enters in the shared buffer, otherwise the machine is blocked until a
new space becomes available in the dedicated buffer or shared buffer. This reaction
Table 4. Real case: processing rates [part/min] and MTTFs and MTTRs [min] of machines
Table 5. Real case: buffer capacities of real system and alternatives with shared buffer
Table 6. Real case: comparison between real system and alternatives with shared buffer
Fig. 13. Real case: average production rate vs number of pallets (±4 part/h)
the alternative 2. It can be noticed how the blocking of machines decreases from
the real system to the system with the shared buffer and starvation increases due to
the fact that parts circulate in the shared buffer. However the disadvantages due to
the increase of starvation do not compensate the advantages due to the reduction
of blocking in the system.
A long term investment analysis must consider the Net Present Value (NPV)
related to the investment, i.e. the sum of all discounted cash flows during the life
of the system. The NPV considers the initial investment cost (fundamentally ma-
chines, buffers), the future discounted cash flows during the normal running of the
system (revenues and production costs) and the residual value of the system after
the planning horizon of the investment. In the investment analysis only the discrim-
inating voices have to be considered. In this case the three alternative systems differ
in the buffer capacity since machines remain the same. Thus the investment cost of
machines is not differential and is not considered. The investment cost of buffers
is different only for the alternative 1. Revenues are discriminant if the additional
production capacity of the proposed alternatives is converted in additional sales.
114 A. Matta et al.
Fig. 14. Real case: Machine 1 states vs number of pallets for real system and alternative 2
A reasonable assumption is that the unit production cost does not differ because
the process technology is the same for each alternative. Again, if the additional
capacity is exploited in new sales the total variable production costs are a differ-
ential item. The residual value of a production system at the end of the planning
horizon is very difficult to estimate. However, the difference between the initial
investment with the real system is limited to 44 kEuro for the first alternative and
null for the second one and we can imagine that the difference between the residual
value will be smaller, so we can neglect the residual value voice in the analysis. The
second alternative has the same investment cost of the real system with a higher
production rate, and in this case the NPV analysis is not necessary because the real
system is dominated by the alternative. For the first alternative we have to impose
Automated flow lines with shared buffer 115
Fig. 15. Real case: Machine 2 states vs number of pallets for real system and alternative 2
Fig. 16. Real case: Machine 3 states vs number of pallets for real system and alternative 2
4 Practical considerations
In order to introduce manufacturing flow lines with shared buffer in real shop floors
several aspects, both technological and economic, have to be clarified. First of all,
a necessary condition for the common buffer exploitation is that, in order to be
able to dispatch parts to the different machines, parts must be tracked during their
movements in the system. To do this, several technologies are available at low costs.
Lasers markers can sculpture codes, easily readable by optical devices, on metal
components in a fast and cheap way. Standard devices like chips can save informa-
tion and exchange it with the system supervisor. Also radio frequency technology
is now ready to be used in shop floors to exchange information without the limiting
Automated flow lines with shared buffer 117
Fig. 17. Real case: Machine 4 states vs number of pallets for real system and alternative 2
Fig. 18. Real case: Machine 5 states vs number of pallets for real system and alternative 2
Flexible Manufacturing Systems, which represent the first case of shared buffer in
automated manufacturing systems.
Another important aspect that is normally taken into consideration in the design
phase of a flow line is the floor space occupied by the system. Theoretically, it is
necessary an additional space to locate the common buffer in a manufacturing flow
line. However, it is also true that the space occupied by dedicated buffers decreases
and consequently the machines of the line are closer. Therefore, it is not possible
a priori to say anything about the effects of the shared buffer on the occupied floor
space since this aspect is closely connected to the lay-out of the real system.
Automated flow lines with shared buffer 119
The paper addresses the problem of fully using buffer spaces in flow lines. The idea
is to exploit recent technological devices to move in reasonable time pieces from
a machine to a common buffer area of the system and vice versa. In such a way
machines can avoid their blocking moving pieces to the shared buffer area. The
decrease of blocking in flow lines has a positive impact on their production rate.
The numerical analysis reported in the paper demonstrates the validity of the idea
pointing out also the factors that affect the improvement of the proposed system
architecture in terms of productivity.
In conclusion, several practical aspects have to be investigated before to state
that shared buffers can be successfully adopted in real manufacturing flow lines,
however the first results shown in this paper and the technologies now available
motivate further research in this direction. Ongoing research is dedicated to identify
the potential sectors for practical applications of the new concepts proposed in
this paper. Then, further research will focus on new key-issues never addressed in
literature and introduced by the architecture with the shared buffer:
– Allocation of dedicated and shared buffers. Traditionally only capacities of
dedicated buffers have been considered in the design phase of manufacturing
flow lines. In our opinion the buffer allocation problem in the case of shared
buffer will be easier in comparison than the traditional one because the new
system architecture is more robust, i.e. the system performance is stable in
several conditions and do not decay after some changes in the design of the
line.
– Performance evaluation of flow lines with shared buffer. New analytical meth-
ods are necessary to estimate performance of new system architectures. The
method of Tempelmeier et al. [10], originally developed to evaluate the per-
formance of Flexible Manufacturing Systems with blocking, could be adopted
also for flow lines with shared buffer. This method is being tested in terms of
accurateness of provided results.
– Management of flow lines with shared buffer. New dispatching rules could
be necessary to avoid deadlock in the new system architecture when pieces
converge to the same area coming from different positions.
References
6. Law AM, Kelton WD (2000) Simulation modelling and analysis. McGraw–Hill, New
York
7. Shantikumar JG, Yao DD (1989) Queueing networks with finite buffers. In: Perros
HG, Altiok T (eds) chapter Monotonicity and concavity properties in cyclic queueing
networks with finite buffers, pp. 325–344. North Holland, Amsterdam
8. Tempelmeier H (2003) Practical considerations in the optimization of flow production
systems. International Journal of Production Research 41(1): 149–170
9. Tempelmeier H, Kuhn H (1993) Flexible manufacturing systems – Decision support
for design and operation. Wiley, New York
10. Tempelmeier H, Kuhn H, Tetzlaff U (1989) Performance evaluation of flexible manu-
facturing systems with blocking. International Journal of Production Research, 27(11):
1963–1979
Integrated quality and quantity modeling
of a production line
Jongyoon Kim and Stanley B. Gershwin
Department of Mechanical Engineering, Massachusetts Institute of Technology,
Cambridge, MA 02139-4307, USA (e-mail: [email protected])
Abstract. During the past three decades, the success of the Toyota Production Sys-
tem has spurred much research in manufacturing systems engineering. Productivity
and quality have been extensively studied, but there is little research in their inter-
section. The goal of this paper is to analyze how production system design, quality,
and productivity are inter-related in small production systems. We develop a new
Markov process model for machines with both quality and operational failures,
and we identify important differences between types of quality failures. We also
develop models for two-machine systems, with infinite buffers, buffers of size zero,
and finite buffers. We calculate total production rate, effective production rate (ie,
the production rate of good parts), and yield. Numerical studies using these models
show that when the first machine has quality failures and the inspection occurs only
at the second machine, there are cases in which the effective production rate in-
creases as buffer sizes increase, and there are cases in which the effective production
rate decreases for larger buffers. We propose extensions to larger systems.
1 Introduction
1.1 Motivation
During the past three decades, the success of the Toyota Production System has
spurred much research in manufacturing systems design. Numerous research pa-
pers have tried to explore the relationship between production system design and
We are grateful for support from the Singapore-MIT Alliance, the General Motors Re-
search and Development Center, and PSA Peugeot-Citroën.
Correspondence to: S.B. Gershwin
122 J. Kim and S.B. Gershwin
productivity, so that they can show ways to design factories to produce more prod-
ucts on time with less resources (such as people, material, and space). On the other
hand, topics in quality research have captured the attention of practitioners and re-
searchers since the early 1980s. The recent popularity of Statistical Quality Control
(SQC), Total Quality Management (TQM), and Six Sigma have demonstrated the
importance of quality.
These two fields, productivity and quality, have been extensively studied and
reported separately both in the manufacturing systems research literature and the
practitioner literature, but there is little research in their intersection. The need for
such work was recently described by authors from the GM Corporation based on
their experience [13]. All manufacturers must satisfy these two requirements (high
productivity and high quality) at the same time to maintain their competitiveness.
Toyota Production System advocates admonish factory designers to combine
inspections with operations. In the Toyota Production System, the machines are
designed to detect abnormalities and to stop automatically whenever they occur.
Also, operators are equipped with means of stopping the production flow whenever
they note anything suspicious. (They call this practice jidoka.) Toyota Production
System advocates argue that mechanical and human jidoka prevent the waste that
would result from producing a series of defective items. Therefore jidoka is a means
to improve quality and increase productivity at the same time [23], [24]. But this
statement is arguable: quality failures are often those in which the quality of each
part is independent of the others. This is the case when the defect takes place due
to common (or chance or random) causes of variations [16]. In this case, there is
no reason to stop a machine that has made a bad part because there is no reason
to believe that stopping it will reduce the number of bad parts in the future. In this
case, therefore, stopping the operation does not influence quality but it does reduce
productivity. On the other hand, when quality failures are those in which once a
bad part is produced, all subsequent parts will be bad until the machine is repaired
(due to special or assignable or systematic causes of variations) [16], catching bad
parts and stopping the machine as soon as possible is the best way to maintain high
quality and productivity.
Non-stock or lean production is another popular buzzword in manufacturing
systems engineering. Some lean manufacturing professionals advocate reducing
inventory on the factory floor since the reduction of work-in-process (WIP) reveals
the problems in the production lines [3]. Thus, it can help improve product quality.
It is true in some sense: less inventory reduces the time between making a defect
and identifying the defect. But it is also true that productivity would diminish
significantly without stock [5]. Since there is a tradeoff, there must be optimal
stock levels that are specific to each manufacturing environment. In fact, Toyota
recently changed their view on inventory and are trying to re-adjust their inventory
levels [9].
What is missing in discussions of factory design, quality, and productivity is a
quantitative model to show how they are inter-related. Most of the arguments about
this are based on anecdotal evidence or qualitative reasoning that lack a sound sci-
entific quantitative foundation. The research described here tries to establish such
a foundation to investigate how production system design and operation influence
Integrated quality and quantity modeling of a production line 123
1.2 Background
1.2.1 Quality models. There are two extreme kinds of quality failures based on
the characteristics of variations that cause the failures. In the quality literature,
these variations are called common (or chance or random) cause variations and
assignable (or special or unusual) cause variations [18].
Figure 1 shows the types of quality failures and variations. Common cause
failures are those in which the quality of each part is independent of the others.
Such failures occur often when an operation is sensitive to external perturbations
like defects in raw material or when the operation uses a new technology that is
difficult to control. This is inherent in the design of the process. Such failures can be
represented by independent Bernoulli random variables, in which a binary random
variable, which indicates whether or not the part is good, is chosen each time a
part is operated on. A good part is produced with probability π, and a bad part
is produced with probability 1 − π. The occurrence of a bad part implies nothing
about the quality of future parts, so no permanent changes can have occurred in the
machine. For the sake of clarity, we call this a Bernoulli-type quality failure. Most
of the quantitative literature on inspection allocation assumes this kind of quality
failure [21]. In this case, if bad parts are destined to be scrapped, it is useful to catch
them as soon as possible because the longer before they are scrapped, the more they
consume the capacity of downstream machines. However, there is no reason to stop
a machine that has produced a bad part due to this kind of failure.
The quality failures due to assignable cause variations are those in which a
quality failure only happens after a change occurs in the machine. In that case, it is
very likely that once a bad part is produced, all subsequent parts will be bad until
the machine is repaired. Here, there is much more incentive to catch defective parts
and stop the machine quickly. In addition to minimizing the waste of downstream
capacity, this strategy minimizes the further production of defective parts. For this
kind of quality failure, there is no inherent measure of yield because the fractions
Persistent-
type quality
failure
Bernoulli-
type quality
failure
Repair takes place
Upper
Specification
Limit
Mean
Lower
Random Variation
Specification
Limit
Assignable Variation
(tool breakage) takes
place
of parts that are good and bad depend on how soon bad parts are detected and how
quickly the machine is stopped for repair. In this paper, we call this a persistent-type
quality failure. Most quantitative studies in Statistical Quality Control (SQC) are
dedicated to finding efficient inspection policies (sampling interval, sample size,
control limits, and others) to detect this type of quality failure [26].
In reality, failures are mixtures of Bernoulli-type quality failures and persistent-
type quality failures. It can be argued that the quality strategy of the Toyota Produc-
tion System [17], in which machines are stopped as soon as a bad part is detected,
is implicitly based on the assumption of the persistent-type quality failure. In this
paper, we focus on persistent failures.
1.2.2 System yield. System yield is defined here as the fraction of input to a sys-
tem that is transformed into output of acceptable quality. This is an important metric
because customers observe the quality of products only after all the manufacturing
processes are done and the products are shipped. The system yield is a complex
function of how the factory is designed and operated, as well as of the characteristics
of the machines. Some of influencing factors include individual operation yields,
inspection strategies, operation policies, buffer sizes, and other factors. Compre-
hensive approaches are needed to manage system yield effectively. This research
aims to develop mathematical models to show how the system yield is influenced
by these factors.
1.2.3 Quality improvement policy. System yield is a complex function of various
factors such as inspection, individual operation yields, buffer size, operation poli-
cies, and others. There are many ways to affect the system yield. Inspection policy
has received the most attention in the literature. Research on inspection policies
can be divided into optimizing inspection parameters at a single station and the
inspection station allocation problem. The former issue has been investigated ex-
tensively in the SQC literature [26]. Here, optimal SQC parameters such as control
limits, sampling size, and frequency are sought for an optimal balance between the
inspection cost and the cost of quality. The latter research looks for the optimal
distributions of inspection stations along production lines [21].
Improving individual operation yield is another important way to increase the
system yield. Many studies in this field try to stabilize the process either by finding
root causes of variation and eliminating them or by making the process insensitive
to external noise. The former topic has numerous qualitative research papers in the
fields of Total Quality Management (TQM) [2] and Six Sigma [19]. Quantitative
research is more oriented toward the latter topic. Robust engineering [20] is an area
that has gained substantial attention.
It has been argued that inventory reduction is an effective means to improve
system yield. Many lean manufacturing specialists have asserted that less inventory
on the factory floor reveals problems in the manufacturing lines more quickly and
helps quality improvement activities [1, 17].
There also have been investigations to explain the relationship between plant
layout design and quality [7]. They argue that U-shaped lines are better than straight
lines for producing higher quality products since there are more points of contact
between operators. There is also less material movement, and there are other rea-
sons.
Integrated quality and quantity modeling of a production line 125
There are many ways to improve system yield, but using only a single method
will give limited gains. The effectiveness of each method is greatly dependent on
the details of the factory. Thus, there is need to determine which method or which
combination of methods is most effective in each case. The quantitative tools that
will be developed from this research can help fulfill this need.
1.3 Outline
2 Mathematical models
There are many possible ways to characterize a machine for the purpose of simul-
taneously studying quality and quantity issues. Here, we model a machine as a
discrete state, continuous time Markov process. Material is assumed continuous,
and µi is the speed at which Machine i processes material while it is operating and
not constrained by the other machine or the buffer. It is a constant, in that µi does
not depend on the repair state of the other machine or the buffer level.
Figure 2 shows the proposed state transitions of a single machine with persistent-
type quality failures. In the model, the machine has three states:
∗ State 1: The machine is operating and producing good parts.
∗ State -1: The machine is operating and producing bad parts, but the operator does
not know this yet.
∗ State 0: The machine is not operating.
The machine therefore has two different failure modes (i.e. transition to failure
states from state 1):
∗ Operational failure: transition from state 1 to state 0. The machine stops pro-
ducing parts due to failures like motor burnout.
∗ Quality failure: transition from state 1 to state -1. The machine stops producing
good parts (and starts producing bad parts) due to a failure like a sudden tool
damage.
When a machine is in state 1, it can fail due to a non-quality related event.
It goes to state 0 with transition probability rate p. After that an operator fixes it,
and the machine goes back to state 1 with transition rate r. Sometimes, due to an
assignable cause, the machine begins to produce bad parts, so there is a transition
126 J. Kim and S.B. Gershwin
g f
r
Fig. 2. States of a machine
from state 1 to state -1 with a probability rate g. Here g is the reciprocal of the Mean
Time to Quality Failure (MTQF). A more stable operation leads to a larger MTQF
and a smaller g.
The machine, when it is in state -1, can be stopped for two reasons: it may
experience the same kind of operational failure as it does when it is in state 1; and
the operator may stop it for repair when he learns that it is producing bad parts.
The transition from state -1 to state 0 occurs at probability rate f = p + h where h
is the reciprocal of the Mean Time To Detect (MTTD). A more reliable inspection
leads to a shorter MTTD and a larger f . (The detection can take place elsewhere,
for example at a remote inspection station.) Note that this implies that f > p. Here,
for simplicity, we assume that whenever a machine is repaired, it goes back to state
1. All the indicated transitions are assumed to follow exponential distributions.
A flow (or transfer) line is a manufacturing system with a very special structure. It
is a linear network of service stations or machines (M1 , M2 , ..., Mk ) separated by
buffer storages (B1 , B2 , ..., Bk−1 ). Material flows from outside the system to M1 ,
then to B1 , then to M2 , and so forth until it reaches Mk , after which it leaves. Figure
3 depicts a flow line. The rectangles represent machines and the circles represent
buffers.
M1 B1 M2 B2 M3 B3 M4 B4 M5
M1 B
M2
M1 B M2
of the last machine. Thus, the first machine is never starved, and the last machine
is never blocked. Also, failures are assumed to be operation dependent (ODF).
Finally, we assume that each machine works on a different feature. For example,
the two machines may be making two different holes. We do not consider cases
where the both machines work on the same hole, in which the first machine does
a roughing operation and the second does a finishing operation. This allows us to
assume that the failures of the two machines are independent.
An infinite buffer case is a special 2M1B line in which the size of the buffer (B)
is infinite. This is an extreme case in which the first machine (M1 ) never suffers
from blockage. To derive expressions for the total production rate and the effective
production rate, we observe that when there is infinite buffer capacity between two
machines (M1 , M2 ), the total production rate of the 2M1B system is a minimum
of the total production rates of M1 and M2 . The total production rate of machine i
is given by (8), so the total production rate of the 2M1B system is
! "
µ1 (1 + g1 /f1 ) µ2 (1 + g2 /f2 )
PT∞ = min , (11)
1 + (p1 + g1 )/r1 + g1 /f1 1 + (p2 + g2 )/r2 + g2 /f2
The probability that machine Mi does not add non-conformities is
Pi (1) fi
Yi = = (12)
Pi (1) + Pi (−1) fi + gi
Since there is no scrap and rework in the system, the system yield becomes
f1 f2
(13)
(f1 + g1 )(f2 + g2 )
As a result, the effective production rate is
f1 f2
PE∞ = P∞ (14)
(f1 + g1 )(f2 + g2 ) T
The effective production rate evaluated from (14) has been compared with a
discrete-event, discrete-part simulation. Table 1 shows good agreement. The pa-
rameters for these cases are shown in Appendix B.
As indicated in Section 2.1, the detection of quality failures due to machine M1
need not occur at that machine. For example, the inspection of the feature that M1
works on could take place at an inspection station at M2 , and this inspection could
trigger a repair of M1 . (We call this quality information feedback. See Section 4.)
In that case, the MTTD of M1 (and therefore f1 ) will be a function of the amount
of material in the buffer. We return to this important case in Section 4.
The zero buffer case is one in which there is no buffer space between the machines.
This is the other extreme case where blockage and starvation take place most
frequently.
Integrated quality and quantity modeling of a production line 129
In the zero-buffer case in which machines have different operation times, when-
ever one of the machines stops, the other one is also stopped. In addition, when
both of them are working, the production rate is min[µ1 , µ2 ]. Consider a long time
interval of length T during which M1 fails m1 times and M2 fails m2 times. If we
assume that the average time to repair M1 is 1/r1 and the average time to repair
M2 is 1/r2 , then the total system down time will be close to D = m 1 m2
r1 + r2 .
Consequently, the total up time will be approximately
m1 m2
U =T −D =T −( + ) (15)
r1 r2
Since we assume operation-dependent failures, the rates of failure are reduced
for the faster machine. Therefore,
min(µ1 , µ2 ) b min(µ1 , µ2 ) min(µ1 , µ2 )
pbi = pi , gi = gi , and fib = fi (16)
µi µi µi
The reduction of pi is explained in detail in [10]. The reductions of gi and fi
are done for the same reasons.
Table 2 lists the possible working states α1 and α2 of M1 and M2 . The third
column is the probability of finding the system in the indicated state. The fourth
and fifth columns indicate the expected number of transitions to down states during
the time interval from each of the states in column 1.
f1b f2b
1 1 f1b +g1b f b +g b pb1 U π(1, 1) pb2 U π(1, 1)
2 2
f1b b
g2
1 −1 f1b +g1
b f b +g b pb1 U π(1, −1) f2b U π(1, −1)
2 2
b
g1 f2b
−1 1 f1b +g1
b f b +g b f1b U π(−1, 1) pb2 U π(−1, 1)
2 2
b b
g1 g2
−1 −1 f1b +g1
b f b +g b f1b U π(−1, −1) f2b U π(−1, −1)
2 2
130 J. Kim and S.B. Gershwin
3 2-machine-1-finite-buffer line
The two-machine line is the simplest non-trivial case of a production line. In the
existing literature on the performance evaluation of systems in which quality is not
considered, two-machine lines are used in decomposition approximations of longer
lines (see [10]).
We define the model here and show the solution technique in Appendix A.
Integrated quality and quantity modeling of a production line 131
3.2.1 Internal transition equations. When buffer B is neither empty nor full, its
level can rise or fall depending on the states of adjacent machines. Since it can
change only a small amount during a short time interval, it is reasonable to use a
continuous probability density f (x, α1 , α2 ) and differential equations to describe
its behavior. The probability of finding both machines at state 1 with a storage level
between x and x + δx at time t + δt is given by f (x, 1, 1, t + δt)δx, where
f (x, 1, 1, t + δt) = {1 − (p1 + g1 + p2 + g2 )δt}f (x + (µ2 − µ1 )δt, 1, 1) (20)
+r2 δtf (x − µ1 δt, 1, 0) + r1 δtf (x + µ2 δt, 0, 1) + o(δt)
Except for the factor of δx, the first term is the probability of transition from
between (x + (µ2 − µ1 )δt, 1, 1) and (x + (µ2 − µ1 )δt + δx, 1, 1) at time t to
between (x, 1, 1) and (x + δx, 1, 1) at time t + δt. This is because
∗ The probability of neither machine failing between t and t + δt is
{1 − (p1 + g1 )δt}{1 − (p2 + g2 )δt} {1 − (p1 + g1 + p2 + g2 )δt} (21)
∗ If there are no failures between t and t + δt and the buffer level is between x and
x + δx at time t + δt, then it could only have been between x + (µ2 − µ1 )δt and
x + (µ2 − µ1 )δt + δx at time t.
The other terms, which represent the probabilities of transition from (1) machine
states (1,0) with buffer level between x − µ1 δt and x − µ1 δt + δx and (2) machine
states (0,1) with buffer level between x + µ2 δt and (x + µ2 δt + δx can be found
similarly. No other transitions are possible. After linearizing, and letting δt → 0,
this equation becomes
∂f (x, 1, 1) ∂f (x, 1, 1)
= (µ2 − µ1 ) − (p1 + g1 + p2 + g2 )f (x, 1, 1)
∂t ∂x
+r2 f (x, 1, 0) + r1 f (x, 0, 1) (22)
∂f
In steady state ∂t = 0. Then, we have
df (x, 1, 1)
(µ2 − µ1 ) − (p1 + g1 + p2 + g2 )f (x, 1, 1) + r2 f (x, 1, 0)
dx
+r1 f (x, 0, 1) = 0 (23)
132 J. Kim and S.B. Gershwin
In the same way, the eight other internal transition equations for the probability
density function are
df (x, 1, 0)
p2 f (x, 1, 1) − µ1 − (p1 + g1 + r2 )f (x, 1, 0) + f2 f (x, 1, −1)
dx
+r1 f (x, 0, 0) = 0 (24)
df (x, 1, −1)
g2 f (x, 1, 1) + (µ2 − µ1 ) − (p1 + g + f2 )f (x, 1, −1)
dx
+r1 f (x, 0, −1) = 0 (25)
df (x, 0, 1)
p1 f (x, 1, 1) + µ2 − (r1 + p2 + g2 )f (x, 0, 1) + r2 f (x, 0, 0)
dx
+f1 f (x, −1, 1) = 0 (26)
p1 f (x, 1, 0) + p2 f (x, 0, 1) − (r1 + r2 )f (x, 0, 0) + f2 f (x, 0, −1)
+f1 f (x, −1, 0) = 0 (27)
df (x, 0, −1)
p1 f (x, 1, −1) + g2 f (x, 0, 1) − (r1 + f2 )f (x, 0, −1) + µ2
dx
+f1 f (x, −1, −1) = 0 (28)
df (x, −1, 1)
g1 f (x, 1, 1) − (p2 + g2 + f1 )f (x, −1, 1) + (µ2 − µ1 )
dx
+r2 f (x, −1, 0) = 0 (29)
df (x, −1, 0)
g1 f (x, 1, 0) − µ1 − (r2 + f1 )f (x, −1, 0) + p2 f (x, −1, 1)
dx
+f2 f (x, −1, −1) = 0 (30)
df (x, −1, −1)
g1 f (x, 1, −1) + g2 f (x, −1, 1) + (µ2 − µ1 )
dx
−(f1 + f2 )f (x, −1, −1) = 0 (31)
3.2.2 Boundary transition equations. While the internal behavior of the system
can be described by probability density functions, there is a nonzero probability of
finding the system in certain boundary states. For example, if µ1 < µ2 and both
machines are in state 1, the level of storage tends to decrease. If both machines
remain operational for enough time, the storage will become empty (x = 0). Once
the system reaches state (0, 1, 1), it will remain there until a machine fails. There are
18 probability masses for boundary states (P (N, α1 , α2 ) and P (0, α1 , α2 ) where
α1 = −1, 0 or 1, and α2 = −1, 0 or 1) and 22 boundary equations for the µ1 = µ2
case.
To arrive at state (0, 1, 1) at time t + δt when µ1 = µ2 , the system may have
been in one of two states at time t. It could have been in state (0, 1, 1) without any
of operational failures and quality failures for both of machines. It could have been
in state (0, 0, 1) with a repair of the first machine. (The second machine could not
have failed since it was starved). If the second order terms are ignored,
P (0, 1, 0) = 0 (35)
g2b P (0, 1, 1) − (p1 + g1 + f2b )P (0, 1, −1) + r1 P (0, 0, −1) = 0 (36)
p1 P (0, 1, 1) − r1 P (0, 0, 1) + µ2 f (0, 0, 1) + f1 P (0, −1, 1)
+r2 P (0, 0, 0) = 0 (37)
−(r1 + r2 )P (0, 0, 0) = 0 (38)
p1 P (0, 1, −1) − r1 P (0, 0, −1) + µ2 f (0, 0, −1) + f1 P (0, −1, −1) = 0 (39)
g1 P (0, 1, 1) − (f1 + pb2 + g2b )P (0, −1, 1) =0 (40)
P (0, −1, 0) = 0 (41)
g1 P (0, 1, −1) + g2b P (0, −1, 1) − (f1 + f2b )P (0, −1, −1) = 0 (42)
−(pb1 + g1b + p2 + g2 )P (N, 1, 1) + r2 P (N, 1, 0) = 0 (43)
p2 P (N, 1, 1) − r2 P (N, 1, 0) + µ1 f (N, 1, 0) + f2 P (N, 1, −1)
+r1 P (N, 0, 0) = 0 (44)
g2 P (N, 1, 1) − (pb1 + g1b + f2 )P (N, 1, −1) = 0 (45)
P (N, 0, 1) = 0 (46)
−(r1 + r2 )P (N, 0, 0) = 0 (47)
P (N, 0, −1) = 0 (48)
g1b P (N, 1, 1) − (f1b + g2 + p2 )P (N, −1, 1) + r2 P (N, −1, 0) = 0 (49)
−r2 P (N, −1, 0) + µ1 f (N, −1, 0) + f2 P (N, −1, −1)
+p2 P (N, −1, 1) = 0 (50)
g1b P (N, 1, −1) + g2 P (N, −1, 1) − (f1b + f2 )P (N, −1, −1) = 0 (51)
µ1 f (0, 1, 0) = r1 P (0, 0, 0) + pb2 P (0, 1, 1) + f2b P (0, 1, −1) (52)
µ1 f (0, −1, 0) = pb2 P (0, −1, 1) + f2b P (0, −1, −1) (53)
µ2 f (N, 0, 1) = r2 P (N, 0, 0) + pb1 P (N, 1, 1) + f1b P (N, −1, 1) (54)
µ2 f (N, 0, −1) = pb1 P (N, 1, −1) + g2 P (N, 0, 1) + f1b P (N, −1, −1) (55)
3.2.4 Performance measures. After finding all probability density functions and
probability masses, we can calculate the average inventory in the buffer from
⎡N ⎤
$
x= ⎣ xf (x, α1 , α2 )dx + N P (N, α1 , α2 )⎦ (57)
α1 =−1,0,1 α2 =−1,0,1 0
The probability that the first machine produces a non-defective part is then Y1 =
PE1 /PT . Similarly, the probability that the second machine finishes its operation
without adding a bad feature to a part is Y2 = PE2 /PT , where
$N
PE2 = µ2 [ f (x, α1 , 1)dx
α1 =−1,0,1 0
+P (N, α1 , 1)] + µ1 {P (0, −1, 1) + P (0, 1, 1)} (60)
3.3 Validation
The 2M1B systems with the same machine speed (µ1 = µ2 ) are solved in Ap-
pendix A. As we have indicated, we represent discrete parts in this model as a
continuous fluid and time as a continuous variable. We compare analytical and
simulation results in this section. In the simulation, both material and time are
discrete. Details are presented in [14].
Figure 5 shows the comparison of the effective production rate and the average
inventory from the analytic model and the simulation. 50 cases are generated by
changing machines and buffer parameters and % errors are plotted in the vertical
Integrated quality and quantity modeling of a production line 135
10.00% 10.00%
8.00% 8.00%
6.00% 6.00%
4.00% 4.00%
E
% error of Inv
2.00% 2.00%
% error of P
0.00% 0.00%
1
4
7
1
4
7
10
13
16
19
22
25
28
31
34
37
40
43
46
49
10
13
16
19
22
25
28
31
34
37
40
43
46
49
-2.00% -2.00%
-4.00% -4.00%
-6.00% -6.00%
-8.00% -8.00%
-10.00% -10.00%
axis. The parameters for theses cases are given in Appendix B. The % error in the
effective production rate is calculated from
PE (A) − PE (S)
PE %error = × 100(%) (62)
PE (S)
where PE (A) and PE (S) are the effective production rates estimated from the
analytical model and the simulation respectively. But the % error in the average
inventory is calculated from
InvE (A) − InvE (S)
InvE %error = × 100(%) (63)
0.5 × N
where InvE (A) and InvE (S) are average inventory estimated from the analytical
model and the simulation respectively and N is a buffer size1 .
The average absolute value of the % error in the effective production rate esti-
mation is 0.76% and it is 1.89% for average inventory estimation.
Factory designers and managers know that it is ideal to have inspection after every
operation. However, it is often costly to do this. As a result, factories are usually
designed so that multiple inspections are performed at a small number of stations.
In this case, inspection at downstream operations can detect bad features made by
upstream machines. We call this quality information feedback. A simple example of
the quality information feedback in 2M1B systems is when M1 produces defective
features but does not have inspection and M2 has inspection and it can detect bad
features made by M1 . In this situation, as we demonstrate below, the yield of a line
is a function of the size of buffer. This is because when buffer gets larger, more
material can accumulate between an operation (M1 ) and the inspection of that
operation (M2 ). All such material will be defective if a persistent quality failure
1
This is an unbiased way to calculate the error in average inventory. If it were calculated
in the same way as the error in the effective production rate, the error would depend on the
relative speeds of the machines. This is because there will be a lower error when the buffer
is mostly full (ie, when M1 is faster than M2 ) and a higher error when the buffer is empty
(when M1 is faster than M2 ).
136 J. Kim and S.B. Gershwin
takes place. In other words, if buffer is larger, there tends to be more material in the
buffer and consequently more material is defective. In addition it takes longer to
have inspections after finishing operations. We can capture this phenomenon with
the adjustment of a transition probability rate of M1 from state -1 to state 0.
Let us define f1q as a transition rate of M1 from state -1 to state 0 when there
is a quality information feedback and f1 as the transition rate without the quality
information feedback. The adjustment can be done in a way that the yield of M1 is
Z1g
the same as Z g +Z b where
1 1
∗ Z1b : the expected number of bad parts generated by M1 while it stays in state -1.
∗ Z1g : the expected number of good parts produced by M1 from the moment when
M1 leaves the -1 state to the next time it arrives at state -1.
From (10), the yield of M1 is
P (1) fq
= q 1 (64)
P (1) + P (−1) f1 + g1
Suppose that M1 has been in state 1 for a very long time. Then all parts in the
buffer B are non-defective. Suppose that M1 goes to state -1. Defective parts will
then begin to accumulate in the buffer. Until all the parts in the buffer are defective,
the only way that M1 can go to state 0 is due to its own inspection or its own
operation failure. Therefore, the probability of a transition to 0 before M1 finishes
a part is
f1
≡ χ11
µ1
Eventually all the parts in the buffer are bad so that defective parts reach M2 . Then,
there is another way that M1 can move to state 0 from state -1: quality information
feedback. The probability that the inspection at M2 detects a nonconformity made
by M1 is
h21
χ21 ≡
µ2
where 1/h21 is the mean time until the inspection at M2 detects a bad part made
by M1 after M2 receives the bad part.
The expected value of the number of bad parts produced by M1 before it is
stopped by either operational failures or quality information feedback is
Z1b = [χ11 + 2χ11 (1 − χ11 ) + 3χ11 (1 − χ11 )2 + . . . + wχ11 (1 − χ11 )w−1 ]
+[(w + 1)(1 − χ11 )w χ21 + (w + 2)(1 − χ11 )w+1 χ21 (1 − χ21 ) + . . .](65)
where w is average inventory in the buffer B. This is an approximate formula since
we simply use the average inventory rather than averaging the expected number of
bad parts produced by M1 depending on different inventory levels wi . After some
mathematical manipulation,
1 − (1 − χ11 )w
Z1b = − w(1 − χ11 )w
χ11
(1 − χ11 )w χ21 [(w + 1) − w(1 − χ11 )(1 − χ21 )]
+ (66)
[1 − (1 − χ11 )(1 − χ21 )]2
Integrated quality and quantity modeling of a production line 137
Since the average inventory is a function of f1q and f1q is dependent on the
average inventory, an iterative method is required to determine these values.
15.00% 15.00%
10.00% 10.00%
5.00% 5.00%
% error of Inv
% error of P E
0.00% 0.00%
1
9
13
17
21
25
29
33
37
13
17
21
25
29
33
37
-5.00% -5.00%
-10.00% -10.00%
-15.00% -15.00%
Case Number Case Number
Figure 6 shows the comparison of the effective production rate and the average
inventory from the analytic model and the simulation. 50 cases are generated by
selecting different machine and buffer parameters and % errors are plotted in the
y-axis. The parameters for theses cases are given in Appendix B. % errors in the
effective production rate and average inventory are calculated using equations (62)
and (63) respectively. The average absolute value of the % error in PE and x
estimations are 1.01% and 3.67% respectively.
5.1.1 Production rates. Having quality information feedback means having more
inspection than otherwise. Therefore, machines tend to stop more frequently. As
a result, the total production rate of the line decreases. However, the effective
production rate can increase since added inspections prevent the making of defective
parts. This phenomenon is shown in Figure 7. Note that the total production rate PT
without quality information feedback is consistently higher than PT with quality
information feedback regardless of buffer size and the opposite is true for the
138 J. Kim and S.B. Gershwin
effective production rate PE . Also it should be noted that in this case, both the total
production rate and the effective production rate increase with buffer size, with or
without quality information feedback.
0.8
0.8
with feedback
0.75
0.75
without feedback
with feedback
0.7
0.7
0.65
0.65 0 5 10 15 20 25 30 35 40 45 50
0 5 10 15 20 25 30 35 40 45 50
Buffer Size Buffer Size
5.1.2 System yield and buffer size. Even though a larger buffer increases both total
and effective production rates in this case, it decreases yield. As explained in Section
4, the system yield is a function of the buffer size if there is quality information
feedback. Figure 8 shows system yield decreasing as buffer size increases when
there is quality information feedback. This happens because when the buffer gets
larger, more material accumulates between an operation and the inspection of that
operation. All such material will be defective when the first machine is at state -1
but the inspection at the first machine does not find it. This is a case in which a
smaller buffer improves quality, which is widely believed to be generally true. If
there is no quality information feedback, then the system yield is independent of
the buffer size (and is substantially less).
0.97
0.96
0.95
without feedback
with feedback
0.94
System Yield
0.93
0.92
0.91
0.9
0.89
0 5 10 15 20 25 30 35 40 45 50
Buffer Size
1
1
Without feedback
With feedback
Without feedback
With feedback
0.5
0.5
0
0 5 10 15 20 25 30 35 40 45 50 0
0 5 10 15 20 25 30 35 40 45 50
Buffer Size Buffer Size
0.9
0.8
0.7
System Yield
0.6
Without feedback
With feedback
0.5
0.4
0.3
0.2
0.1
0
0 5 10 15 20 25 30 35 40 45 50
Buffer Size
5.2.2 System yield. The system yield for this case is shown in Figure 10. Note
that the yield decreases dramatically as the buffer size increases. In this case, the
decrease of the system yield is more than the increase of the total production rate
so that the effective production rate monotonically decreases as buffer size gets
bigger.
There are two major ways to improve quality. One is to increase the yield of in-
dividual operations and the other is to perform more rigorous inspection. Having
extensive preventive maintenance on manufacturing equipment and using robust
engineering techniques to stabilize operations have been suggested as tools to in-
crease yield of individual operations. Both approaches increase the Mean Time to
Quality Failure (MTQF) (i.e. decrease g). On the other hand, the inspection policy
aims to detect bad parts as soon as possible and prevent their flow toward down-
stream operations. More rigorous inspection decreases the mean time to detect
(MTTD) (i.e. increases h and therefore increases f ). It is natural to believe that
using only one kind of method to achieve a target quality level would not give the
most cost efficient quality assurance policy. Figure 11 indicates that the impact of
individual operation stabilization on the system yield decreases as the operation
becomes more stable. It also shows that effect of improving inspection (MTTD) on
the system yield decreases. Therefore, it is optimal to use a combination of both
methods to improve quality.
1
1
0.9
0.9
0.8
0.8
0.7
0.7
System Yield
0.6
System Yield
0.6
0.5
0.5
0.4
0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
MTQF f=p+h
Improving the stand-alone throughput of each operation and increasing the buffer
space are typical ways to increase the production rate of manufacturing systems.
If operations are apt to have quality failures, however, there may be other ways to
increase the effective production rate: increasing the yield of each operation and
conducting more extensive inspections. Stabilizing operations, thus improving the
yield of individual operations, will increase effective throughput of a manufacturing
Integrated quality and quantity modeling of a production line 141
system regardless of the type of quality failure. On the other hand, reducing the
mean time to detect (MTTD) will increase the effective production rate only if the
quality failure is persistent but it will decrease the effective production rate if the
quality failure is Bernoulli. This is because the quality of each part is independent
of the others when the quality failure is Bernoulli. Therefore, stopping the line does
not reduce the number of bad parts in the future.
In a situation in which machines produce defective parts frequently and in-
spection is poor, increasing inspection reliability is more effective than increasing
buffer size to boost the effective production rate. Figure 12 shows this. Also, in other
situations in which machines produce defective parts frequently and inspection is
reliable, increasing machine stability is more effective than increasing buffer size
to enhance effective production rate. Figure 13 shows this phenomenon.
0.8
0.75
MTTD = 20
MTTD = 10
MTTD = 2
Effective Production Rate
0.7
0.65
0.6
0.55
0.5
0 5 10 15 20 25 30 35 40
Buffer Size
0.8
Effective Production Rate
0.7
MTQF = 20
0.6
MTQF = 100
MTQF = 500
0.5
0.4
0.3
0 5 10 15 20 25 30 35 40
Buffer Size
6 Future research
The 2-Machine-1-Buffer (2M1B) model with µ1 = / µ2 is analyzed in [14]. This
case is more challenging because the number of roots of the internal transition
equations depends on parameters of machine. A more general 2M1B model with
multiple-yield quality failures (a mixture of Bernoulli- and persistent-type quality
failures) should also be studied. A long line analysis using decomposition is under
the development. Refer to Kim [14] for more detailed information.
Appendix
A Solution technique
It is natural to assume an exponential form for the solution to the steady state density
functions since equations (23)–(31) are coupled ordinary linear differential equa-
tions. A solution of the form eλx K1α1 K2α2 worked successfully in the continuous
material two-machine line with perfect quality [10]. Therefore, a solution of a form
f (x, α1 , α2 ) = eλx G1 (α1 )G2 (α2 ) (69)
is assumed here. This form satisfies the transition equations if all of the following
equations are met. Equations (23)–(31) become, after substituting (69),
{(µ2 − µ1 )λ − (p1 + g1 + p2 + g2 )G1 (1)G2 (1)}
+r2 G1 (1)G2 (0) + r1 G1 (0)G2 (1) = 0 (70)
−{µ1 λ + (p1 + g1 + r2 )}G1 (1)G2 (0) + p2 G1 (1)G2 (1) + f2 G1 (1)G2 (−1)
+r1 G1 (0)G2 (0) = 0 (71)
{(µ2 − µ1 )λ − (p1 + g1 + f2 )}G1 (1)G2 (−1) + g2 G1 (1)G2 (1)
+r1 G1 (0)G2 (−1) = 0 (72)
{µ2 λ − (r1 + p2 + g2 )}G1 (0)G2 (1) + p1 G1 (1)G2 (1) + r2 G1 (0)G2 (0)
+f1 G1 (−1)G2 (1) = 0 (73)
p1 G1 (1)G2 (0) + p2 G1 (0)G2 (1) − (r1 + r2 )G1 (0)G2 (0) + f2 G1 (0)G2 (−1)
+f1 G1 (−1)G2 (0) = 0 (74)
{µ2 λ − (r1 + f2 )}G1 (0)G2 (−1) + p1 G1 (1)G2 (−1) + g2 G1 (0)G2 (1)
+f1 G1 (−1)G2 (−1) = 0 (75)
{(µ2 − µ1 )λ − (p2 + g2 + f1 )}G1 (−1)G2 (1) + g1 G1 (1)G2 (1)
+r2 G1 (−1)G2 (0) = 0 (76)
−{µ1 λ + (r2 + f1 )}G1 (−1)G2 (0) + g1 G1 (1)G2 (0) + p2 G1 (−1)G2 (1)
+f2 G1 (−1)G2 (−1) = 0 (77)
{(µ2 − µ1 )λ − (f1 + f2 )}G1 (−1)G2 (−1) + g1 G1 (1)G2 (−1)
+g2 G1 (−1)G2 (1) = 0 (78)
These are nine equations in seven unknowns (λ, G1 (1), G2 (0), G1 (−1), G2 (1),
G2 (0), and G2 (−1)). Thus, there must be seven independent equations and two
dependent ones.
Integrated quality and quantity modeling of a production line 143
If we divide equations (70) – (78) by G1 (0)G2 (0) and define new parameters
Gi (1) Gi (−1)
Γ i = pi − ri + fi (79)
Gi (0) Gi (0)
Gi (0)
Ψi = −pi − gi + ri (80)
Gi (1)
Gi (1)
Θi = −fi + gi (81)
Gi (−1)
then equations (70)–(78) can be rewritten as
Γ1 + Γ2 = 0 (82)
−µ2 λ = Γ1 + Ψ2 (83)
µ1 λ = Γ2 + Ψ1 (84)
(µ1 − µ2 )λ = Ψ1 + Ψ2 (85)
(µ1 − µ2 )λ = Θ1 + Θ2 (86)
µ1 λ = Γ2 + Θ1 (87)
−µ2 λ = Γ1 + Θ2 (88)
(µ1 − µ2 )λ = Ψ2 + Θ1 (89)
(µ1 − µ2 )λ = Ψ1 + Θ2 (90)
From equations (82)–(90), it is clear that only seven equations are independent.
After much mathematical manipulation [14], these equations become
{(M + r1 )(µ1 N − 1) − f1 }2
0=
(f1 − p1 )(µ1 N − 1)
{(p1 + g1 − f1 ) + r1 (µ1 N − 1)}{(M + r1 )(µ1 N − 1) − f1 }
− − r1 (91)
(f1 − p1 )(µ1 N − 1)
{(−M + r2 )(µ2 N − 1) − f2 }2
0=
(f2 − p2 )(µ2 N − 1)
{(p2 +g2 −f2 )+r2 (µ2 N −1)}{(−M +r2 )(µ2 N −1)−f2 }
− −r2 =0 (92)
(f2 −p2 )(µ2 N −1)
where
G1 (1) G1 (−1) G2 (1) G2 (−1)
p1 − r1 + f1 = − p2 − r2 + f2 =M (93)
G1 (0) G1 (0) G2 (0) G2 (0)
⎛ ⎞
1 ⎝ 1 ⎠=
1+
µ1 G1 (1)/ + G1 (−1)/
G1 (0) G1 (0)
⎛ ⎞
1 ⎝ 1 ⎠=N
1+ (94)
µ2 G2 (1)/ + G2 (−1)/
G2 (0) G2 (0)
Now all the equations and unknowns are simplified into two unknowns and
two equations. By solving equations (91) and (92) simultaneously we can calculate
144 J. Kim and S.B. Gershwin
1
N
−1
−2
−3
−3 −2 −1 0 1 2 3
M
These are high order polynomial equations for which no general analytical
solution exists. A numerical approach is required to find the roots of the equations.
A special algorithm to find the solutions has been developed [14] based on the
characteristics of the equations. Once we find roots of equations (91) and (92),
Gi (1)
we can get ratios G i (0)
and GGi i(−1)
(0) (i = 1, 2) from equation (94). By setting
G1 (0) = G2 (0) = 1, we can calculate G1 (1), G1 (−1), G2 (1), and G2 (−1). After
some mathematical manipulation, we find that λ can be expressed as
−p1 − g1 + r1/G1 (1) − p1 G1 (1) + r1 − f1 G1 (−1)
λ= (95)
µ1
Therefore, we can get a probability density function f (x, α1 , α2 ) corresponding
to an (M, N ) pair. The number of roots in equations (91) and (92) depends on
machine parameters. There are only 3 roots when µ1 = µ2 regardless of other
parameters. Therefore, a general expression of the probability density function in
this case is
f (x, α1 , α2 ) = c1 f1 (x, α1 , α2 ) + c2 f2 (x, α1 , α2 ) + c3 f3 (x, α1 , α2 ) (96)
where f1 (x, α1 , α2 ), f2 (x, α1 , α2 ), f3 (x, α1 , α2 ) are the roots of the equations
(91) and (92).
The remaining unknowns, including c1 , c2 , c3 and probability masses at the
boundaries, can be calculated by solving boundary equations ((34)–(55)) and the
normalization equation (56) with fi (x, α1 , α2 ) given by equation (96).
Integrated quality and quantity modeling of a production line 145
Table 4. Machine parameters for infinite buffer case and zero buffer case
Case # µ1 µ2 r1 r2 p1 p2 g1 g2 f1 f2
1 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2
2 1.0 1.0 0.3 0.3 0.005 0.005 0.05 0.05 0.5 0.5
3 1.0 1.0 0.2 0.05 0.01 0.01 0.01 0.01 0.2 0.2
4 1.0 1.0 0.1 0.1 0.05 0.005 0.01 0.01 0.2 0.2
5 1.0 1.0 0.1 0.1 0.01 0.01 0.05 0.005 0.2 0.2
6 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.5 0.1
7 2.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.5 0.1
8 3.0 2.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2
9 1.0 2.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2
10 2.0 3.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2
µ1 µ2 r1 r2 p1 p2 g1 g2 f1 f2
1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.1 0.9
µ1 µ2 r1 r2 p1 p2 g1 g2 f1 f2
2.0 2.0 0.5 0.1 0.005 0.05 0.5 0.005 0.02 0.9
µ1 µ2 r1 r2 p1 p2 g1 g2 f1 f2
1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2
µ1 µ2 r1 r2 p1 p2 g1 g2
1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01
µ1 µ2 r1 r2 p1 p2 f1 f2
1.0 1.0 0.1 0.1 0.01 0.01 0.2 0.2
146 J. Kim and S.B. Gershwin
Case # µ1 µ2 r1 r2 p1 p2 g1 g2 f1 f2 N
1 1.0 1.0 0.1 0.1 0.01 0.01 0.02 0.01 0.1 0.2 30
2 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 5
3 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 10
4 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 15
5 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 20
6 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 25
7 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 35
8 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 40
9 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 45
10 0.5 0.5 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
11 1.5 1.5 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
12 2.0 2.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
13 2.5 2.5 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
14 3.0 3.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
15 1.0 1.0 0.01 0.01 0.01 0.01 0.01 0.01 0.2 0.2 30
16 1.0 1.0 0.05 0.05 0.01 0.01 0.01 0.01 0.2 0.2 30
17 1.0 1.0 0.2 0.2 0.01 0.01 0.01 0.01 0.2 0.2 30
18 1.0 1.0 0.5 0.5 0.01 0.01 0.01 0.01 0.2 0.2 30
19 1.0 1.0 0.8 0.8 0.01 0.01 0.01 0.01 0.2 0.2 30
20 1.0 1.0 0.1 0.1 0.001 0.001 0.01 0.01 0.2 0.2 30
21 1.0 1.0 0.1 0.1 0.005 0.005 0.01 0.01 0.2 0.2 30
22 1.0 1.0 0.1 0.1 0.02 0.02 0.01 0.01 0.2 0.2 30
23 1.0 1.0 0.1 0.1 0.05 0.05 0.01 0.01 0.2 0.2 30
24 1.0 1.0 0.1 0.1 0.1 0.1 0.01 0.01 0.2 0.2 30
25 1.0 1.0 0.1 0.1 0.01 0.01 0.001 0.001 0.2 0.2 30
26 1.0 1.0 0.1 0.1 0.01 0.01 0.005 0.005 0.2 0.2 30
27 1.0 1.0 0.1 0.1 0.01 0.01 0.02 0.02 0.2 0.2 30
28 1.0 1.0 0.1 0.1 0.01 0.01 0.05 0.05 0.2 0.2 30
29 1.0 1.0 0.1 0.1 0.01 0.01 0.10 0.10 0.2 0.2 30
30 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.02 0.02 30
31 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.05 0.05 30
32 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.1 0.1 30
33 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.5 0.5 30
34 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.95 0.95 30
35 1.0 1.0 0.5 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
36 1.0 1.0 0.01 0.1 0.01 0.01 0.01 0.01 0.2 0.2 30
37 1.0 1.0 0.1 0.5 0.01 0.01 0.01 0.01 0.2 0.2 30
38 1.0 1.0 0.1 0.01 0.01 0.01 0.01 0.01 0.2 0.2 30
39 1.0 1.0 0.1 0.1 0.1 0.01 0.01 0.01 0.2 0.2 30
40 1.0 1.0 0.1 0.1 0.001 0.01 0.01 0.01 0.2 0.2 30
41 1.0 1.0 0.1 0.1 0.01 0.1 0.01 0.01 0.2 0.2 30
42 1.0 1.0 0.1 0.1 0.01 0.001 0.01 0.01 0.2 0.2 30
43 1.0 1.0 0.1 0.1 0.01 0.01 0.1 0.01 0.2 0.2 30
44 1.0 1.0 0.1 0.1 0.01 0.01 0.001 0.01 0.2 0.2 30
45 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.1 0.2 0.2 30
46 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.001 0.2 0.2 30
47 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.9 0.2 30
48 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.05 0.2 30
49 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.9 30
50 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 0.05 30
Integrated quality and quantity modeling of a production line 147
Case # µ1 µ2 r1 r2 p1 p2 g1 g2 f1 f2 N
1 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 10
2 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 0
3 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 5
4 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 20
5 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 30
6 1.0 1.0 0.01 0.01 0.01 0.01 0.01 0.01 0.01 1.0 10
7 1.0 1.0 0.05 0.05 0.01 0.01 0.01 0.01 0.01 1.0 10
8 1.0 1.0 0.4 0.4 0.01 0.01 0.01 0.01 0.01 1.0 10
9 1.0 1.0 0.8 0.8 0.01 0.01 0.01 0.01 0.01 1.0 10
10 1.0 1.0 0.1 0.1 0.001 0.001 0.01 0.001 0.01 1.0 10
11 1.0 1.0 0.1 0.1 0.005 0.005 0.01 0.005 0.01 1.0 10
12 1.0 1.0 0.1 0.1 0.02 0.02 0.01 0.01 0.02 1.0 10
13 1.0 1.0 0.1 0.1 0.1 0.1 0.01 0.01 0.1 1.0 10
14 1.0 1.0 0.1 0.1 0.01 0.01 0.001 0.001 0.01 1.0 10
15 1.0 1.0 0.1 0.1 0.01 0.01 0.005 0.005 0.01 1.0 10
16 1.0 1.0 0.1 0.1 0.01 0.01 0.02 0.02 0.01 1.0 10
17 1.0 1.0 0.1 0.1 0.01 0.01 0.05 0.05 0.01 1.0 10
18 0.5 0.5 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 10
19 1.5 1.5 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 10
20 2.0 2.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 1.0 10
21 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.05 1.0 10
22 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 1.0 10
23 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.5 1.0 10
24 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.8 1.0 10
25 1.0 1.0 0.5 0.1 0.01 0.01 0.01 0.01 0.01 1.0 10
26 1.0 1.0 0.01 0.1 0.01 0.01 0.01 0.01 0.01 1.0 10
27 1.0 1.0 0.1 0.5 0.01 0.01 0.01 0.01 0.01 1.0 10
28 1.0 1.0 0.1 0.01 0.01 0.01 0.01 0.01 0.01 1.0 10
29 1.0 1.0 0.1 0.1 0.1 0.01 0.01 0.01 0.1 1.0 10
30 1.0 1.0 0.1 0.1 0.001 0.01 0.01 0.01 0.001 1.0 10
31 1.0 1.0 0.1 0.1 0.01 0.1 0.01 0.01 0.01 1.0 10
32 1.0 1.0 0.1 0.1 0.01 0.001 0.01 0.01 0.01 1.0 10
33 1.0 1.0 0.1 0.1 0.01 0.01 0.05 0.01 0.01 1.0 10
34 1.0 1.0 0.1 0.1 0.01 0.01 0.001 0.01 0.01 1.0 10
35 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.05 0.01 1.0 10
36 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.001 0.01 1.0 10
37 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.5 1.0 10
38 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.2 1.0 10
39 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 0.8 10
40 1.0 1.0 0.1 0.1 0.01 0.01 0.01 0.01 0.01 0.2 10
References
Abstract. We consider a cyclic flow line model that repetitively produces multiple
items in a cyclic order. We examine performance of stochastic cyclic flow line mod-
els with finite buffers of which processing times have exponential or phase-type
distributions. We develop an exact method for computing a two-station model by
making use of the matrix geometric structure of the associated Markov chain. We
present a computationally tractable approximate performance computing method
that decomposes the line model into a number of two-station submodels and param-
eterizing the submodels by propagating the starvation and blocking probabilities
through the adjacent submodels. We discuss performance characteristics including
comparison with random order processing and effects of the job variation and the
job processing sequence. We also report the accuracy of our proposed method.
1 Introduction
a, b, b, c, c, c, which is the same as the release sequence of the items into the line.
Cyclic flow lines are widely used for assembly lines or serial processing lines where
multiple types of items are simultaneously produced and the setup times are not
significant. Advantages of cyclic production over conventional batch production
or random order production include better utilization of the machines, simplified
flow control, continuous and smooth supply of complete part sets for downstream
assembly, timely delivery, and reduced work-in-progress inventory [14].
There have been studies on cyclic shops. Essential issues can be found in [1,
7, 10, 11, 14–18, 22]. Cyclic flow lines are often used for printed circuit board as-
sembly and electronics or other home appliance assembly, and integrated with an
accumulation-type conveyor system. Such a conveyor system allows only a few
parts to wait before each station. Such cyclic flow lines with blocking have been
examined [1, 18, 22]. They deal with scheduling issues for the cases where process
times are completely known. However, cyclic shops are subject to random disrup-
tions such as tool jamming and recovery, retrials of an assembly operation, etc.
These tend to contribute to random variation in job processing times. Scheduling
models often neglect transport times or tend to increase the processing times by the
transport times. Such approximate modeling simplifies the scheduling model, but
adds randomness of the transport times to the combined process times.
There are a few works on stochastic cyclic shop models. Rao and Jackson [20]
develop an approximate algorithm to compute the average cycle time for a cyclic
job shop with general processing time distributions, which makes use of Clark’s
approximation method for stochastic PERT networks. Bowman and Muckstadt [2]
deliberately develop a finite Markov chain model for a cyclic job shop with expo-
nential processing times and compute the average cycle time, but do not discuss
the queue length. Zhang and Graves [23] find the schedules that are least dis-
turbed by machine failures in a re-entrant cyclic flow shop. For cyclic flow lines,
Seo and Lee [21] examine the queue length distributions of the cases that have
exponential processing times and infinite buffers. Stochastic cyclic flow lines with
limited buffers have distinct performance characteristics and require a different per-
formance analysis method due to blocking. Therefore, it is necessary to examine
performance of stochastic flow lines with blocking. Karabati and Tan [13] propose
a heuristic procedure for scheduling stochastic cyclic transfer lines that move jobs
between the stations synchronously.
Stochastic cyclic flow lines are comparable to conventional tandem queues with
multiple customer classes. While the former produces different types of items in
a cyclic order, the latter processes the items in random order. Therefore, stochas-
tic flow lines require a distinct performance analysis method. Nonetheless, it is
expected that some ideas for analyzing tandem queues also will be useful for ex-
amining stochastic cyclic flow lines.
An important technique for analyzing a tandem queue model is to decompose
the model into multiple two-station models, each of which is modeled by an appro-
priately parameterized single-queue model, and approximate the performance of
the tandem queue from the performance estimates of the decomposed single-queue
models [4–6, 8]. While Dallery et al. [4, 5] and Gershwin [8] propose such decom-
position technique for transfer lines with unreliable machines and finite buffers,
Stochastic cyclic flow lines with blocking: Markovian models 151
the technique is popularly used for tandem queues. Various decomposed single-
queue models and different approximation schemes can be found in the survey on
modeling and analysis of tandem queues [6]. We note that most works on tandem
queues assume single customer class while a stochastic cyclic flow line processes
multiple customer classes simultaneously. It is expected that stochastic cyclic flow
lines with blocking require yet another decomposition and approximation method.
In this paper, we examine performance of cyclic flow line models that have finite
buffers and processing times of exponential or phase-type distributions. Phase-type
distributions are more realistic for modeling processing time distributions since any
distribution can be arbitrarily closely approximated by phase-type distributions.
While such a cyclic flow line model would be modeled by a finite continuous-
time Markov chain, the number of states tends to explode and the chain easily
becomes computationally intractable as the number of stations, the buffer capac-
ities, the number of job types, and the number of phases in the processing time
distributions increase. Therefore, we present a computationally tractable perfor-
mance approximation method that decomposes the line model into a number of
two-station submodels and appropriately parameterizes the mean processing times
of the decomposed submodels. We examine the performance characteristics and
report the experimental accuracy of the proposed algorithm. We also compare the
performance of cyclic production with that of random order production. The effect
of the job processing sequence is also discussed.
We first explain stochastic cyclic flow line models. We consider a cyclic flow line
that consists of (K + 1) stations (S0 , S1 , . . ., SK ). Each station has a single
machine. The first station S0 has an unlimited buffer. Each subsequent station
Si (i = 1, . . . , K) has an input buffer of capacity B − 1 (that is, each station has
capacity B). Each station can process the next job in the buffer after the previous
one completes and leaves the station. A job completed at a station immediately
leaves the station and enters the input buffer of the next station. When the next
input buffer is full, the job cannot leave the station and waits until the next buffer is
available. Such waiting is called blocking, more specifically blocking after service
(BAS). When there is no job available at the input buffer, the station is idle. This is
called starvation. The transport times of jobs between the stations are negligible or
included in the processing times. The jobs in an input buffer are processed in the
order of first come and first service. A job being processed at a station cannot be
preempted. We assume that the stations are all reliable and there is no breakdown.
There are enough jobs available and hence no shortage or starvation at the first
station S0 . The last station SK has no blocking since there is no next station. The
J types of jobs are repetitively loaded into the line in a predefined cyclic order.
Therefore, each station repeats the identical cyclic order of processing the jobs.
We assume that the processing times of the jobs at a station have exponential or
phase-type distributions. The setup times are negligible.
152 Y.-D. Lee and T.-E. Lee
3 Two-station models
λ1 λ2 λ3 λJ
µ1 µ2 µ3 µJ
211 321 431 1J1
λ2 λ4 µJ
λ3 λ1
µ1 µJ
µ2
µJ
µ1 µ2 µ3
J1(J-1) 12(J-1) 23(J-1) (J-1)J(J-1)
λJ λ1 λ2 λ J −1
µ1 µ2 µ3 µJ
11J 22J 33J JJJ
µJ
λ1 λ2 λ3 λJ
µ1 µ2 µ3 µJ
21(J+1) 32(J+1) 43(J+1) 1J(J+1)
λ2 λ4 µJ
λ3 λ1
µ1 µ2
µJ
µ1 µ2 µ3
b1B b2B b3B bJB
are governed by exponential processing times, the state transition process forms a
continuous-time Markov chain.
We observe that the diagram repeats an identical structure each multiple of
J for state variable n. The transition rates are marked on the corresponding
arcs. We therefore expect that the generator of the Markov chain has a repeat-
ing pattern. Define r ≡ B J
. We let π(m1 , m2 , n) denote the probability that
the line is at state (m1 , m2 , n) in the steady state. For exposition convenience,
we explain the case of J = 2. Define the steady state distribution vector as
π ≡ (πs , π0 , π1 , . . . , πk , . . . , πr−1 , πb ), where πs ≡ (π(1, s, 0), π(2, s, 0)), πk ≡
(π(2, 1, 2k+1), π(1, 2, 2k+1), π(1, 1, 2k+2), π(2, 2, 2k+2)), k = 0, 1, . . . , r−1,
and πb ≡ (π(b, 1, B), π(b, 2, B)). From the state transition diagram, we have the
generator matrix Q that is represented by block matrices with special structures as
⎛ ⎞
S1 S2 ···
⎜ ⎟
⎜S S A ··· ⎟
⎜ 3 4 0 ⎟
⎜ ⎟
⎜ A2 A1 A0 · · · ⎟
⎜ ⎟
⎜ ⎟
⎜ ⎟
⎜
Q=⎜ .
.. ⎟,
⎟
⎜ ⎟
⎜ ⎟
⎜ A 2 A1 A0 ⎟
⎜ ⎟
⎜ ⎟
⎜ A2 B4 B3 ⎟
⎝ ⎠
B1 B2
⎛ ⎞
−λ1 0 λ1 0
⎜ ⎟
⎜ 0 −λ2 0 λ2 ⎟
⎜ ⎟
where S1 = ⎜ ⎟,
⎜ 0 µ1 −(µ1 + λ2 ) 0 ⎟
⎝ ⎠
µ2 0 0 −(µ2 + λ1 )
⎛ ⎞
0 0 00
⎜ ⎟
⎜ 0 0 0 0⎟
⎜ ⎟
S 2 = A0 = ⎜ ⎟,
⎜ λ2 0 0 0 ⎟
⎝ ⎠
0 λ1 0 0
⎛ ⎞
−(µ1 +λ1 ) 0 λ1 0
⎜ ⎟
⎜ 0 −(µ2 +λ2 ) 0 λ2 ⎟
⎜ ⎟
S4 = B4 = A1 = ⎜ ⎟,
⎜ 0 µ1 −(µ1 +λ2 ) 0 ⎟
⎝ ⎠
µ2 0 0 −(µ2 +λ1 )
154 Y.-D. Lee and T.-E. Lee
⎛ ⎞
0 0 0 µ1
⎜ ⎟
⎜ 0 0 µ2 0 ⎟ 0 0 0 µ1 −µ1 0
⎜ ⎟
S 3 = A2 = ⎜ ⎟ , B1 = , B2 = ,
⎜0 0 0 0 ⎟ 0 0 µ2 0 0 −µ2
⎝ ⎠
00 0 0
⎛ ⎞
0 0
⎜ ⎟
⎜ 0 0 ⎟
⎜ ⎟
and B3 = ⎜ ⎟.
⎜ λ2 0 ⎟
⎝ ⎠
0 λ1
Generator Q for J > 2 that has the same block structure also can be similarly
identified. The size of each block matrix is determined by the number of job types,
J, and the station capacity, B. Ai is a J 2 × J 2 square matrix regardless of B.
It is easily seen that Q is irreducible and the finite Markov chain is positive
recurrent and hence ergodic. Therefore, the steady state probability π is the solution
of the balance equation πQ = 0. The balance equation can be efficiently solved
since Q has a special structure called a generalized birth-and-death process. Such
a structure is also a special case of the general matrix geometric structure. There
are two generally known strategies for solving the balance equation for such a
structured generator, matrix geometric technique [19] and recursive technique [3].
Buzacott and Kostelski [3] report that there is no significant difference between the
two methods in their accuracy, but the recursive method is more efficient than the
matrix geometric algorithm. There can be different implementations of the recursive
method depending on the detailed matrix structure. We adapt the recursive algorithm
of Hong et al. [12] that is used for two-station tandem queues with random failures
of stations.
From πQ = 0, we obtain the following equations.
πs S1 + π0 S3 = 0, (1)
πs S2 + π0 S4 + π1 A2 = 0, (2)
πk−1 A0 + πk A1 + πk+1 A2 = 0, k = 1, 2, . . . , r − 2, (3)
πr−2 A0 + πr−1 B4 + πb B1 = 0, and (4)
πr−1 B3 + πb B2 = 0. (5)
From (1), (2), (3), and (4), we derive following relationships. From (1) and (2),
πs = π0 T0 , where T0 = −S3 S1−1 , and (6)
π0 = π1 T1 , where T1 = −A2 (S4 + T0 S2 )−1 . (7)
From (3), we can derive
πk = πk+1 Tk+1 , where Tk+1 = −A2 (A1 + Tk A0 )−1 , k = 1, . . . , r − 2. (8)
From (4), we have
πr−1 = πb Tb , where Tb = −B1 (B4 + Tr−1 A0 )−1 . (9)
Stochastic cyclic flow lines with blocking: Markovian models 155
Algorithm: Two-station
Step 1. Set πb = 1 and SAV E = πb = 1.
Step 2. Compute πr−1 from (9).
Step 3. Compute πb from (5).
Step 4. If ||SAV E − πb || ≤ , go to Step 5.
Else SAV E = πb , and go to Step 2.
Step 5. Compute πr−1 , πr−2 , . . . , π0 , πs from (6)–(9).
Step 6. Normalize the steady state distribution vector π.
Vectors πb , πk , and πs are computed recursively by matrix operations. After
taking some initial value of πb , compute πk ’s and πb recursively until πb value con-
verges to a finite value. Then, π vector is normalized. The steady state queue length
distributions, the starvation probability, the blocking probability, the throughput
rate, and the mean queue length are computed, respectively, as
J
J
pn = π(m1 , m2 , n), n = 1, . . . , B,
m1 =1 m2 =1
J
ps = π(m1 , s, 0),
m1 =1
J
pb = π(b, m2 , B),
m2 =1
J
B
J
T = µj [ π(m1 , j, y)] + µJ π(b, J, B), and
j=1 y=1 m1 =1
B
J J
J
L= yπ(m1 , m2 , y) + Bπ(b, m2 , B).
y=1 m1 =1 m2 =1 m2 =1
Note that 1/T is the mean cycle time for all types of items. The mean cycle time
of job sets is J/T .
In order to analyze the performance of a line model with more than two stations, we
extend the decomposition technique that has been used for tandem queues [8]. The
156 Y.-D. Lee and T.-E. Lee
S
S0 B1 S1 B2 S2 B3 S3 B4 S4
S(1)
u
S (1) B(1) Sd(1)
S(2)
u
S (2) B(2) Sd(2)
S(3)
S (3) B(3) Sd(3)
u
S(4)
u
S (4) B(4) Sd(4)
procedure is outlined as follows. First, the line model is decomposed into K two-
station submodels as shown in Figure 2. Each two-station submodel S(i) consists
of upstream station S u (i), downstream station S d (i), and buffer B(i) between
them with the same capacity B − 1 as in the original line model S. Stations S u (i)
and S d (i) are parameterized to have the performances close to those of stations
Si−1 and Si , respectively, in the original line, which are subject to starvation and
blocking.
We first examine the decomposition method for the case where all processing times
are exponentially distributed. Let tj (i) denote the mean processing time of job j
at station i in the original line model. Let t(i) ≡ (t1 (i), . . . , tJ (i)) , i = 0, . . . , K,
be the mean processing time vector at station Si . Let tu (i) ≡ (tu1 (i), . . . , tuJ (i))
and td (i) ≡ (td1 (i), . . . , tdJ (i)) , i = 1, . . . , K, be the mean processing times of
S u (i) and S d (i) in submodel S(i), respectively. The processing capacity at each
station of a decomposed two-station submodel is parameterized to be as close as
possible to the effective processing capacity of the corresponding station in the
original line. The processing capacity of a station in the original line is reduced due
to starvation or blocking at the station. Therefore, the processing times tu (i) of the
upstream station S u (i) of each submodel S(i) are extended as much as the delays
due to starvation of the corresponding station Si−1 in the original line. Similarly,
the processing times td (i) of the downstream station S d (i) of each submodel S(i)
are extended as much as the delays due to blocking of the corresponding stations Si
in the original line. However, for the first submodel S(1), the processing times of
station S u (1) are kept same as those of S0 in the original line. It is because the first
station S0 is never starved. Similarly, for the last submodel S(K), the processing
times of S d (K) are kept same as those of SK in the original line because the last
Stochastic cyclic flow lines with blocking: Markovian models 157
S(i) S(i)
J1 J1
J2 J2
J3 J3
u d u
S (i) S (i) S (i) Sd(i)
a b
Fig. 3a,b. Propagation of starvation and blocking. a Starvation propagation. b Blocking
propagation
The mean processing times of the downstream station S d (i) of each submodel S(i)
are recursively computed from the mean processing times of the downstream station
S d (i + 1) and the blocking probability of the succeeding submodel S(i + 1). Such
kind of recursion is often called blocking propagation [6]. Figure 3 illustrates the
propagation mechanism of starvation and blocking. The mean processing time of a
specific job(shaded box) at S(i) is elongated by the starvation (or blocking) time,
which is the remaining processing time of the job in progress at the preceding (or
succeeding) station.
We now have a simultaneous equation system that has 2JK unknown param-
eters, tu (i) and td (i), i = 1, . . . , K, and 2JK independent equations. The perfor-
mance of the original line is then approximated by the decomposed submodels that
are parameterized by the decomposition procedure summarized below. Once each
submodel is parameterized based on the starvation and blocking probabilities of
the adjacent submodels, the starvation and blocking probabilities of each submodel
change due to the modifications in its processing times. Therefore, the submod-
els should be parameterized again based on the changed starvation and blocking
probabilities. The algorithm repeats such computing cycle until the process times
of the submodels do not change anymore. We note that our decomposition algo-
rithm is structurally similar to the well-known decomposition procedures of [4, 5, 8]
for conventional transfer lines or tandem queues. It is known that such algorithms
based on propagation of starvation and blocking converge. In fact, our algorithm
converged quickly, mostly within 10 iterations, for all experimental cases, which
are explained in Section 5. The computation times were within 1∼2 CPU seconds
at Pentimum 1 GHz PC.
Algorithm: Decomposition
Step 1. Initialize.
Set tu (1) ≡ t(0) and td (K) ≡ t(K).
Let td (i) = t(i), i = 1, . . . , K − 1.
Step 2. For i = 2, . . . , K,
compute ps (i − 1) from submodel S(i − 1), and
compute tu (i) using equation (12).
Step 3. For i = K − 1, . . . , 1,
compute pb (i + 1) from submodel S(i + 1), and
compute td (i) using equation (13).
Stochastic cyclic flow lines with blocking: Markovian models 159
5 Experiments
ͥ͡ ͦͥ͟
ͤͪ
ͥͪ͟
ͤͩ
v d10
v d10
Number of waiting
ͤͨ
Cycle time
ͥͥ͟
ͤͧ
v d3 v d3 max
max
ͤͦ
min ͤͪ͟ min
ͤͥ estimate estimate
ͤͤ
ͤͥ͟
jobs
ͤͣ
ͤ͢ ͣͪ͟
ͤ͡
ͣͪ ͣͥ͟
͢ ͦ ͢͡ ͣ͡ ͢ ͦ ͢͡ ͣ͡
Buffer size Buffer size
computer simulation program, which are hard to control. In fact, Figure 4 shows
that the confidence intervals are not sufficiently reduced. Nonetheless, we roughly
figure out the relative accuracy of the estimates by our proposed algorithm and how
much job variation increment increases the errors of our estimates. For instance,
among 8 comparisons in the table, 6 estimates by our approximation algorithm fall
within the confidence intervals. The figure shows that the errors tend to be larger
when the buffer size is smaller. It is because when the blocking probabilities are
higher, the blocking propagation procedure amplifies the blocking approximation
errors more. We also observe that job variation increment tends to increase the errors
of our algorithm in the cycle time estimates. It is because higher job variation causes
more blocking especially when the buffer size is small.
Stochastic cyclic flow lines with blocking: Markovian models 163
Cycle time
55
B=1, v=10
50
B=1, v=3
45
Seconds
99% C.I.
1 estimate
40
35 B=10, v=10
B=10, v=3
30 1 2 3 4 5 6 Random
Sequence
Fig. 5. Effects of job processing orders for M (4, 4)
We examine the effects of the job processing sequence on the performance and
compare the performance with that of random order job processing. There are
(J − 1)! cyclic sequences of processing J types of jobs.
Since the exact method is not available for the larger cases such as M (4, 4),
we should resort to simulation or our proposed approximation method. Although
the errors of our approximation method are small, mostly within 1∼2 percent, they
are biased. However, simulation tends to give less biased point estimates because
the point estimates are obtained by averaging out the estimates from many inde-
pendent replications. Furthermore, the confidence intervals provide information on
the relative accuracy of the point estimates. Therefore, we primarily use simula-
tion estimates to have more consistent performance estimates for examining the
effects of the job processing order. Table 3 shows the performance estimates for
each processing sequence for line model M(4,4) with W(0.8, 0.8, 0.8, υ). Figure 5
visualizes the cycle time estimates. The point estimates and the 99% confidence
intervals are obtained from 100 replications of simulation. CV in the table indi-
cates 100 times of the coefficient of variation of the performance estimates for the
sequences. For random order processing, the first station that is considered as an
arrival generator is modified to generate arrivals in random order while the long-
run proportion of job types is maintained. From Figure 5, we observe that while
the mean processing times are kept to be identical for all four cases, the relative
performance between the processing sequences are different for each case. There-
fore, the processing sequence should be carefully taken based on the performance
estimates. Our proposed procedure can efficiently compute the approximate cycle
time estimates within 1∼2 percent errors. As seen in Figures 4 and 5, the estimates
164 Y.-D. Lee and T.-E. Lee
CT L CT L
CT Q CT Q
mostly fall within or are close to the confidence intervals. Further, Figure 5 shows
that their changes for different sequences are consistent with those of the simulation
estimates. Even though our estimates tend to have larger errors when the buffer size
is smaller, the performance differences between the processing sequences become
larger for such case. Therefore, our approximation procedure can be effectively
used for selecting the optimal or near-optimal processing sequence.
We observe that the processing sequence significantly affects the performance,
especially when job variation υ is high and the buffer sizes are small. The cyclic
processing sequences outperform random order processing in most cases. When the
job variation is higher and the buffer capacities are smaller, the cyclic sequences
have larger performance differences, but the optimal cyclic sequence has much
better performance than random order processing. A good choice of the processing
sequence tends to minimize both the cycle time and the queue length.
Stochastic cyclic flow lines with blocking: Markovian models 165
7 Final remarks
References
1. Ahmadi RH, Wurgaft H (1994) Design for synchronized flow manufacturing. Manage-
ment Science 40(11): 1469–1483
2. Bowman RA, Muckstadt JA (1993) Stochastic analysis of cyclic schedules. Operations
Research 41(5): 947–958
3. Buzacott JA, Kostelski D (1987) Matrix-geometric and recursive algorithm solution of
a two-stage unreliable flow line. IIE Transaction 19(4): 429–438
4. Dallery Y, David R, Xie XL (1988) An efficient algorithm for analysis of transfer lines
with unreliable machines and finite buffers. IIE Transactions 20(3): 280–283
5. Dallery Y, David R, Xie XL (1989) Approximate analysis of transfer lines with un-
reliable machines and finite buffers. IEEE Transactions on Automatic Control 34(9):
943–953
6. Dallery Y, Gershwin SB (1992) Manufacturing flow lines: a review of models and
analytical results. Queueing Systems 12(1): 3–94
7. Dobson G, Yano CA (1994) Cyclic scheduling to minimize inventory in a batch flow
line. European Journal of Operational Research 75(2): 441–461
8. Gershwin SB (1987) An efficient decomposition method for the approximate evaluation
of tandem queues with finite storage space and blocking. Operations Research 35(2):
291–305
9. Gershwin SB, Schick IC (1983) Modeling and analysis of three-stage transfer lines
with unreliable machines and finite buffers. Operations Research 31(2): 354–380
10. Graves SC, Meal HC, Stefek D, Zeghmi AH (1983) Scheduling of re-entrant flow shops.
Journal of Operations Management 3(4): 197–207
11. Hall NG, Lee TE, Posner ME (2002) The complexity of cyclic shop scheduling prob-
lems. Journal of Scheduling 5(4): 307–327
12. Hong Y, Glassey CR, Seong D (1992) The analysis of a production line with unreliable
machines and random processing times. IIE Transactions 24(1): 77–83
13. Karabati S, Tan B (1998) Stochastic cyclic scheduling problem in synchronous as-
sembly and production lines. The Journal of the Operational Research Society 49(11):
1173–1187
14. Lee TE, Posner ME (1997) Performance measures and schedules in periodic job shops.
Operations Research 45(1): 72–91
15. Lee TE (2000) Stable earliest starting schedules for cyclic job shops: a linear system
approach. International Journal of Flexible Manufacturing Systems 12(1): 59–80
166 Y.-D. Lee and T.-E. Lee
16. Seo JW, Lee TE (2002) Steady state analysis of cyclic job shops with overtaking.
International Journal of Flexible Manufacturing Systems 14(4): 291–318
17. Kim JH, Lee TE, Lee HY, Park DB (2003) Scheduling analysis of time-constrained
dual-armed cluster tools. IEEE Transactions on Semiconductor Manufacturing 16(3):
521–534
18. McCormick ST, Pinedo ML, Shenker S, Wolf B (1989) Sequencing in an assembly line
with blocking to minimize cycle time. Operations Research 37(6): 925–935
19. Neuts MF (1981) Matrix-geometric solutions in stochastic models: an algorithmic ap-
proach. The Johns Hopkins University Press, Baltimore, MD
20. Rao US, Jackson PL (1996) Estimating performance measures in repetitive manufac-
turing environments via stochastic cyclic scheduling. IIE Transactions 28(11): 929–939
21. Seo JW, Lee TE (1996) Stochastic cyclic flow lines: non-blocking, Markovian models.
Journal of Operational Research Society 49(5): 537–548
22. Wittrock RJ (1985) Scheduling algorithm for flexible flow lines. IBM Journal of Re-
search and Development 29(4): 401–412
23. Zhang H, Graves SC (1997) Cyclic scheduling in a stochastic environment. Operations
Research 45(6): 894–903
Section III: Queueing Network Models
of Manufacturing Systems
Performance analysis of multi-server tandem queues
with finite buffers and blocking
Marcel van Vuuren1 , Ivo J.B.F. Adan1 , and Simone A.E. Resing-Sassen2
1
Eindhoven University of Technology, P.O. Box 513, 5600 MB Eindhoven,
The Netherlands (e-mail: [email protected], [email protected])
2
CQM BV, P.O. Box 414, 5600 AK Eindhoven, The Netherlands
(e-mail: [email protected])
Abstract. In this paper we study multi-server tandem queues with finite buffers
and blocking after service. The service times are generally distributed. We develop
an efficient approximation method to determine performance characteristics such as
the throughput and mean sojourn times. The method is based on decomposition into
two-station subsystems, the parameters of which are determined by iteration. For the
analysis of the subsystems we developed a spectral expansion method. Comparison
with simulation shows that the approximation method produces accurate results.
So it is useful for the design and analysis of production lines.
1 Introduction
Queueing networks with finite buffers have been studied extensively in the litera-
ture; see, e.g., Dallery and Gershwin [6], Perros [17, 18], and Perros and Altiok [19],
and the references therein. Most studies, however, consider single-server models.
The few references dealing with multi-server models typically assume exponential
service times. In this paper we focus on multi-server tandem queues with general
service times, finite buffers and Blocking After Service (BAS).
Models with finite buffers and phase-type service times can be represented by
finite state Markov chains. Hence, in theory, they can be analyzed exactly. However,
the number of states of the Markov chain can be very large, which makes numerical
solutions intractable. In practice, only small systems with one or two queues can
be solved exactly; for exact methods we refer to Perros [18].
We develop an efficient method to approximate performance characteristics
such as the throughput and the mean sojourn time. The method only needs the
first two moments of the service time and it decomposes the tandem queue into
170 M. van Vuuren et al.
Fig. 1. The tandem queue L and its decomposition into three subsystems L1 , L2 and L3
In this section we describe how the service times of the arrival and departure servers
in subsystem Li are modelled.
The service-time Di of a departure-server in subsystem Li is approximated as
follows. We define bi,j as the probability that just after service completion of a server
in server-group Mi , exactly j servers of server-group Mi are blocked. This means
that, with probability bi,j , a server in server-group Mi has to wait for one residual
inter-departure time and j − 1 full inter-departure times of the next server-group
Mi+1 before the customer can leave the server. The inter-departure times of server-
group Mi+1 are assumed to be independent and distributed as the inter-departure
times of the superposition of mi+1 independent service processes, each with service
times Di+1 ; the residual inter-departure time is approximated by the equilibrium
residual inter-departure time of the superposition of these service processes. Let
the random variable SDi+1 denote the inter-departure time of server-group Mi+1
and RSDi+1 the residual inter-departure time. Figure 2 displays a representation
of the service time of a departure-server of subsystem Li .
In the appendix it is explained how the rates and coefficients of variation of
SDi+1 and RSDi+1 can be determined. If also the blocking probabilities bi,j
Performance analysis of multi-server tandem queues 173
are known, then we can determine the rate µd,i and coefficient of variation cd,i
of the service time Di of a departure-server of subsystem Li . The distribution of
Di is approximated by fitting an Erlangk−1,k or Coxian2 distribution on µd,i and
cd,i , depending on whether c2d,i is less or greater than 1/2. More specifically, if
c2d,i > 1/2, then the rate and coefficient of variation of the Coxian2 distribution
with density
µ1 µ2 −µ2 t
f (t) = (1 − q)µ1 e−µ1 t + q e − e−µ1 t , t ≥ 0,
µ1 − µ2
matches with µd,i and cd,i , provided the parameters µ1 , µ2 and q are chosen as (cf.
Marie [14]):
1
µ1 = 2µd,i , q = 2 , µ2 = µ1 q. (1)
2cd,i
If 1/k ≤ c2d,i ≤ 1/(k − 1) for some k > 2, then the rate and coefficient of variation
of the Erlangk−1,k with density
tk−2 −µt tk−1 −µt
f (t) = pµk−1 e + (1 − p)µk e , t ≥ 0,
(k − 2)! (k − 1)!
matches with µd,i and cd,i if the parameters µ and p are chosen as (cf. Tijms [22]):
kc2d,i − k(1 + c2d,i ) − k 2 c2d,i
p= , µ = (k − p)µd,i . (2)
1 + c2d,i
Of course, also other phase-type distributions may be fitted on the rate and coef-
ficient of variation of Di , but numerical experiments suggest that other distributions
only have a minor effect on the results, as shown in [10].
The service times Ai of the arrival-servers in subsystem Li are modelled simi-
larly. Instead of bi,j we now use si,j defined as the probability that just after service
completion of a server in server-group Mi , exactly j servers of Mi are starved.
This means that, with probability si,j , a server in server-group Mi has to wait one
174 M. van Vuuren et al.
residual inter-departure time and j − 1 full inter-departure times from the preced-
ing server-group Mi−1 . Figure 3 displays a representation of the service time of an
arrival-server of subsystem Li .
indicates the number of blocked arrival servers. The state variable j (k) indicates
the phase of the service time of the super arrival (departure) server. If i ≤ md + b,
then the service time of the super arrival server consists of l phases; the number of
phases depends on i for i > md + b. Similarly, the number of phases of the service
time of the super departure server is n for i ≥ md , and it depends on i for i < md .
The steady-state distribution of this Markov process can be determined effi-
ciently by using the spectral expansion method, see e.g. Mitrani [16]. Using the
spectral expansion method, Bertsimas [1] analysed a multi-server system with an
infinite buffer; we will adapt this method for finite buffer systems. The advantage of
the spectral expansion method is that the time to solve a subsystem is independent
of the size of the buffer.
Below we formulate the equilibrium equations for the equilibrium probabilities
P (i, j, k). Only the equations in the states (i, j, k) with md < i < md + b are
presented; the form of the equations in the other states appears to be of minor
importance to the analysis.
So, for md < i < md + b we have:
l n
P (i, 1, 1)(λ1 +µ1 ) = (1 − pj )λj P (i−1, j, 1)+ (1−qk )µk P (i+1, 1, k)(3)
j=1 k=1
n
P (i, j, 1)(λj + µ1 ) = pj−1 λj−1 P (i, j − 1, 1) + (1 − qk )µk P (i + 1, j, k),
k=1
j = 2, . . . , l (4)
l
P (i, 1, k)(λ1 + µk ) = qk−1 µk−1 P (i, 1, k − 1) + (1 − pj )λj P (i − 1, j, k),
j=1
k = 2, . . . , n (5)
P (i, j, k)(λj + µk ) = pj−1 λj−1 P (i, j − 1, k) + qk−1 µk−1 P (i, j, k − 1),
j = 2, . . . , l, k = 2, . . . , n. (6)
We are going to use the separation of variables technique presented in Mickens
[15], by assuming that the equilibrium probabilities P (i, j, k) are of the form
P (i, j, k) = Dj Rk wi , md ≤ i ≤ md + b, 2 ≤ j ≤ l, 2 ≤ k ≤ n. (7)
Substituting (7) in the equilibrium equations (3)–(6) and dividing by common pow-
ers of w yields:
l n
1
D1 R1 (λ1 +µ1 ) = (1 − pj )λj Dj R1 + w (1 − qk )µk D1 Rk (8)
w j=1
k=1
n
Dj R1 (λj +µ1 ) = pj−1 λj−1 Dj−1 R1 +w (1−qk )µk Dj Rk , 2≤j≤l (9)
k=1
l
1
D1 Rk (λ1 +µk ) = (1−pj )λj Dj Rk +qk−1 µk−1 D1 Rk−1 , 2≤k≤n (10)
w j=1
Dj Rk (λj +µk ) = pj−1 λj−1 Dj−1 Rk + qk−1 µk−1 Dj Rk−1
2 ≤ j ≤ l, 2 ≤ k ≤ n (11)
176 M. van Vuuren et al.
mb ≤ i ≤ md + b, 1 ≤ j ≤ l, 1 ≤ k ≤ n,
where Bt = D1,t R1,t is some constant. Since the equilibrium equations are linear,
any linear combination of the above solutions satisfies (3)–(6). Hence, the general
solution of (3)–(6) is given by
l+n
j−1
k−1
- pr λ r - qr µr
P (i, j, k) = Bt wti ,
t=1 r=1
x(w t ) + λ r+1 r=1
−x(wt ) + µr+1
mb ≤ i ≤ md + b, 1 ≤ j ≤ l, 1 ≤ k ≤ n.
Performance analysis of multi-server tandem queues 177
5 Iterative algorithm
We now describe the iterative algorithm for approximating the performance char-
acteristics of tandem queue L. The algorithm is based on the decomposition of L
in M − 1 subsystems L1 , L2 , . . . , LM −1 . Before going into detail in Section 5.2,
we present the outline of the algorithm in Section 5.1.
Step 0: Initialization: The first step of the algorithm is to set bi,j = 0 for all i and
j. This means that we initially assume that there is no blocking. This also means
that the random variables Di are initially the same as the service times Si .
T i = ⎝1 − pi,j ⎠ mi µd,i +
(n) (n) (n−1) (n) (n−1)
pi,j jµd,i . (19)
j=0 j=1
We also determine new estimates for the probabilities bi−1,j that j servers of
server-group Mi−1 are blocked after service completion of a server in server-group
Mi−1 and the probabilities si,j that j servers of server-group Mi are starved after
service completion of a server in server-group Mi .
We perform Step 1 for every subsystem from L1 up to LM −1 .
Performance analysis of multi-server tandem queues 179
Step 2: Service times of departure servers: Now we have new information about
the departure processes of the subsystems. So we can again calculate the first two
moments of the service times of the departure-servers, starting from DM −2 down
to D1 . Note that DM −1 is always the same as SM −1 , because the servers in server-
group MM −1 can never be blocked.
A new estimate for the rate µd,i of Di is determined from (cf. (18))
(n)
(n) Ti+1
µd,i = mi −1 (n) mi −1 (n)
(20)
(1 − j=0 pi,j )mi + j=1 pi,j j
The calculation of a new estimate for the coefficient of variation cd,i of Di is similar
to the one of Ai .
Convergence criterion: After Step 1 and 2 we check whether the iterative algo-
rithm has converged by comparing the departure rates in the (n − 1)-th and k-th
iteration. We decide to stop when the sum of the absolute values of the differ-
ences between these rates is less than ε; otherwise we repeat Step 1 and 2. So the
convergence criterion is
M −1
(n) (n−1)
µd,i − µd,i < ε.
i=1
Of course, we may use other stop-criteria as well; for example, we may consider
the throughput instead of the departure rates. The bottom line is that we go on until
all parameters do not change anymore.
Remark. Equality of throughputs.
It is easily seen that, after convergence, the throughputs in all subsystems are
(n) (n−1)
equal. Let us assume that the iterative algorithm has converged, so µd,i = µd,i
for all i = 1, . . . , M − 1. From equations (19) and (20) we find the following:
⎛ ⎞
m i −1 m
i −1
T i = ⎝1 − pi,j ⎠ mi µd,i +
(n) (n) (n−1) (n) (n−1)
pi,j jµd,i
j=0 j=1
⎛ ⎞
m
i −1 m
i −1
= ⎝1 − pi,j ⎠ mi µd,i +
(n) (n) (n) (n)
pi,j jµd,i
j=0 j=1
(n)
= Ti+1 .
Hence we can conclude that the throughputs in all subsystems are the same after
convergence.
of phases of the distribution of one arrival process and kd is the number of phases
of the distribution of one departure process. Then, the time complexity of one it-
eration becomes O(M maxi ((mi + mi−1 )(ki ki−1 )3 )). This means that the time
complexity is polynomial and it doesn’t depend on the sizes of the buffers.
6 Numerical results
In this section we present some numerical results. To investigate the quality of our
method we compare it with discrete event simulation. After that, we compare our
method with the method developed by Tahilramani et al. [21], which is implemented
in QNAT [25].
In order to investigate the quality of our method we compare the throughput and
the mean sojourn time with the ones produced by discrete event simulation. We are
especially interested in investigating for which set of input-parameters our method
gives satisfying results. Each simulation run is sufficiently long such that the widths
of the 95% confidence intervals of the throughput and the mean sojourn time are
smaller than 1%.
In order to test the quality of the method we use a broad set of parameters. We
test two different lengths M of tandem queues, namely with 4 and 8 server-groups.
For each tandem queue we vary the number of servers mi in the server-groups; we
use tandems with 1 server per server-group, 5 servers per server-group and with the
sequence (4, 1, 2, 8). We also vary the level of balance in the tandem queue; every
server-group has a maximum total rate of 1 and the group right after the middle
can have a total rate of 1, 1.1, 1.2, 1.5 and 2. The coefficient of variation of the
service times varies between 0.1, 0.2, 0.5, 1, 1.5 and 2. Finally we vary the buffer
sizes between 0, 2, 5 and 10. This leads to a total of 720 test-cases. The results for
each category are summarized in Table 1 up to 5. Each table lists the average error
in the throughput and the mean sojourn time compared with the simulation results.
Each table also gives for 4 error-ranges the percentage of the cases which fall in
that range. The results for a selection of 54 cases can be found in Tables 6 and 7.
Table 1. Overall results for tandem queues with different buffer sizes
0 5.7% 55.0% 35.0% 4.4% 5.6% 6.8% 42.8% 35.0% 14.4% 7.8%
2 3.2% 76.1% 22.8% 1.1% 0.0% 4.7% 57.2% 35.0% 7.2% 0.6%
5 2.1% 90.6% 9.4% 0.0% 0.0% 4.5% 60.6% 32.2% 7.2% 0.0%
10 1.4% 95.6% 4.4% 0.0% 0.0% 5.1% 53.3% 34.4% 12.2% 0.0%
Performance analysis of multi-server tandem queues 181
Table 2. Overall results for tandem queues with different balancing rates
1.0 3.3% 76.4% 20.8% 1.4% 1.4% 3.4% 74.3% 22.2% 2.1% 1.4%
1.1 3.1% 78.5% 18.1% 2.1% 1.4% 4.0% 68.1% 27.1% 3.5% 1.4%
1.2 3.0% 79.2% 18.8% 0.7% 1.4% 4.6% 59.7% 34.7% 4.2% 1.4%
1.5 3.0% 81.3% 16.0% 1.4% 1.4% 6.5% 38.2% 43.1% 16.7% 2.1%
2.0 3.1% 81.3% 16.0% 1.4% 1.4% 7.9% 27.1% 43.8% 25.0% 4.2%
Table 3. Overall results for tandem queues with different coefficients of variation of the
service times
0.1 4.4% 54.2% 44.2% 1.7% 0.0% 3.1% 77.5% 21.7% 0.8% 0.0%
0.2 2.6% 88.3% 11.7% 0.0% 0.0% 3.4% 75.8% 22.5% 1.7% 0.0%
0.5 2.2% 90.8% 9.2% 0.0% 0.0% 4.5% 60.8% 32.5% 6.7% 0.0%
1.0 1.5% 93.3% 2.5% 4.2% 0.0% 4.1% 64.2% 30.0% 5.0% 0.8%
1.5 3.0% 82.5% 13.3% 0.0% 4.2% 7.5% 25.8% 54.2% 15.0% 5.0%
2.0 4.8% 66.7% 26.7% 2.5% 4.2% 9.1% 16.7% 44.2% 32.5% 6.7%
Table 4. Overall results for tandem queues with a different number of servers per server-
group
All 1 2.9% 83.8% 9.2% 2.9% 4.2% 5.9% 46.3% 39.2% 10.0% 4.6%
All 5 3.8% 68.3% 30.8% 0.8% 0.0% 4.6% 60.0% 29.2% 10.8% 0.0%
Mixed 2.6% 85.8% 13.8% 0.4% 0.0% 5.3% 54.2% 34.2% 10.0% 1.7%
We may conclude the following from the above results. First, we see in Table 1
that the performance of the approximation becomes better when the buffer sizes
increase. This may be due to less dependencies between the servers-groups when
the buffers are large.
We also notice that the performance is better for balanced lines (Table 2); for
unbalanced lines, especially the estimate for the mean sojourn time is not as good
as for balanced lines. If we look at the coefficients of variation of the service times
(Table 3), we get the best approximations for the throughput when the coefficients
182 M. van Vuuren et al.
4 2.3% 87.2% 12.2% 0.6% 0.0% 4.7% 57.5% 32.8% 9.7% 0.0%
8 3.9% 71.4% 23.6% 2.2% 2.8% 5.8% 49.4% 35.6% 10.8% 4.2%
of variation are 1, and also the estimate for the mean sojourn time is better for small
coefficients of variation.
The quality of the results seems to be rather insensitive to the number of servers
per server-group (Table 4), in spite of the super-server approximation used for
multi-server models. Finally we may conclude from Table 5 that the results are
better for shorter tandem queues.
Most crucial to the quality of the approximation of the throughput appears to be
the buffer-size. For the sojourn time this appears to be the coefficient of variation of
the service time. In Figures 4 and 5 we present a scatter-plot of simulation results
versus approximation results for the throughput and mean sojourn times; the plotted
cases are the same as in Tables 6 and 7. The results of the throughput are split-up
184 M. van Vuuren et al.
according to the buffer-size; the one for the sojourn times are split-up according to
the squared coefficient of variation of the service times.
Overall we can say that the approximation produces accurate results in most
cases. In the majority of the cases the error of the throughput is within 5% of the
simulation and the error of the mean sojourn time is within 10% of the simulation
(see also Tables 6 and 7). The worst performance is obtained for unbalanced lines
with zero buffers and high coefficients of variation of the service times. But these
cases are unlikely (and undesired) to occur in practice.
The computation times are very short. On a modern computer the computation
times are much less than a second in most cases, only in cases with service times
with low coefficients of variation and 1 server per server-group the computation
times increase to a few seconds. Therefore, for the design of production lines, this
is a very useful approximation method.
We also compare the present method with QNAT, a method developed by Tahilra-
mani et al. [21]. We use a tandem queue with four server-groups. It was only possible
to test cases where the first server-group consists of a single exponential server. The
reason is that the two methods assume a different arrival process to the system. Both
processes, however, coincide for the special case of a single exponential server at
the beginning of the line. We varied the number of servers per server-group and the
size of buffers. Table 8 shows the results.
Performance analysis of multi-server tandem queues 185
Fig. 5. Scatter-plot of the mean sojourn time of 54 cases split up by coefficient of variation
Table 8. Comparison of our method with QNAT
We see that the present approximation method is much more stable than QNAT
and gives in almost all cases better results. Especially the approximation of the mean
sojourn time is much better; in a number of cases QNAT is not able to produce an
approximation of the mean sojourn time. Of course, one should be careful with
drawing conclusions from this limited set of cases. Table 8 only gives an indication
of how the two methods perform.
186 M. van Vuuren et al.
Table 9. Parameters for the production line for the production of bulbs
1 2 5.73 0.96 −
2 8 1.53 0.09 21
3 4 3.43 0.80 11
4 1 32.18 0.57 34
5 4 16.12 0.96 19
We only have data of the throughput and not of the mean sojourn time of the
line, so we can only test the approximation for the throughput. The output of the
production line based on the measured data is 11.34 products per time unit. If we
simulate this production line, we obtain a throughput of 11.41 products per time
unit. The throughput given by our approximation method is 11.26, so in this case
the approximation is a good prediction for the actual throughput.
7 Concluding remarks
We assume that the service times of each service process are Coxian2 distributed
with the same parameters. The rate of the first phase is µ1 , the rate of the second
phase is µ2 and the probability that the second phase is needed is q. The distribution
of an arbitrary inter-departure time of the superposition of m service processes can
be described by a phase-type distribution with m+1 phases, numbered 0, 1, . . . , m.
In phase i exactly i service processes are in the second phase of the service time and
m − i service processes are in the first phase. A phase diagram of the phase-type
distribution of an arbitrary inter-departure time is shown in Figure 6. The probability
to start in phase i is denoted by ai , i = 0, . . . , m − 1. The sojourn time in phase
i is exponentially distributed with rate R(i), and pi is the probability to continue
with phase i + 1 after completion of phase i. Now we explain how to compute the
parameters ai , R(i) and pi .
The probability ai can be interpreted as follows. It is the probability that i service
processes are in phase 2 just after a departure (i.e., service completion). There is
at least one process in phase 1, namely the one that generated the departure. Since
the service processes are mutually independent, the number of service processes
in phase 2 is binomially distributed with m − 1 trials and success probability p.
188 M. van Vuuren et al.
The success probability is equal to the fraction of time a single service process is
in phase 2, so
qµ1
p= .
qµ+ µ2
Hence, for the initial probability ai we get
i m−1−i
m−1 qµ1 µ2
ai = (21)
i qµ1 + µ2 qµ1 + µ2
To determine the rate R(i), note that in state i there are i processes in phase 2 and
m − i in phase 1, so the total rate at which one of the service processes completes
a service phase is equal to
R(i) = (m − i)µ1 + iµ2 (22)
It remains to find pi , the probability that there is no departure after phase i. In phase
i three things may happen:
– Case (i): A service process completes phase 1 and immediately continues with
phase 2;
– Case (ii): A service process completes phase 1 and generates a departure;
– Case (iii): A service process completes phase 2 (and thus always generates a
departure).
Clearly, pi is the probability that case (i) happens, so
q(m − i)µi
pi = (23)
R(i)
Now the parameters of the phase-type distribution are known, we can determine
its first two moments. Let Xi denote the total sojourn time, given that we start in
phase i, i = 0, 1, . . . , m. Starting with
1 2 2
EXm = , EXm = ,
R(m) R(m)2
the first two moments of Xi can be calculated from i = m − 1 down to i = 0 by
using
1
EXi = + pi EXi , (24)
R(i)
2 2EXi+1
EXi2 = + p i + EX 2
i+1 . (25)
R(i)2 R(i)
Then the rate µs and coefficient of variation cs of an arbitrary inter-departure time
of the superposition of m service processes follow from
m
1 1 q
µ−1
s = ai EX i = + , (26)
i=0
m µ1 µ2
m
c2s = µ2s ai EXi2 − 1 (27)
i=0
Performance analysis of multi-server tandem queues 189
Now the service times of each service process are assumed to be Erlangk−1,k
distributed, i.e., with probability p (respectively 1 − p) a service time consists of
k − 1 (respectively k) exponential phases with parameter µ. Clearly, the time that
elapses until one of the m service processes completes a service phase is exponential
with parameter mµ. The number of service phases completions before one of the
service processes generates a departure ranges from 1 up to m(k − 1) + 1. So the
distribution of an arbitrary inter-departure time of the superposition of m service
processes is a mixture of Erlang distributions; with probability pi it consists of i
exponential phases with parameter mµ, i = 1, . . . , m(k − 1) + 1. Figure 7 depicts
the phase diagram. Below we show how to determine the probabilities pi .
An arbitrary inter-departure time of the superposition of m service processes is
the minimum of m − 1 equilibrium residual service times and one full service time.
Both residual and full service time have a (different) mixed Erlang distribution. In
particular, the residual service consists with probability ri of i phases with parameter
µ, where
1/(k − p), i = 1, 2, . . . , k − 1;
ri =
(1 − p)/(k − p), i = k.
The minimum of two mixed Erlang service times has again a mixed Erlang distribu-
tion; below we indicate how the parameters of the distribution of the minimum can
be determined. Then repeated application of this procedure yields the minimum of
m mixed Erlang service times.
Let X1 and X2 be two independent random variables with mixed Erlang dis-
tributions, i.e., with probability qk,i the random variable Xk (k = 1, 2) consists of
i exponential phases with parameter µk , i = 1, . . . , nk . Then the minimum of X1
To determine qi (j) note that the ith phase transition of the minimum can be due to
either X1 or X2 . If X1 makes the last transition, then X1 clearly consists of exactly
j phases and X2 of at least i − j + 1 phases; the probability that X2 makes i − j
transitions before the jth transition of X1 is negative-binomially distributed with
parameters j and µ1 /(µ1 + µ2 ). The result is similar if X2 instead of X1 makes
the last transition. Hence, we obtain
⎛ ⎞
j i−j n2
i−1 µ1 µ2
qi (j) = q1,j ⎝ q2,k ⎠
j−1 µ1 + µ2 µ1 + µ2
k=i−j+1
⎛ ⎞
j i−j n1
i−1 µ1 µ2 ⎝
+ q1,k ⎠ q2,i−j ,
j µ1 + µ2 µ1 + µ2
k=j+1
1 ≤ i ≤ n1 + n2 − 1, 0 ≤ j ≤ i,
where by convention, q1,0 = q2,0 = 0.
By repeated application of the above procedure we can find the probability pi
that the distribution of an arbitrary inter-departure time of the superposition of m
Erlangk−1,k service processes consists of exactly i service phases with parameter
mµ, i = 1, 2, . . . , m(k − 1) + 1. It is now easy to determine the rate µs and
coefficient of variation cs of an arbitrary inter-departure time, yielding
−1 1 p(k − 1) (1 − p)k k−p
µs = + = ,
m µ µ mµ
and, by using that the second moment of an Ek distribution with scale parameter
µ is k(k + 1)/µ2 ,
m(k−1)+1 m(k−1)+1
i(i + 1) 1
c2s = µ2s pi − 1 = −1 + pi i(i + 1).
i=1
(mµ)2 (k − p)2 i=1
To determine the first two moments of the equilibrium residual inter-departure time
of the superposition of m independent service processes we adopt the following
simple approach.
Let the random variable D denote an arbitrary inter-departure time and let R
denote the equilibrium residual inter-departure time. It is well known that
E(D2 ) E(D3 )
E(R) = , E(R2 ) = .
2E(D) 3E(D)
Performance analysis of multi-server tandem queues 191
In the previous sections we have shown how the first two moments of D can be
determined in case of Coxian2 and Erlangk−1,k service times. Its third moment is
approximated by the third moment of the distribution fitted on the first two moments
of D, according to the recipe in Section 3.
References
1 Introduction
In 1960, Clark and Scarf [10] initiated the research on the coordination of multi-
stage, serial, uncapacitated inventory systems with stochastic demand and constant
lead times. Their work received considerable attention in the years that followed
and spawned a large amount of follow-on research. Much of that research evolved
around variants of the base stock control system. Research on the coordination of
multi-stage, serial, production/inventory systems having networks of stations with
limited capacity, on the other hand, has been directed mostly towards variants of
the kanban control system. In this paper, we develop an analytical approximation
method for the performance evaluation of an echelon kanban control system, used
for the coordination of production in a multi-stage, serial production/inventory
system. We test the behavior of this method with several numerical examples.
The term “echelon kanban” was introduced in [19]. The basic principle of the
operation of the echelon kanban control system is very simple: When a part leaves
Correspondence to: G. Liberopoulos
194 S. Koukoumialos and G. Liberopoulos
the last stage of the system to satisfy a customer demand, a new part is demanded
and authorized to be released into each stage. It is worth noting that the echelon
kanban control system is equivalent to the integral control system described in [8].
The echelon kanban control system differs from the conventional kanban control
system, which is referred to as installation kanban control system or policy in [19],
in that in the conventional kanban control system, a new part is demanded and
authorized to be released into a stage when a part leaves this particular stage and
not when a part leaves the last stage, as is the case with the echelon kanban control
system. This implies that in the conventional kanban control system, the placement
of a demand and an authorization for the production of a new part into a stage is
based on local information from this stage, whereas in the echelon kanban control
system, it is based on global information from the last stage. This constitutes a
potential advantage of the echelon kanban control system over the conventional
kanban control system. Moreover, the echelon kanban control system, just like the
conventional kanban control system, depends on only one parameter per stage, the
number of echelon kanbans, as we will see later on, and is therefore simpler to
optimize and implement than more complicated kanban-type control systems that
depend of two parameters per stage, such as the generalized kanban control system
[7] and the extended kanban control system [12]. These two apparent advantages of
the echelon kanban control system motivated our effort to develop an approximation
method for its performance evaluation.
Kanban-type production/inventory systems have often been modeled as queue-
ing networks in the literature. Consequently, most of the techniques that have been
developed for the analysis of kanban-type production/inventory systems are based
on methods for the performance evaluation of queueing networks. Exact analytical
solutions exist for a class of queueing networks known as separable, in which the
steady-state joint probabilities have a product-form solution. Jackson [18] was the
first to show that the steady-state joint probability of an open queueing network with
Poisson arrivals, exponential service times, probabilistic routing, and first-come-
first-served (FCFS) service disciplines has a product-form solution, where each
station of the network can be analyzed in isolation as an M/M/1 queue. For closed
queueing networks of the Jackson type, Gordon and Newell [17] showed that an
analytical, product-form solution also exists. The performance parameters of such
networks can be obtained using efficient algorithms, such as the mean value analy-
sis (MVA) algorithm [22] and the convolution algorithm [9]. The BCMP theorem
[1] summarizes extensions of product-form networks that incorporate alternative
service disciplines and several classes of customers.
Since the class of queueing networks for which an exact solution is known
(separable networks) is too restrictive for modeling and analyzing real systems,
much work has been devoted to the development of approximation methods for the
analysis of non-separable networks. Whitt [27] presented an approximation method
for the analysis of a general open queueing network that is based on decomposing
the network into a set of GI/GI/1 queues and analyzing each queue in isolation. In
the case of closed queueing networks, the approximation methods are for the most
part based on two approaches. The first approach relies on heuristic extensions of
the MVA algorithm (e.g. [23]). The second approach relies on approximating the
An analytical method for the performance evaluation 195
M1 M2 M3 M4 M5 M6 M7 M8 M9
In this section, we give a precise description of the operation of the echelon kan-
ban control system by means of a simple example. In this example, we consider
a production system that consists of M = 9 machines in series, labeled M1 to
M9, produces a single part type, and does not involve any batching, reworking or
scrapping of parts. Each machine has a random processing time. All parts visit
successively machines M1 to M9. The production system is decomposed into N =
3 stages. Each stage is a production/inventory system consisting of a manufactur-
ing process and an output buffer. The output buffer stores the finished parts of the
stage. The manufacturing process consists of a subset of machines of the original
manufacturing system and contains parts that are in service or waiting for service
on the machines. These parts represent the work in process (WIP) of the stage and
are used to supply the output buffer. In the example, each stage consists of three
machines. More specifically, the sets of machines {M1, M2, M3}, {M4, M5, M6}
and {M7, M8, M9} belong to stages 1, 2 and 3, respectively. The decomposition
of the production system into three stages is illustrated in Figure 1.
Each stage has associated with it a number of echelon kanbans that are used
to demand and authorize the release of parts into this stage. An echelon kanban
of a particular stage traces a closed path through this stage and all its downstream
stages. The number of echelon kanbans of stage i is fixed and equal to Ki . There
must be at least one echelon kanban of stage i available in order to release a new
part into this stage. If such a kanban is available, the kanban is attached onto the
part and follows it through the system until the output buffer of the last stage. Since
an echelon kanban of stage i is attached to every part in any stage from i to N , the
number of parts in stages i to N is limited by Ki .
Parts that are in the output buffer of stage N are the finished parts of the
production system. These parts are used to satisfy customer demands. When a
customer demand arrives to the system, a demand for the delivery of a finished
part from the output buffer of the last stage to the customer is placed. If there
are no finished parts in the output buffer of the last stage, the demand cannot be
immediately satisfied and is backordered until a finished part becomes available. If
there is at least one finished part in the output buffer of the last stage, this part is
delivered to the customer after releasing the kanbans of all the stages (1, 2, and 3,
in the example) that were attached to it, hence the demand is immediately satisfied.
The released kanbans are immediately transferred upstream to their corresponding
stages. The kanban of stage i carries with it a demand for the production of a new
An analytical method for the performance evaluation 197
stage−i finished part and an authorization to release a finished part from the output
buffer of stage i − 1 into stage i. When a finished part of stage i − 1 is transferred
to stage i, the stage-i kanban is attached to it on top of the kanbans of stages 1 to
i − 1, which have already been attached to the part at previous stages. With this in
mind, we can just as well assume that
Ki ≥ Ki+1 , i = 1, . . ., N − 1. (1)
L1 J1 L2 J2 L3 J3
S1 S2 S3 PA1 S4 S5 S6 PA2 S7 S8 S9 PA3
DA2 DA3 D4
Customer
K3 Demands
K2
K1
Fig. 2. Queueing network model of the echolon kanban control system of Figure 1
198 S. Koukoumialos and G. Liberopoulos
the output buffer of stage i and contains stage-i finished parts, each of which has
attached to it a kanban from each stage from 1 to i. DAi+1 contains demands for
the production of new stage-(i + 1) parts, each of which has attached to it a stage-
(i + 1) kanban. The synchronization station operates as follows. As soon as there
is one entity in each queue P Ai and DAi+1 , the stage-i finished part engages the
stage-(i + 1) kanban without releasing the kanbans from stages 1 to i that were
already attached to it, and joins the first station of stage i + 1. Note that at stage
1, as soon as a stage-1 kanban is available, a new part is immediately released into
stage 1 since there are always raw parts at the input of the system.
Let us now consider the last synchronization station JN (J3 in the example). JN
synchronizes queues P AN , and DN +1 . P AN represents the output buffer of stage
N and contains stage-N finished parts, each of which has attached to it a kanban
from each stage from 1 to N . DN +1 contains customer demands. When a customer
demand arrives to the system, it joins DN +1 , thereby demanding the release of a
finished part from P AN to the customer. If there is a finished part in queue P AN ,
it is released to the customer and the demand is satisfied. In this case, the finished
part in P AN releases the kanbans that were attached to it, and these kanbans are
transferred upstream to queues DAi (i = 1, . . ., N ). The kanban of stage i carries
along with it a demand for the production of a new stage-i(i = 1, . . ., N ) finished
part and an authorization for the release of a finished part from queue P Ai−1 into
stage i. If there are no finished parts in queue P AN , the customer demand remains
on hold in DN +1 as a backordered demand.
An important special case of the echelon kanban control system in the case
where there are always customer demands for finished parts. This case is known
as the saturated echelon kanban control system. Its importance lies in the fact that
its throughput determines the maximum capacity of the system. In the saturated
system, when there are finished parts at stage N , they are immediately consumed
and an equal number of parts enter the system. As far as the queueing network
corresponding to this model is concerned, the synchronization station JN can be
eliminated since queue DN +1 is never empty and can therefore be ignored. In
the saturated echelon kanban control system, when the processing of a part is
completed at stage N , this part is immediately consumed after releasing the kanbans
of stages 1,. . . , N that were attached to it and sending them back to queues DAi (i =
1, . . ., N ).
It is worth noting that the echelon kanban control system contains the make-
to-stock CONWIP system [23] as a special case. In the make-to-stock CONWIP
system, as soon as a finished part leaves the production system to be delivered to a
customer, a new part enters the system to begin its processing. An echelon kanban
control system with K1 ≤ Ki , i = / 1, behaves exactly like the make-to-stock
CONWIP system.
The dynamic behavior of the echelon kanban control system depends on the
manufacturing processes, the arrival process of customer external demands, and
the number of echelon kanbans of each stage. Among the performance measures
that are of particular interest are the average work in process (WIP) and the average
number of finished parts in each stage, the average number of backordered (not
immediately satisfied) demands, and the average waiting time and percentage of
An analytical method for the performance evaluation 199
backordered demands. In the case of the saturated echelon kanban control system,
the main performance measure of interest is its production rate, Pr , i.e. the average
number of finished parts leaving the output buffer of stage N per unit of time. Pr
represents the maximum rate at which customer demands can be satisfied. With this
in mind, the average arrival rate of external customer demands in the unsaturated
system, say λD , must be strictly less than Pr in order for the system to meet all the
demands in the long run. In other words, the stability condition for the unsaturated
system is
λD < Pr . (2)
concerning parts that arrive from subnetwork LN −1 , and the other at synchroniza-
tion station ON concerning customer demands. Subsystem Ri , i = 2, . . ., N − 1,
is synchronized with only one external arrival process at synchronization station
I i concerning parts that arrive from subnetwork Li−1 . Subsystem R1 is a closed
network; therefore it is not synchronized with any external arrival processes. As
can be seen from Table 3, each synchronization station Ji of the original network
R, linking stage i to stage i + 1, is represented only once in the decomposition.
To completely characterize each subsystem Ri , i = 2, . . ., N , we assume that
each of the external arrival processes to Ri is a state-dependent, continuous-time
Markov process. Let λi (ni ) denote the state-dependent arrival rate of stage-i raw
parts at the upstream synchronization station I i of subsystem Ri , where ni is the
state of subsystem Ri and is defined as the number of parts in this subsystem.
Let Qiu and QiI be the two queues of synchronization station I i , containing niu
and niI customers, respectively, where niu is the number of finished parts of stage
i-1 waiting to enter subnetwork Li , and niI is the number of free stage-i kanbans
waiting to authorize the release of stage-(i − 1 <) finished parts into subnetwork
Li . Then, it is clear that the only possible states of the synchronization station are
An analytical method for the performance evaluation 201
the states (niI , 0), for niI = 0, . . ., Ki , and (0, niu ), for niu = 0, . . ., Ki−1 − Ki ;
therefore, the state ni of subsystem Ri can be simply obtained from niu and niI
using the following relation:
Ki − niI if niI = / 0,
i
n = (3)
Ki + niu if niI = 0.
The above relation implies that 0 ≤ ni ≤ Ki−1 . Also, since the number of raw
parts at the input of stage i cannot be more than the number of stage-(i−1) kanbans,
λi (Ki−1 ) = 0. In subsystem RN , besides the arrival rate of stage-N raw parts at
I N , λN (nN ), there is also the external arrival rate of customer demands at ON ,
λD . Subsystem R1 , as was mentioned above, is a closed network and therefore has
no external arrival processes to define.
To obtain the performance of the original network R, the following two prob-
lems must be addressed: 1) How to analyze each subsystem Ri , i = 1, . . ., N ,
assuming that the external arrival rates are known (except in the case of the first
subsystem R1 , where there are no external arrivals), and 2) how to determine the
unknown external arrival rates. These two problems are addressed in Sections 5
and 6, respectively. Once these two problems have been solved, the performance
of each stage of the original network R can be obtained from the performances of
subsystems Ri , i = 1, . . ., N .
In this section, we describe how to analyze each subsystem in isolation using Marie’s
approximate analysis of general closed queueing networks [20]. Throughout this
analysis, the state-dependent rates of the external arrival processes, λi (ni ), 0 ≤
ni ≤ Ki−1 , i = 2, . . ., N , are assumed to be known. To analyze each subsystem
using Marie’s method, we first view the subsystem as a closed queueing network.
For subsystems Ri , i = 2, . . ., N , this is done by considering the kanbans of stage
i as the customers of the closed network, and the parts and demands (in the case of
the last subsystem RN ) as external resources. Note that the queueing network asso-
ciated with subsystem R1 is already being modeled as a closed queueing network
in the decomposition. Its customers are the kanbans of stage 1.
The closed queueing network associated with subsystem RN is partitioned
into mN + 2 stations, namely, the synchronization stations I N and ON and the
mN stations of subnetwork LN . Similarly, the closed queueing network asso-
ciated with each subsystem Ri is partitioned into mi + 2 stations, namely, the
synchronization station I i , the mi stations of subnetwork Li , and station Ŝi . Fi-
nally, the closed queueing network associated with subsystem R1 is partitioned into
m1 +1 stations, namely, the m1 stations of subnetwork L1 , and station Ŝ1 . Each sta-
tion is approximated by an exponential-service station with load-dependent service
rates. The resulting network associated with each subsystem is a Gordon-Newell,
product-form network [17] consisting of Ki customers and mi + 2 stations for
subsystems Ri , i = 2, . . ., N , and m1 + 1 stations for subsystem R1 . The stations
within each subsystem Ri , i = 1, . . ., N , will be denoted by the index k ∈ Mi ,
202 S. Koukoumialos and G. Liberopoulos
where Vk is the average visit ratio of station k in the original system and is given
from the routing matrix of the original system, and G(K) is the normalization
constant.
To determine the unknown parameters µk (nk ), nk = 1, . . ., K, for each station
k ∈ M , in the product-form solution (4), each station is analyzed in isolation
as an open system with a state-dependent, Poisson arrival process, whose rate
λk (nk ) depends on the total number of customers, nk , present in the station. Let
Tk denote this open system. Assume that the rates λk (nk ) are known for nk =
1, . . ., K − 1. The open queueing system Tk can then be analyzed in isolation
using any appropriate technique to obtain the steady-state probabilities of having
nk customers in the isolated system, say Pk (nk ). The issue of analyzing each
queueing system Tk in isolation will be discussed immediately after Algorithm 1,
below. Once the probabilities Pk (nk ) are known, the conditional throughput of Tk
when its population is nk , which is denoted by vk (nk ), can be derived using the
relation [12],
Pk (nk − 1)
vk (nk ) = λk (nk − 1) , for nk = 1, . . ., K. (5)
Pk (nk )
The load-dependent service rates of the k-th station of the closed product-form
network are then set equal to the conditional throughputs of the corresponding open
station in isolation, i.e.:
µk (nk ) = vk (nk ), for nk = 1, . . ., K. (6)
Once the rates µk (nk ) have been obtained, the state-dependent arrival rates
λk (nk ) can be obtained from the generalized, product-form solution as [6,12]:
Gk (K − nk − 1)
λk (nk ) = Vk , for nk = 1, . . ., K − 1, and λk (K) = 0, (7)
Gk (K − nk )
where Gk (n) is the normalization constant of the closed, product-form network with
station k removed (complementary network) and population n. Gk (n) is a function
of the parameters µk (nk ) for all k =
/ k and nk = 1, . . ., K, and can be computed
efficiently using any computational algorithm for product-form networks [6,9]. An
An analytical method for the performance evaluation 203
iterative procedure can then be used to determine these unknown quantities. This
procedure is described by the following algorithm.
Algorithm 1: Analysis of a Subsystem in Isolation.
Step 0. (Initialization) Set µk (nk ) to some initial value, for k ∈ M and nk =
1, . . ., K.
Step 1. For k ∈ M :
Calculate the state-dependent arrival rates λk (nk ), for nk = 0, . . ., K −1, using
(7).
Step 2. For k ∈ M :
1. Analyze the open queueing system Tk .
2. Derive the steady state probabilities Pk (nk ) of having nk customers, for nk =
1, . . ., K.
3. Calculate the conditional throughputs vk (nk ), for nk = 1, . . ., K, using (5).
Step 3. For k ∈ M :
Set the load-dependent service rates µk (nk ), for nk = 1, . . ., K, in the closed,
product-form network using (6).
Step 4. Go to Step 1 until convergence of the parameters µk (nk ).
Next, we show how to analyze each open queueing system Tk . To do this, we
reintroduce index i to denote subsystem Ri . Step 2.1 of Algorithm 1 above requires
the analysis of the open queueing systems Tki for k ∈ Mi and i = 1, . . ., N .
There are four different types of queueing systems: 1) the synchronization station
ON in subsystem RN , 2) the synchronization stations I i in subsystems Ri , i =
2, . . ., N, 3) the mi stations in each subnetwork Li , i = 1, . . ., N , and 4) the pseudo-
stations Ŝi in subsystems Ri , i = 1, . . ., N − 1.
First, consider the analysis of synchronization station ON in subsystem RN .
N
O is a synchronization station fed by a continuous-time Markov arrival process
with state-dependent rates, λN N N
O (nO ), 0 ≤ nO ≤ KN , and an external Poisson
process with fixed rate λD . An exact solution for this system is easy to obtain by
solving the underlying continuous-time Markov chain. Namely, the steady-state
probabilities PON (nN N
O ) of having nO customers in subsystem O
N
can be derived,
N N
and the conditional throughput vO (nO ) can be estimated using (5) (see [11] and
Appendix A).
The synchronization station I i in each subsystem Ri , i = 2, . . ., N , is a syn-
chronization station fed by two continuous-time Markov arrival processes with
state-dependent rates, λiI (niI ), 0 ≤ niI ≤ Ki , and λi (ni ), 0 ≤ ni ≤ Ki−1 . An
exact solution for this system is also easy to obtain by solving the underlying
continuous-time Markov chain. (see [14] and Appendix B).
The analysis in isolation of any station k ∈ {1, . . ., mi } in each subnetwork
Li , i = 1, . . ., N , reduces to the analysis of a λik (nik )/Gi /1/N queue. Classical
methods can be used to analyze this queue to obtain the steady-state probabilities
Pki (nik ). For instance, if the service time distribution is Coxian, the algorithms
given in [21] may be used. For multiple-server stations, we can use the numerical
204 S. Koukoumialos and G. Liberopoulos
technique presented in [26]. The conditional throughput vki (nik ) can then be derived
from the state probabilities using (5). In the special case where the service time is
exponentially distributed, the conditional throughput vki (nik ) is simply equal to the
load-dependent service rate µik (nik ) [12].
Finally, as was mentioned earlier, pseudo-station Ŝi in subsystem Ri , i =
1, . . ., N − 1, is an aggregate representation of subsystem Ri+1 , which is nested
inside subsystem Ri . Therefore, the conditional throughput of pseudo-station Ŝi ,
vŜi (niŜ ), is set equal to the conditional throughput of subsystem Ri+1 . The condi-
tional throughput of any subsystem Ri , i = 2, . . ., N , is denoted by v i (ni ) and can
be estimated by the following simple expression [3]:
i
λI (Ki − ni ) for 1 ≤ ni ≤ Ki ,
i i
v (n ) = (8)
λiI (0) for Ki ≤ ni ≤ Ki−1 .
Set i = i − 1.
Else
Solve subsystem Ri using Algorithm 1 and calculate the arrival rate λiŜ (niŜ ),
niŜ = 0, . . . , Ki , and the throughput v i (ni ), ni = 1, . . ., Ki−1 , from (8).
If λiŜ (ni+1 ) = λi+1 (ni+1 ), ni+1 = 0, . . . , Ki ,
Set i = i − 1
Else
Set λi+1 (ni+1 ) = λiŜ (ni+1 ), ni+1 = 0, . . . , Ki , and set i = i + 1
Endif
Endif
Endwhile
In the case of the saturated echelon kanban control system, we can use the
same algorithm. The only difference is in the analysis of subsystem RN in Al-
gorithm 1, where there is no downstream synchronization station ON . As far as
the convergence properties of Algorithms 1 and 2 are concerned, in all of the nu-
merical examples that we examined (see Sect. 7), both algorithms converged. The
convergence criterion was that the relative difference between the values of every
unknown parameter at two consecutive iterations should be less that 10−4 .
Once Algorithm 2 has converged, all the performance parameters of the system
can be calculated. Indeed, from the analysis of each subsystem Ri using Algo-
rithm 1, it is possible to derive the performance parameters of stage i in the original
network R, especially the throughput and the average length of each queue, includ-
ing the queues of the synchronization stations. Thus, in the case of the saturated
206 S. Koukoumialos and G. Liberopoulos
echelon kanban system, we can derive the throughput, the average WIP, the average
number of finished parts, and the average number of free echelon kanbans for each
stage. In the case of the echelon kanban control system with external demands,
some other important performance measures can be derived from the analysis of
subsystem RN , namely, the proportion of backordered demands, pB , the average
number of backordered demands, QD , and the average waiting time of backordered
demands, WB . These performance measures can be derived as follows [11,14]:
1 QD
pB = PON (0), QD = PON (0) λN (0) , WB = ,
O
−1 pB λ D
λD
where λN N
O (0) is the arrival rate of finished parts at synchronization station O when
N
there are no finished parts at that station and PO (0) is the steady-state probability
of having no finished parts at synchronization station ON .
7 Numerical results
In this section, we test the approximation method for the performance evaluation
of the echelon kanban control system that we developed in Sections 4–6 on sev-
eral numerical examples. The approximation method was implemented on an Intel
Celeron PC @ 433 MHz, and its results are compared to simulation results obtained
using the simulation software ARENA on an AMD Athlon PC @ 1400 MHz. For
each simulation experiment we run a single replication. The length of this replica-
tion was set equal to the time needed for the system to produce 68 million parts. The
initial condition of the system at the beginning of the replication was set to a typical
regenerative state, namely the state where all customer demands and demands for
the production of new parts at all stages have been satisfied. This permitted us to
set the warm-up period at the beginning of the replication equal to zero. In all simu-
lation experiments we used 95% confidence intervals. The numerical examples are
organized into Sections 7.1 and 7.2. In Section 7.1, we study the accuracy and ra-
pidity of the approximation method as well as the influence of some key parameters
of the echelon kanban control system on system performance. In Section 7.2, we
use the approximation method to optimize the design parameters (echelon kanbans)
of the system.
In this section, we test the accuracy and rapidity of the approximation method
with two numerical examples in which we vary the number of stages, the number
of kanbans in each stage, and the service-time distributions of the manufacturing
process of each stage. For each example, we consider first the case of the saturated
system and then the case of the system with external demands. In each example,
we compare the performance of the system obtained by the approximation method
to that obtained by simulation. We also compare the performance of the echelon
kanban control system obtained by the approximation method and by simulation to
An analytical method for the performance evaluation 207
Table 1. Production capacity of the saturated echelon kanban control system (Example 1)
Simulation Approximation
Table 2. Production capacity of the saturated conventional kanban control system (Exam-
ple 1)
Simulation Approximation
For the system with backordered demands, the main performance parameters
of interest are the proportion of backordered demands, pB , the average number of
backordered demands, QD , and the average waiting time of backordered demands,
WB , as defined at the end of Section 6. Table 3 shows these performance parameters
obtained by the approximation method and by simulation, for the configurations of
parameters 1.3, 1.8, and 1.13 of Table 1, i.e. for K = 5, and different values of the
customer demand rate, λD . The same table also shows the 95% confidence interval
for the simulation results and the number of iterations of Algorithm 2 that are
needed to reach convergence. Table 4 shows the same results for the conventional
kanban control system obtained in [14].
From the results in Table 3, we note that as the customer demand arrival rate
increases, the number of iterations of Algorithm 2 also increases, though not dra-
matically. As far as the average number of backordered demands, QD , is concerned,
we note that the analytical method is fairly accurate. This is not true for the average
waiting time of backordered demands, WB , where in some cases the difference
between the approximation method and simulation are significant. Comparing the
results between Tables 3 and 4, we note that the echelon kanban control system
always has a smaller average number of backordered demands, QD , than the con-
ventional kanban control system, given that the two systems have the same value
of K. The difference in the average number of backordered demands is more pro-
nounced when the two systems are highly loaded, i.e. when λD is close to the
production capacity.
Table 5 shows the results for the average number of finished parts (FP) and the
average work-in-process (WIP) at each stage for the configurations of parameters
1.17 and 1.19 in Table 3. Table 6 shows the same results for the conventional kanban
control system.
Comparing the results between Tables 5 and 6, we note that the echelon kanban
control system has slightly higher average WIP and lower FP than the conventional
kanban control system, when the two systems are highly loaded (i.e. λD is close
to Pr ), and given that the two systems have the same value of K. When the two
systems are not highly loaded, the difference in average WIP and FP between the
two systems is very small. Finally, it appears that the difference in average WIP
and FP between the echelon kanban control system and the conventional kanban
control system is higher in upstream stages than in downstream stages.
Although the above observations hold for the particular configurations of pa-
rameters examined, we expect that they should also hold for the other configurations
of Table 1 and different values of the customer demand rate, λD , because to a large
extent they are due to the fact that the echelon kanban control system always re-
sponds faster to customer demands than the conventional kanban control system,
given that the two systems have the same value of K.
Finally, we should note that the approximation method for the performance
evaluation of the conventional kanban control system developed in [14] is also
based on decomposing a system of N stages into N subsystems. The total number
of the unknown parameter sets (the arrival rates of the external arrival processes
to the subsystems) that must be determined for the conventional kanban control
system, however, is twice as big as that which must be determined for the echelon
210 S. Koukoumialos and G. Liberopoulos
Table 3. Average number of backordered demands, average waiting time of backordered de-
mands, and proportion of backordered demands for the echelon kanban system (Example 1)
1.16: N = 3; K = 5; λD = 0.1
Approximation 0.0 0.0 0.0 6
Simulation 0.0 0.0 0.0
1.17: N = 3; K = 5; λD = 0.5
Approximation 0.035 4.069 1.729 7
Simulation 0.034 (±0.9%) 2.066 (±1.2%) 3.337
1.18: N = 3; K = 5; λD = 0.625
Approximation 0.221 4.594 7.687 7
Simulation 0.213 (±0.1%) 3.014 (±14.2%) 11.32
1.19: N = 3; K = 5; λD = 0.8
Approximation 4.176 10.791 48.38 8
Simulation 4.095 (±3.6%) 9.755 (±7%) 52.47
1.20: N = 5; K = 5; λD = 0.1
Approximation 0.0 0.0 0.0 16
Simulation 0.0 0.0 0.0
1.21: N = 5; K = 5; λD = 0.5
Approximation 0.035 4.070 1.71 16
Simulation 0.032 (±0.007%) 3.189 (±0.003%) 2.03
1.22: N = 5; K = 5; λD = 0.8
Approximation 6.774 14.440 58.69 22
Simulation 6.5686 (±0.08%) 12.895 (±0.02%) 63.67
1.23: N = 10; K = 5; λD = 0.1
Approximation 0.0 0.0 0.0 20
Simulation 0.0 0.0 0.0
1.24: N = 10; K = 5; λD = 0.5
Approximation 0.035 4.070 1.72 39
Simulation 0.023 (±0.005%) 3.512 (±0.002%) 1.28
1.25: N = 10; K = 5; λD = 0.77
Approximation 3.817 10.709 46.3 61
Simulation 3.131 (±0.003%) 9.064 (±0.001%) 49.3
kanban control system (namely, there are 2(N − 1) external arrival rates for the
conventional kanban control system compared to N − 1 external arrival rates for
the echelon kanban control system). Yet, for both examples examined, the number
of iterations needed for the convergence of the parameters is significantly lower for
the conventional kanban control system than for the echelon kanban control system,
given the same convergence criterion for the two systems, as can be seen from Tables
1–4. This is due to the fact that the coordination of production is decentralized in
An analytical method for the performance evaluation 211
1.16: N = 3; K = 5; λD = 0.1
Approximation 0.0 0.0 0.0 1
Simulation 0.0 0.0 0.0
1.17: N = 3; K = 5; λD = 0.5
Approximation 0.035 2.06 3.4 2
Simulation 0.033 (±30%) 2.16 (±17%) 3.1
1.18: N = 3; K = 5;
λD = 0.625 Approximation 0.222 3.00 11.82 3
Simulation 0.230 (±17%) 3.26 (±15%) 11.78
1.19: N = 3; K = 5; λD = 0.8
Approximation 4.56 10.1 56.3 4
Simulation 4.26 (±19%) 10.3 (±13%) 52.1
1.20: N = 5; K = 5; λD = 0.1
Approximation 0.0 0.0 0.0 1
Simulation 0.0 0.0 0.0
1.21: N = 5; K = 5; λD = 0.5
Approximation 0.0353 2.07 3.40 2
Simulation 0.038 (±30%) 2.16 (±9%) 3.58
1.22: N = 5; K = 5; λD = 0.8
Approximation 11.26 19.3 73.0 7
Simulation 8.93 (±22%) 17.2 (±15%) 65.2
1.23: N = 10; K = 5; λD = 0.1
Approximation 0.0 0.0 0.0 1
Simulation 0.0 0.0 0.0
1.24: N = 10; K = 5; λD = 0.5
Approximation 0.0353 2.07 3.40 2
Simulation 0.0368 (±30%) 2.18 (±17%) 3.38
1.25: N = 10; K = 5; λD = 0.77
Approximation 6.89 13.9 64.2 11
Simulation 5.95 (±22%) 13.7 (±14%) 56.9
Table 5. Average work in progress (WIP) and average number of finished parts (FP) in each
stage for the echelon kanban control system (Example 1)
1.17: N = 3; K = 5; λD = 0.5
Simulation 0.988 4.039 0.978 4.022 0.961 4.011
(±0.1%) (±0.09%) (±0.1%) (±0.1%) (±0.1%) (±0.1%)
Approximation 0.999 4.031 0.995 4.005 0.969 4.000
Error +1.1% −0.2% +1.7% −0.4% +0.8% −0.3%
1.19: N = 3; K = 5; λD = 0.8
Simulation 3.363 2.392 3.068 2.018 2.589 1.569
(±0.5%) (±0.3%) (±0.3%) (±0.3%) (±0.3%) (±0.5%)
Approximation 3.479 2.349 3.159 1.902 2.655 1.455
Error +3.3% −1.8% +2.9% −6.1% +2.5% −7.8%
Table 6. Average work in progress (WIP) and average number of finished products (FP) in
each stage for the conventional kanban control system (Example 1)
1.17: N = 3; K = 5; λD = 0.5
Simulation 0.94 4.06 0.95 4.02 0.94 4.04
(±3.2%) (±0.7%) (±3.1%) (±0.7%) (±3.2%) (±0.8%)
Approximation 0.97 4.03 0.97 4.01 0.97 4.00
Error +3% −0.7% +2% −0.2% +3% −1%
1.19: N = 3; K = 5; λD = 0.8
Simulation 2.54 2.47 2.52 1.98 2.55 1.58
(±3.0%) (±4.0%) (±3.2%) (±5.0%) (±3.1%) (±6.3%)
Approximation 2.61 2.38 2.58 1.85 2.66 1.40
Error +2.7% −3.6% +2.4% −6.5% +4% −11%
Simulation Approximation
Simulation Approximation
From the results in Table 7, we note that when the variability of the service
time distribution increases, the production capacity decreases, as is expected. The
results obtained by the approximation method are fairly accurate when compared to
the simulation results. Comparing the results between Tables 7 and 8, we note that
for all the service-time distributions, the production capacity of the echelon kanban
control system is higher than that of the conventional kanban control system. The
results for the analytical solution and simulation for the case of the echelon kanban
system with backordered demands is shown in Figure 4. More specifically, Figure 4
depicts the proportion of backordered demands, pB , as a function of the arrival rate
of demands, λD , for the three different service time distributions. It appears that as
the cv2 of the service time distribution increases, the difference between simulation
and analytical results tends to increase.
Fig. 4. Proportion of backordered demands versus the average arrival rate of demands for
different values of the service-time squared coefficient of variation (Example 2)
time and, therefore, building up inventory (inventory holding cost). In this paper, we
consider an optimization problem where the objective is to meet a certain quality
of service constraint with minimum inventory holding cost.
We examine two quality-of-service measures as in [15]. The first measure is
the probability that when a customer demand arrives, it is backordered. The second
measure is the probability that when a customer demand arrives, it sees more than
nwaiting demands, excluding itself. The first measure is denoted by Prupt and con-
cerns the situation where the demands must be immediately satisfied. The second
measure is denoted by P (Q > n) and concerns the situation where we have the
prerogative to introduce a delay in filling orders, which is equivalent to authorizing
demands to wait. Specifically, Prupt is the marginal stationary probability of having
no finished parts in the last synchronization station, which is given by equation (18)
in Appendix A. Similarly, P (Q > n) is the stationary probability of having more
than n customers waiting and can be computed from the following expression:
∞
n
P (Q > n) = P (Q = x) = 1 − P (Q = y), (10)
x=n+1 y=0
The cost function that we want to minimize is the long-run, expected, average cost
of holding inventory,
N
Ctotal = hi E [W IPi + F Pi ], (13)
i=1
where hi is the unit cost of holding W IPi + FPi inventory per unit time in stage i.
In the remaining of this section, we optimize the echelon kanbans of an echelon
kanban control system consisting of N = 5 stages, where each stage contains a
single machine with exponentially distributed service times with mean equal to
1, for different combinations of inventory holding cost rates, hi , i = 1, . . ., 5,
and demand arrival rate λD = 0.5. In all cases we assume that there is value
added to the parts at every stage so that the inventory holding cost increases as
the stage increases, i.e. h1 < h2 < . . . < h5 . If this were not the case, i.e. if
h1 = h2 = . . . = h5 , then clearly it would make no sense to block the passage of
parts from one stage to another via the use of echelon kanbans, because this would
not lower the inventory holding cost but would worsen the quality of service. This
implies that if h1 = h2 = . . . = h5 , the optimal echelon kanbans satisfy K1 ≤ Ki ,
i = 2, . . ., 5, in which case the echelon kanban control system is equivalent to the
make-to-stock CONWIP system [23] with a WIP-cap on the total number of parts
in the system equal to K1 .
Table 9 shows the optimal design parameters (K1 , . . ., K5 ) and associated min-
imum, long-run, expected, average cost of holding inventory, for λD = 0.5 and
different quality of service constraints and inventory holding cost rates h1 , . . ., h5 ,
where h1 < h2 < . . . < h5 . The quality of service constraints that we use are
Prupt ≤ 0.02 and P (Q > n) ≤ 0.02, for n = 2, 5, 10.
From the results in Table 9, we see that the higher the number of backordered
demands n in the quality of service definition, P (Q > n), the lower the optimal
number of echelon kanbans, and hence the inventory holding cost. As the difference
between the holding cost rates hi , i = 1, . . ., 5, increases, the difference between the
optimal values of Ki , i = 1, . . ., 5, also increases, since the behavior of the echelon
kanban control system diverts further from that of the make-to-stock CONWIP
system. When the relative difference between the holding cost rates hi , i = 1, . . ., 5,
is low, the behavior of the echelon kanban control system tends to that of the make-
to-stock CONWIP system.
Table 10 shows the optimal design parameter K1 and associated minimum
inventory holding cost for λD = 0.5 and different quality of service constraints and
inventory holding cost rates h1 , . . ., h5 , for the make-to-stock CONWIP system.
The last column of Table 10 shows the relative increase in cost of the optimal
make-to-stock CONWIP system compared to the optimal echelon kanban control
system. Comparing the results between Tables 9 and 10, we note that the optimal
make-to-stock CONWIP system performs considerably worse than the optimal
echelon kanban control system, particularly when the relative difference between
the holding cost rates hi , i = 1, . . ., 5, is high and/or the number of backordered
demands n in the quality of service definition, P (Q > n), is high, indicating that
the quality of service is low.
216 S. Koukoumialos and G. Liberopoulos
Table 9. Opimal configuration and associated costss for λD = 0.5 and different values of
h1 , . . . , h5 , for the echelon kanban control system
Table 10. Optimal configuration and associated costs for λD = 0.5 and different values of
h1 , . . . , h5 , for the CONWIP system
h1 = 1, h2 = 6, h3 = 11, h4 = 16, h5 = 21
Prupt ≤ 0.02 14 244.163 10.43%
P (Q > 2) ≤ 0.02 12 202.415 11.98%
P (Q > 5) ≤ 0.02 10 161.006 28.2%
P (Q > 10) ≤ 0.02 8 120.307 36.39%
h1 = 1, h2 = 11, h3 = 21, h4 = 31, h5 = 41
Prupt ≤ 0.02 14 474.326 11.37%
P (Q > 2) ≤ 0.02 12 392.830 13.11%
P (Q > 5) ≤ 0.02 10 312.012 29.1%
P (Q > 10) ≤ 0.02 8 232.613 37.64%
h1 = 1, h2 = 2, h3 = 4, h4 = 8, h5 = 16
Prupt ≤ 0.02 14 175.160 17.86%
P (Q > 2) ≤ 0.02 12 143.407 21.59%
P (Q > 5) ≤ 0.02 10 111.986 41.2%
P (Q > 10) ≤ 0.02 8 81.260 50.86%
h1 = 1, h2 = 3, h3 = 9, h4 = 27, h5 = 81
Prupt ≤ 0.02 14 850.927 25.59%
P (Q > 2) ≤ 0.02 12 690.358 31.65%
P (Q > 5) ≤ 0.02 10 531.715 56.47%
P (Q > 10) ≤ 0.02 8 377.102 63.12%
8 Conclusions
pN N N N N N
O (nO , 0)λD = pO (nO − 1, 0)λO (nO − 1) for nN
O = 1, ..., KN (14)
pN N
O (0, nD )λO (0) = pN
O (0, nD − 1)λD for nD > 0 (15)
The marginal probabilities PON (nN
O ) are then simply given by
PON (nN N N
O ) = pO (nO , 0) for nN
O = 1, ..., KN , (16)
∞
PON (0) = pN
O (0, nD ). (17)
nD =0
N
The conditional throughputs of subsystem O are obtained from (5), (14) and (16),
as follows:
N
vO (nN
O ) = λD for nN
O = 2, ..., KN (19)
From (5), (14), (16) and (18), we also get
N 1
vO (1) = . (20)
1
λD − λN1(0)
O
An analytical method for the performance evaluation 219
λi (Ki + n − 1)
n=1
piI (0, niu ) = i niu piI (0, 0). (22)
λI (0)
The marginal probabilities, PIi (niI ), can then be derived by summing up the prob-
abilities above as follows:
⎡ i ⎤
nI
- i
λI (n − 1) ⎦ i
PIi (niI ) = ⎣ i (K − n)
pI (0, 0) for niI = 1, . . . , Ki , (23)
n=1
λ i
⎡ n i ⎤
,u i
⎢ Ki−1 −Ki
λ (Ki + n − 1) ⎥
⎢ n=1 ⎥ i
PIi (0) = ⎢1 + i niu ⎥ pI (0, 0). (24)
⎣ λI (0) ⎦
i nu =1
Fig. 6. Continuous-time Markov chain describing the state (niI ,niu ) of synchronization sta-
tion I i
220 S. Koukoumialos and G. Liberopoulos
⎡ n i ⎤
,u
⎢ Ki−1 −Ki
λi (Ki + n − 1) ⎥
⎢ n=1 ⎥
vIi (1) = λi (Ki − 1) ⎢1 + i niu ⎥. (26)
⎣ λ (0) ⎦
niu =1 I
N Number of stages
Ki Number of echelon kanbans of stage i
Li Subnetwork associated with the manufacturing process of stage i
mi Number of stations of subnetwork Li
Ji Synchronization station at the output of stage i
λD Average arrival rate of external customer demands in the unsaturated
system
Pr Maximum rate at which customer demands can be satisfied
R Queueing network of the echelon kanban control system
Ri Subsystem associated with stage i
Ii Upstream synchronization station of subsystem Ri
ON Downstream synchronization station of subsystem RN
Ŝi Downstream single-server pseudo-station of subsystem Ri
ni State of subsystem Ri
λi (ni ) State-dependent arrival rate of stage-i raw parts at the upstream synchro-
nization station I i of subsystem Ri
v i (ni ) Conditional throughput of subsystem Ri
k ∈ Mi Index denoting the stations within subsystem Ri , where M1 =
{1, . . ., m1 , Ŝ}, Mi = {I, 1, . . ., mi , Ŝ} for i = 2, . . ., N − 1, and
MN = {I, 1, . . ., mN , O}
nik State of station k in subsystem Ri
µik (nik ) Load-dependent service rate of station k in subsystem Ri
µk (nk ) Same as µik (nik ) with index i dropped
Tki Open system representing station k in subsystem Ri
Tk Same as Tki with index i dropped
λik (nik ) Rate of state-dependent Poisson arrival process at Tki
λk (nk ) Same as λik (nik ) with index i dropped
vki (nik ) Conditional throughput of Tki
vk (nk ) Same as vki (nik ) with index i dropped
Pki (nik ) Steady-state probability of Tki
pB Proportion of backordered demands
QD Average number of backordered demands
WB Average waiting time of backordered demands
An analytical method for the performance evaluation 221
References
1. Baskett F, Chandy KM, Muntz RR, Palacios-Gomez F (1975) Open, closed and mixed
networks of queues with different classes of customers. Journal of ACM 22: 248–260
2. Baynat B, Dallery Y (1993) A unified view of product-form approximation techniques
for general closed queueing networks. Performance Evaluation 18(3): 205–224
3. Baynat B, Dallery Y (1993) Approximate techniques for general closed queueing net-
works with subnetworks having population constraints. European Journal of Opera-
tional Research 69: 250–264
4. Baynat B, Dallery Y (1996) A product-form approximation method for general closed
queueing networks with several classes of customers. Performance Evaluation 24: 165–
188
5. Baynat B, Dallery Y, Ross K (1994) A decomposition approximation method for mul-
ticlass BCMP queueing networks with multiple-server stations. Annals of Operations
Research 48: 273–294
6. Bruell SC, Balbo G (1980) Computational algorithms for closed queueing networks.
Elsevier North-Holland, Amsterdam
7. Buzacott JA (1989) Queueing models of kanban and MRP controlled production sys-
tems. Engineering Costs and Production Economics 17: 3–20
8. Buzacott JA, Shanthikumar JG (1993) Stochastic models of manufacturing systems.
Prentice-Hall, Englewood Cliffs, NJ
9. Buzen JP (1973) Computational algorithms for closed queueing networks with expo-
nential servers. Comm. ACM 16(9): 527–531
10. Clark A, Scarf H (1960) Optimal policies for a multi-echelon inventory problem. Man-
agement Science 6: 475–490
11. Dallery Y (1990) Approximate analysis of general open queueing networks with re-
stricted capacity. Performance Evaluation 11(3): 209–222
12. Dallery Y, Cao X (1992) Operational analysis of stochastic closed queueing networks.
Performance Evaluation 14(1): 43–61
13. Dallery Y, Liberopoulos G (2000) Extended kanban control system: combining kanban
and base stock. IIE Transactions 32(4): 369–386
14. Di Mascolo M, Frein Y, Dallery Y (1996) An analytical method for performance eval-
uation of kanban controlled production systems. Operations Research 44(1): 50–64
15. Duri C, Frein Y, Di Mascolo M (2000) Comparison among three pull control policies:
kanban, base stock and generalized kanban. Annals of Operations Research 93: 41–69
16. Frein Y, Di Mascolo M, Dallery Y (1995) On the design of generalized kanban control
systems. International Journal of Operations and Production Management 15(9): 158–
184
17. Gordon WJ, Newell GF (1967) Closed queueing networks with exponential servers.
Operations Research 15: 252–267
18. Jackson JR (1963) Jobshop-like queueing systems. Management Science 10(1): 131–
142
19. Liberopoulos G, Dallery Y (2002) Comparative modeling of multi-stage production-
inventory control policies with lot sizing. International Journal of Production Research
41(6): 1273–1298
20. Marie R (1979) An approximate analytical method for general queueing networks.
IEEE Transactions on Software Engineering 5(5): 530–538
21. Marie R (1980) Calculating equilibrium probabilities for λ(n)/Ck /1/N queues. Perfor-
mance Evaluation Review 9: 117–125
22. Reiser M, Lavenberg SS (1980) Mean value analysis of closed multichain queueing
networks. Journal of ACM 27(2): 313–322
222 S. Koukoumialos and G. Liberopoulos
Abstract. In this paper we consider closed loop two-echelon repairable item sys-
tems with repair facilities both at a number of local service centers (called bases)
and at a central location (the depot). The goal of the system is to maintain a number
of production facilities (one at each base) in optimal operational condition. Each
production facility consists of a number of identical machines which may fail inci-
dentally. Each repair facility may be considered to be a multi-server station, while
any transport from the depot to the bases is modeled as an ample server. At all
bases as well as at the depot, ready-for-use spare parts (machines) are kept in stock.
Once a machine in the production cell of a certain base fails, it is replaced by a
ready-for-use machine from that base’s stock, if available. The failed machine is
either repaired at the base or repaired at the central repair facility. In the case of local
repair, the machine is added to the local spare parts stock as a ready-for-use ma-
chine after repair. If a repair at the depot is needed, the base orders a machine from
the central spare parts stock to replenish its local stock, while the failed machine
is added to the central stock after repair. Orders are satisfied on a first-come-first-
served basis while any requirement that cannot be satisfied immediately either at
the bases or at the depot is backlogged. In case of a backlog at a certain base, that
base’s production cell performs worse.
To determine the steady state probabilities of the system, we develop a slightly
aggregated system model and propose a special near-product-form solution that
provides excellent approximations of relevant performance measures. The depot
repair shop is modeled as a server with state-dependent service rates, of which the
parameters follow from an application of Norton’s theorem for Closed Queuing
Networks. A special adaptation to a general Multi-Class Marginal Distribution
Analysis (MDA) algorithm is proposed, on which the approximations are based.
All relevant performance measures can be calculated with errors which are generally
less than one percent, when compared to simulation results. The approximations
are used to find the stock levels which maximize the availibility given a fixed
configuration of machines and servers and a certain budget for storing items.
1 Introduction
Repairable inventory theory involves designing inventory systems for items which
are repaired and returned to use rather than discarded. The items are less expensive
to repair than to replace. Such items can for example be found in the military, avi-
ation, copying machines, transportation equipment and electronics. The repairable
inventory problem is typically concerned with the optimal stocking of parts at bases
and a central depot facility which repairs failed units returned from bases while pro-
viding some predetermined level of service. Different performance measures may
be used, such as cost, backorders and availability.
Over the past 30 years there has been considerable interest in multi-echelon
inventory theory. Much of this work originates from a model called METRIC, which
was first reported in the literature by Sherbrooke [9]. The model was developed for
the US Air Force at the Rand Corporation for a multi-echelon repairable-item
inventory system. In this model an item at failure is replaced by a spare if one is
available. If none are available a spare is backordered. Of the failed items a certain
proportion is repaired at the base and the rest at a repair depot, thereby creating
a two-echelon repairable-item system. Items are returned from the depot using a
one-for-one reordering policy. The METRIC model determines the optimal level
of spares to be maintained at each of the bases and at the depot.
A shortfall of the METRIC model is that it assumes that failures are Poisson
from an infinite source and that the repair capacity is unlimited. Therefore, others
have continued the research to gain results more useful for real life applications.
Gross, Kioussis and Miller [5], Albright and Soni [1] and Albright [2] focused their
attention on closed queuing network models, thereby dropping the assumption of
Poisson failures from an infinite source. The intensity by which machines enter the
repair shops depends on the number of machines operating in the production cell.
In case of a backlog at a base, this intensity is therefore smaller than in the optimal
case where the maximum number of machines is operating in the production cell.
Also the assumption of unlimited repair capacity is dropped in Gross et al. [5] and
Albright [2].
This paper deals with similar models. It handles closed queuing network mod-
els with limited repair. However, the approximation method differs considerably.
The approximation method builds on the method by Avsar and Zijm [3]. Avsar and
Zijm considered an open queuing network model with limited repair. By a small
aggregation step, the system is changed into a system with a special near-product-
form solution that provides an approximation for the steady state distribution. From
the steady state distribution all relevant performance measures can be computed.
Closed loop two-echelon repairable item systems 225
We will perform a similar aggregation step in this paper and again a special near-
product-form solution will be obtained. However, as opposed to open systems, in a
system with finite sources, the demand rates to the depot also become state depen-
dent; moreover, these demand rates are clearly influenced by the efficiency of the
base repair stations. Nevertheless, we are able to develop relatively simple approxi-
mation algorithms to obtain the relevant performance measures. These performance
measures can ultimately be used within an optimization model to determine such
quantities as the optimal repair capacities and the optimal inventory levels.
The organization of this paper is as follows: In the next section we consider
a very simple two-echelon system, consisting of one base, a base repair shop and
a central repair shop. The repair shops are modeled as single servers. This model
mainly serves to explain the essential elements of the aggregation step. We present
the modified system with near-product-form solution and numerical results to show
the accuracy of the approximation. Next, in Section 3, we turn to more general re-
pairable item network structures, containing multiple bases and transport lines from
the depot to the bases. The repair shops are modeled as multi-servers. The approx-
imation method leading to an adapted Multi-Class MDA algorithm is presented
and some numerical results are discussed. In Section 4, an optimization algorithm
based on this approximation method, is given which finds the stock levels that max-
imize the (weighted) availibility under a given cost constraint. In the last section,
we summarize our results and discuss a number of extensions that are currently
being investigated.
Consider the system as depicted in Figure 1. The system consists of a single base
and a depot. At the base a maximum of J1 machines can be operational in the
production cell.
Operational machines fail at exponential rate λ1 and are replaced by a machine
from the base stock (if available). Both at the base and at the depot there is a repair
shop. Failed machines are base-repairable with probability p1 and consequently
depot-repairable with probability 1 − p1 . The repair shops are modeled as single
servers with exponential service rate µ0 for the depot and exponential service rate
µ1 for the base. In addition to the J1 machines another group of S1 machines is
dedicated to the base to act as spares. When a machine fails, the failed machine goes
to a repair shop while at the same time a request is sent to place a spare machine
from the base stock in the production cell. This request is carried out immediately,
if possible. In case no spare machines are at the base, a backlog occurs. As soon
as there is a repaired machine available, it becomes operational. A number of S0
226 L. Spanjers et al.
Production cell
1 − p1
λ1
λ1
p1
Base repair
µ1 λ1
J1 machines
S1
Depot repair
µ0
S0
machines is dedicated to the depot to act as spares. When a failed machine cannot
be repaired at the base and hence is sent to the depot, a spare machine is shipped
from the depot to the base to replenish the base stock, or - in case of a backlog
- to become operational immediately. When no spares are available at the depot,
a backorder is created. In that case, as soon as a machine is repaired at the depot
repair shop, it is sent to the base. In this simple model, transport times from the
base to the depot and vice versa are not taken into account.
In Figure 1 (and subsequent figures), requests are indicated by dotted lines. The
matching of a request and a ready-for-use machine is modeled as a synchronization
queue, both at the base and at the depot. At the base however, some reflection reveals
that the synchronization queue can be seen as a normal queue where machines
are waiting to be moved into the production cell. This is only possible when the
production cell does not contain the maximum number of machines, that is, if a
machine in the production cell has failed. This leads to the model in Figure 2.
Production cell
1 − p1
λ1
λ1
p1
Base repair
µ1 λ1
k m11
j1 machines
m12 operational
Depot repair
µ0
n1 n2
In this figure the variables n1 , n2 , k, m11 and m12 indicate the lengths of the
various queues in the system. The number of machines in (or awaiting) depot repair
is denoted by the random variable n1 , the number of spare machines at the depot
is denoted by the random variable n2 and the backlog of machines at the depot
is denoted by k. At the base there are m11 machines waiting for repair or being
repaired and m12 machines are acting as spares. In the production cell j 1 machines
are operational. As a result of the operating inventory control policies, for n1 = n1 ,
n2 = n2 , k = k, m11 = m11 , m12 = m12 and j 1 = j1 the following equations
Closed loop two-echelon repairable item systems 227
must hold:
n 1 + n 2 − k = S0 , (1)
n2 · k = 0, (2)
k + m11 + m12 + j1 = S1 + J1 , (3)
m12 · (J1 − j1 ) = 0, (4)
where Equations (2) and (4) follow from the fact that it is impossible to have a
backlog and to have spare machines available at the same time. If spare machines
are available, a request is satisfied immediately. In case of a backlog, a request is
not satisfied until a repair completion. The repaired machine is merged with the
longest waiting request.
From these relations it follows immediately that n1 and m11 completely deter-
mine the state of the system, including the values of n2 , k, m12 and j1 . Therefore,
the system can be modeled as a continuous time Markov chain with state description
(n1 , m11 ). The corresponding transition diagram is displayed in Figure 3.
m11
S1 + J1
( J1 + S − m ) p λ
1 11 1 1
µ0 ( J1 + S − m )(1 − p ) λ
1 11 1 1
µ1
( J1 + S + S − m − n ) p λ
II 0 1 11 1 1 1
S1
µ0 ( J1 + S + S − m − n )(1 − p ) λ
0 1 11 1 1 1
µ1
J1 p λ J1 p λ
1 1 1 1
µ0 µ0
J1 (1 − p )λ J1 (1 − p )λ
1 1 1 1
µ1 µ1
I III IV
S0 S 0 + S1 S0 + S1 + J1 n1
Let P (n1 , m11 ) = P (n1 = n1 , m11 = m11 ) be the steady state probability of
being in state (n1 , m11 ). This steady state probability can be found by solving the
global balance equations of the system. These can be deduced from the transition
diagram. Nevertheless, it is not possible to find an algebraic expression for the
steady state probabilities. Moreover, for larger systems with e.g. multiple bases,
the computational effort becomes prohibitive. Therefore the system will be slightly
adjusted in the next subsection, in order to arrive at a near-product form network.
Note that the analysis presented in this paper, is partly similar to the one given in
Avsar and Zijm [3], where the equivalent open two-echelon network is considered.
For this open network, an algebraic and easily computable product form approxi-
mation is found. In the current paper, a closed network is considered, and an easily
computable algebraic approximation could not be found. However, the aggregated
228 L. Spanjers et al.
network has a product form steady state distribution, and we can use MDA-like
algorithms to find numerical approximations for performance measures.
An alternative approach is to model the number of machines at the depot and
the bases as a level dependent quasi birth death process. This method may yield an
algebraic solution but, here too, the finite state space makes the analysis more com-
plex. Moreover, the transition rates in a given state, do not only depend on the phase
but also on the level. Together, this makes the alternative method computationally
very demanding, if not intractable.
2.2 Approximation
A first step towards an approximation for the steady state probabilities is to aggregate
the state space. The most difficult parts of the transition diagram are regions I and
II, that is, the parts with n1 ≤ S0 or, equivalently, the parts with k = 0. The parts
with k > 0 are equivalent to the states with n1 = k + S0 . A natural aggregation of
the system is a description through the states (k, m11 ). The states (n1 , m11 ) with
n1 = 0, 1, . . . , S0 are then aggregated into one state (0, m11 ). Denote the steady
state probabilities for the new model by P̃ then the following holds for any m11 :
S0
P̃ (k = 0, m11 = m11 ) = P (n1 = n1 , m11 = m11 ), (5)
n1 =0
However, to compute this, one needs to know the steady state distribution of the
original system, which is exactly what we attempt to approximate. Therefore, we
approximate the q(m11 )’s by their weighted average, i.e. we focus on the conditional
probability q defined by
q= q(m11 )P (m11 = m11 |n1 ≤ S0 ) = P (n1 = S0 |n1 ≤ S0 ) (8)
m11
and for every m11 we replace q(m11 ) in the transition diagram by this q. In the
next section it will be explained how a reasonable approximation for this q can be
obtained by means of an application of Norton’s theorem.
Closed loop two-echelon repairable item systems 229
m11
S1 + J1
( J1 + S − m ) p λ
1 11 1 1
( J1 + S1 − m11)(1 − p1)λ1q(m11)
µ1
( J1 + S − m − k ) p λ
1 11 1 1
S1
J1 p λ µ0 ( J1 + S − m − k )(1 − p ) λ
1 1 1 11 1 1
µ1
J1(1 − p1)λ1q(m11)
µ1
J1 p λ
1 1
µ0
J1 (1 − p ) λ
1 1
µ1
k
0 S1 S1 + J1
Fig. 4. Transition diagram for state description (k, m11 )
Lemma 1 The steady state probabilities for the model with state description
(k, m11 ) and transition rates as denoted in Figure 4 with q(m11 ) replaced by
arbitrary q have a product form.
Proof. To find the steady state probabilities, consider both the original model in
Figure 2 and the alternative model in Figure 5.
Production cell
1 − p1
λ1
λ1
p1
Base repair
µ1 λ1
m11
Depot repair j1 machines
µ0 / ∞ m12 operational
k
Expressed in terms of the state variables (k, m11 ), this result immediately leads
to:
Lemma 2 The steady state distribution for the aggregate model is given by
⎧ m11 k
⎪
⎪ Gq p1 λ 1 (1−p1 )λ1
⎪
⎪ ,
⎪
⎪ J1 !J1S1 −k−m11 µ1 µ0
⎪
⎪
⎪
⎪ k+m11 ≤S1 , k>0
⎪
⎪
⎪
⎪ m11 k
⎪
⎪
⎪
⎪ Gq p1 λ 1 (1−p1 )λ1
⎪
⎪ ,
⎪
⎪ (S1 +J1 −k−m11 )! µ1 µ0
⎪
⎨ k+m11 >S1 , k>0
P̃ (k, m11 )= m11 (10)
⎪
⎪
⎪
⎪ G p1 λ 1
⎪
⎪ ,
⎪
⎪ J1 !J1S1 −m11 µ1
⎪
⎪
⎪
⎪ m11 ≤ S1 , k = 0
⎪
⎪
⎪
⎪ m11
⎪
⎪ G p1 λ 1
⎪
⎪ ,
⎪ (S1 + J1 − m11 )!
⎪ µ1
⎪
⎩
m11 > S1 , k = 0
−(J1 +S1 )
with G = G̃λ1 the normalization constant.
The previous lemma gives an explicit expression for the steady state probabili-
ties. For large systems it may be difficult to calculate the normalization constant G.
However, since we are dealing with a product form network, Marginal Distribution
Analysis (see e.g. Buzacott and Shanthikumar [4]) can be used to calculate the
appropriate performance measures directly.
The results presented so far hold true for any value of q ∈ [0, 1]. In the derivation
of the lemmas above the interpretation of q as the conditional probability that a
Closed loop two-echelon repairable item systems 231
request at the depot has to wait given that it finds no other requests in front of it
(see (8)), has not been used. Therefore any q ∈ [0, 1] will do, but it is expected that
a good approximation will be obtained by using a q that does correspond to this
interpretation. In the next subsection Norton’s theorem will be used to find a q with
a meaningful interpretation that gives good results.
Although we have stated in the previous section that the product form does not
depend on q, it is still needed to find a q that gives a good approximation for the
performance measures. In this section, the basic idea of Norton’s theorem (see
Harrison and Patel [6] for an overview) is used to find an approximation for q that
gives good results. This basic idea is that a product form network can be analyzed
by replacing subnetworks by state dependent servers. Norton’s theorem states that
the joint distributions for the numbers of customers in the subnetworks and the
queue lengths at the replacing state dependent servers are the same.
To use this idea, first recall the original model as shown in Figure 2. We want
to find q, the conditional probability that a request corresponding with a machine
failure finds no spare parts in stock at the depot, although there was no backlog so
far. The base, consisting of the production cell and the base repair shop, is taken
apart and replaced by a state dependent server.
Production cell
1 − p1
λ1
λ1
p1
Base repair
µ1 λ1
k m11
j1 machines
TH1(i) m12 operational
Depot repair TH1(i)
µ0
n1 n2
a b
Fig. 6. a The new network with state dependent server. b The short circuited network
The new network with the state dependent server is displayed in Figure 6a. In
order to find the service rates for this state dependent server, the original network is
short circuited by setting the service rate at the depot repair facility to infinity. This
short circuited network is also depicted in Figure 6b. The service rate for the new
state dependent server with i jobs present is equal to the throughput of the short
circuited network with i jobs present, denoted by T H1 (i).
The evolution of n1 = n1 , the number of machines in or awaiting depot re-
pair, can be described as a birth-death process. The transition diagram is shown in
Figure 7.
232 L. Spanjers et al.
0 1 2 S0 − 1 S0 S0 + 1 S0 + S1 + J1 − 1 S0 + S1 + J1
µ0 µ0 µ0 µ0 µ0
Note that this is just an approximation due to the fact that Norton’s theorem is
only valid for product form networks. In case S0 = 0, we would have a product
form network and the results would be exact. From the diagram one can observe
that
2.4 Results
cell, denoted by A, and the expected number of machines operating in the production
cell (Ej 1 ). These are defined as follows:
A = P (j 1 = J1 ) = P (b1 ≥ J1 ) = P (k + m11 ≤ S1 ), (14)
Ej 1 = E(J1 −[k+m11 −S1 ]+ ) = (J1 −[k+m11 − S1 ]+ )P (k, m11 ). (15)
k,m11
In this section the simple system from Section 2 will be extended to a more realistic
one. The system will contain multiple bases and transport lines. Furthermore, the
single servers that are used in the repair shops are replaced by multiple parallel
servers. These adjustments will make the analysis of the system more complicated.
Nevertheless, the basic idea of the aggregation step will be the same.
The system in this section consists of multiple bases, where the number of bases is
denoted by L. A graphical representation of the system is given in Figure 8 for the
case L = 2.
As in the simple system described before, at base = 1, . . . , L at most J
machines are operating in the production cell. The machines fail at exponential rate
λ and are always replaced by a machine from the corresponding base stock (if
available). Failed machines from base are base-repairable with probability p and
depot-repairable with probability 1 − p . In contrast to the simple model described
before, the repair shops are modeled as multi-servers. That is, at the repair shop of
base = 1, . . . , L R repairmen are working, each at exponential rate µ . At the
depot repair shop R0 repairmen are working at exponential rate µ0 . Consistent with
the simple model S machines are dedicated to base to act as spares and S0 spare
machines are dedicated to the depot. Broken machines at a certain base that are
base-repairable are sent to the base repair shop. After repair they fill up the spares
buffer at base or, in case of a backlog at that base, become operational immediately.
Broken machines from base that are considered depot-repairable are sent to the
depot repair shop. When depot spares are available, a spare is immediately sent to
the stock of base . In case there are no spares available a backlog occurs. Machines
that have completed repair are sent to the base that has been waiting the longest. That
is, an FCFS return policy is used. In this model the transportation from the depot to
234 L. Spanjers et al.
Table 1. Results for the simple single base model, p1 = 0.5, λ1 = 1, µ0 = 2J1 , µ1 = J1
Production cell
1 − p1
λ1
λ1
p1
Base 1 repair µ1
λ1
µ1
J1 machines
Transport γ1 S1
µ0 S0 Transport γ2 S2
Production cell
µ2
λ2
Base 2 repair µ2 λ2
p2
λ2
1 − p2
J 2 machines
3URGXFWLRQFHOO
1 − p1
λ1
λ1
p1
%DVHUHSDLU µ1
m11
m12 λ1
µ1
t1 PDFKLQHV
j1
k 7UDQVSRUW γ1 RSHUDWLRQDO
'HSRWUHSDLU µ0 7UDQVSRUWGHSRWWREDVHV
n1
µ0 n2 7UDQVSRUW γ2
t2
3URGXFWLRQFHOO
µ2 m22 λ2
m21
%DVHUHSDLU µ2 λ2
p2
λ2
1 − p2
PDFKLQHV
j2
RSHUDWLRQDO
Fig. 9. The modified multi-base system for L = 2
the bases is taken into account explicitly. The transport lines are modeled as ample
servers with exponential service rate γ for the transport to base = 1, . . . , L. The
number of machines in transport to base is denoted by the random variable t .
The transport from the bases to the depot is not taken into account.
As in the simple model, the synchronization queues at the bases can be replaced
by ordinary queues as is depicted in Figure 9.
The vector m1 = (m11 , m21 , . . . , mL1 ) denotes the number of machines in
base repair ( = 1, . . . , L) and the vector m2 = (m12 , m22 , . . . , mL2 ) denotes
the number of spares at the bases ( = 1, . . . , L). The variable n1 stands for the
number of machines in depot repair and n2 is the number of spare machines at the
236 L. Spanjers et al.
depot. The vector k0 = (k01 , k02 , . . . , k0L ) denotes the backorders at the depot,
L base ( = 1, . . . , L). The total number of backorders at the depot
originating from
equals k = =1 k0 . The machines in transit to the bases are given by the vector
t = (t1 , t2 , . . . , tL ) and the numbers of machines operating in the production cells
are expressed in vector j = (j1 , j2 , . . . , jL ). The sum of the number of machines in
base stock and the number of machines operating in the production cell is denoted
in the vector b = (b1 , b2 , . . . , bL ), where b = m2 + j .
As a result of the operating inventory control policies, for n1 = n1 , n2 = n2 ,
k0 = k0 , t = t , m1 = m1 , m2 = m2 and j = j the following equations must
hold:
n1 + n2 − k = S0 , (16)
n2 · k = 0, (17)
and for = 1, 2, . . . , L :
k0 + t + m1 + m2 + j = S + J , (18)
m2 · (J − j ) = 0. (19)
From these relations it follows immediately that k0 , n1 , t and m1 completely
determine the state of the system. Therefore, the system can be modeled as a
continuous time Markov chain with state description (k0 , n1 , t, m1 ).
Remark 3 In the vector that denotes the number of backorders originating from the
bases, k0 = (k01 , k02 , . . . , k0L ), it is not taken into account that the order of the
backorders matters. Since an FCFS return policy is assumed, this order should be
known. Nevertheless, in this model all states with similar numbers of backorders
per base, are aggregated into one state. This aggregation step will not have a big
influence on the results, but it will considerably simplify the analysis.
3.2 Approximation
Production cell
1 − p1
λ1
λ1
p1
Base 1 repair µ1
m11
m12 λ1
µ1
t1 j1 machines
Transport γ1 operational
Depot repair
min( R0 , S 0 + k ) µ 0 / ∞ Transport depot to bases
k
t2
Transport γ2
Production cell
µ2 m22 λ2
m21
Base 2 repair µ2 λ2
p2
λ2
1 − p2
j2 machines
operational
The first spare that finishes repair will fulfill the just arrived request. With
probability 1 − q spares are available and the arriving request does not have to wait.
This aggregated network is depicted as a Typical-server Closed Queuing Network
in Figure 10. The depot repair shop is modeled as a typical server. In case of no
backlog (k = 0) the service rate equals infinity with probability 1 − q and equals
min(S0 , R0 )µ0 with probability q. In all other cases (k > 0) the service rate equals
min(k + S0 , R0 )µ0 .
To determine q Norton’s theorem is used once more. As in Subsection 2.3 each
base (the transport line, the base repair shop and the production cell) is replaced by a
state dependent server. To determine the transition rate of this state dependent server,
each base-part of the network is short circuited and its throughput is calculated.
This throughput operates as the service rate of the state dependent server. The
new network with the state dependent servers and the short circuited networks are
depicted in Figure 11.
Once again the evolution of n1 can be described as a birth-death process. The
(approximated) transition diagram for n1 = 0, . . . , S0 is given in Figure 12.
Let T H (i) be the throughput of the subnetwork replacing base ( = 1, . . . , L)
with i jobs present. As in the simple model only the behavior for n1 ≤ S0 needs to
238 L. Spanjers et al.
Production cell
1 − p1
λ1
λ1
p1
Base 1 repair µ1
m11
m12 λ1
µ1
TH1 (i) t1 j1 machines
k
TH1(i) Transport γ1 operational
Depot repair µ0
n1
µ0 n2 TH 2 (i) t2
Transport γ2
TH 2 (i) Production cell
µ2 m22 λ2
m21
Base 2 repair µ2 λ2
p2
λ2
1 − p2
j2 machines
operational
a b
Fig. 11. a The new network with state dependent servers. b The short circuited networks
∑ TH l ( J l + S l ) ∑ TH l ( J l + S l ) ∑ TH l ( J l + S l ) ∑ TH l ( J l + S l )
l l l l
0 1 2 i i+1 S 0 −1 S0
µ0 2µ0 min(i, R0 ) µ 0 min(i + 1, R0 ) µ 0 min(S 0 − 1, R0 ) µ 0 min(S 0 , R0 ) µ 0
be studied to determine q. Take δ = T H (J + S )/µ0 , then
1
P (n1 = n1 ) = δ n1 ,n1 P (n1 = 0) for n1 = 0, . . . , S0
k=1 min(k, R0 )
(22)
and
P (n1 = S0 ) P (n = S0 )
q= = S0 1
P (n1 ≤ S0 ) n =0 P (n1 = n1 ) 1
δ S0 S0 1
P (n1 = 0)
k=1 min(k,R0 )
= S0
n1 =0 δ n1 n1 1
min(k,R0 )
P (n1 = 0)
k=1
δ S0 S0 1
min(k,R0 )
= S0 k=1
. (23)
n1 =0 δ n1 n1 1
min(k,R0 )
k=1
The throughputs can be obtained by applying a standard MDA algorithm (see [4])
on the short circuited product form networks as shown in Figure 11.
The steady state marginal probabilities as well as the main performance mea-
sures for the aggregated system can be found by using an adapted Multi-Class
Marginal Distribution Analysis algorithm (see Buzacott and Shanthikumar [4]
for ordinary Multi-Class MDA). To see this, introduce tokens of class with
Closed loop two-echelon repairable item systems 239
2. z:=z+1. L
3. For all states z ∈ {z| =1 z () = z and z () ≤ J + S } execute steps 4
through 6.
4. Compute the sojourn times for = 1, . . . , L for which z () > 0 from:
z−1
() k+1
EW0 (z) = p0 (k|z − e )
min(R0 , S0 + k + 1)µ0
k=1
q
+ p0 (0|z − e ),
min(R0 , S0 + 1)µ0
z−1
() b − J + 1 1
EWlb (z) = plb (b |z − e ) + ,
J λ λ
b =J
z−1
() m1 − R + 1 1
EWlm (z) = plm (m1 |z − e ) + ,
R µ µ
m1 =R
() 1
EWlt (z) = .
γ
()
5. Compute T H0 (z) for = 1, . . . , L if z () > 0 from:
() z ()
T H0 (z) = () () () ()
,
V0 EW0 + i∈{b,m,t} Vi EWi
() ()
and if z () = 0 then T H0 (z) = 0. Compute T Hi (z) for = 1, . . . , L and
i ∈ {b, m, t} from:
() () ()
T Hi (z) = Vi T H0 (z).
6. Compute the marginal probabilities for all stations from:
L
()
µ0 min(R0 , S0 + 1) p0 (1|z) = T H0 (z) q p0 (0|z − e ),
=1
L
()
µ0 min(R0 , S0 + k) p0 (k|z) = T H0 (z)p0 (k−1|z−e ) for k=2, . . ., z,
=1
and for = 1, . . . , L from:
()
λ min(J , b )plb (b |z) = T Hlb (z)plb (b − 1|z − e )
for b = 1, . . . , z,
()
µ min(R , m1 ) plm (m1 |z) = T Hlm (z) plm (m1 − 1|z − e )
for m1 = 1, . . . , z,
()
γ t plt (t |z) = T Hlt (z) plt (t − 1|z − e )
for t = 1, . . . , z.
0
Compute pj (0|z) for j ∈ {lb, lm, lt} ∪ {0} from:
z
pj (0|z) = 1 − pj (y|z).
y=1
Closed loop two-echelon repairable item systems 241
L
7. If z = =1 J + S then stop; else go to step 2.
With the adapted Multi-Class MDA algorithm presented above, the marginal
probabilities of the system as well as the throughputs and the sojourn times can be
approximated. From these, various performance measures can be computed. In the
next section some results obtained by the algorithm will be compared with results
from simulation.
3.3 Results
In this section results obtained by the adapted Multi-Class MDA algorithm from
the previous section will be presented. They will be compared to results obtained by
simulation. For each base we are interested in the availability, that is the probability
that the maximum number of machines is operating in the production cell. For base
this is denoted by A for = 1, . . . , L. Furthermore we are interested in the
expected number of machines operating in the production cell, denoted by Ej for
base =1, . . . , L. For =1, . . . , L the performance measures can be computed by
A = P (j = J ) = P (b ≥ J ) = P (k 0 + m1 ≤ S ), (24)
Ej = E(J − [k 0 + m1 − S ] ) +
= (J − [k0 + m1 − S ]+ )P (k0 , m1 ). (25)
k0 ,m1
In Table 2 and Table 7 in the Appendix, the parameter settings for some repre-
sentative test problems are given. In this section, we consider dual base systems
(L = 2); in the appendix we also have examples of systems with three (L = 3) and
four bases (L = 4). The other parameters in this case are given in Table 2 with
J , the maximum number of working machines at base ,
S , the maximum number of stored items at base (or at the depot),
λ , the breakdown rate of individual machines at base ,
µ , the repair rate of individual machines at base (or at the depot),
R , the number of repairmen at base (or at the depot),
242 L. Spanjers et al.
The first 10 models are symmetric, that is the same parameter values apply to both
bases. The other 10 problems concern asymmetric cases. It is obvious that a large
number of input parameters is required to specify a given problem. This makes it
difficult to vary these parameters in a totally systematic manner. In Albright [2] it is
shown that traffic intensities are good indicators of whether a system will work well
(minimal backorders) and are better indicators than the stock levels. Therefore we
selected most of the test problem parameter settings by selecting values of the traffic
intensities, usually well less than 1, and then selecting parameters to achieve these
traffic intensities. For the base repair facility, the traffic intensity ρ is defined as
ρ = J λ p /R µ , (26)
the maximum failure rate divided by the maximum repair rate. Similarly, the depot
traffic intensity ρ0 is defined as
L
ρ0 = J λ (1 − p )/R0 µ0 . (27)
=1
The results are given in Table 3 and Table 8 in the Appendix. The simulation leads
to 95 % confidence intervals. The simulation method was the so called replication
deletion method where the warmup period was found by Welch’s graphical proce-
dure (cf. Law and Kelton [7]). To compare the approximations with the simulation
results, the deviation from the approximation to the midpoint of the confidence
interval is calculated. These percentage deviations are given as well.
From the results it can be concluded that the approximations are very accurate.
The maximum deviation for the availability as well as the relative deviation for the
expected number of working machines, is well less than 1% and all approximat-
ing values lie within the confidence intervals. Furthermore, all types of problems
exhibited similar levels of accuracy.
4 Optimization
Table 2. Parameter settings for test problems multi-base model with transportation (1)
depot
Problem S0 µ0 R0 ρ0 base J S λ µ R p γ ρ
The aim is to maximize the overall performance of the system under a budget con-
straint for stocking costs. For the overall performance of the two-echelon repairable
item system, the total availability Atot , defined by
L
=1 J λ A
Atot = L
,
=1 J λ
244 L. Spanjers et al.
The most straightforward solution method to find optimal stock levels, is the brute
force method. This method simply checks all feasible allocations and picks the one
which gives the highest total availability. By assuming that Atot is an increasing
function, the brute force can be improved by considering only allocations on the
boundary of the feasible region, that is those allocation where adding another spare
part would lead to an infeasible allocation. Even this improved brute force approach
turns out to be rather time consuming.
In Zijm and Avsar [10], a greedy approximation procedure is given to find the
optimal allocation of stocks for an open two-indenture model. This method can
also be applied on closed two-echelon repairable item systems.
At the start of the heuristic algorithm no spares are allocated. One repeatedly allo-
cates one spare to the location that leads to the maximum increase in total availability
per unit of money invested, under the constraint that the allocation is feasible. The
heuristic continues as long as this maximum increase is positive; it can be presented
as follows:
Algorithm 6 Approximative optimization method (greedy approach)
Let ˆ = arg max ∆ . If ∆ˆ ≤ 0 then stop; otherwise repeat this step after setting
Ŝˆ = Ŝˆ + 1 and Ĉ = Ĉ + cˆ.
3. (Solution) The resulting stock allocation (Ŝ0 , Ŝ1 , . . . , ŜL ) is the approximative
solution to the optimization problem.
The greedy heuristic presented above builds on the observation that
Atot (S0 , . . . , SL ) tends to behave as an increasing multi-dimensional concave func-
tion, in particular for not too small values of Si , i = 1, . . . , L. This observation
246 L. Spanjers et al.
In this section results are obtained for several test problems. The results ob-
tained by the brute force approach are compared to the results found by the greedy
approach. Even when the greedy approach gives a different allocation for spare
items, the total availability only decreases slightly.
In Table 4 several test problems are presented. The parameters in this case are
J , the maximum number of working machines at base ,
λ , the breakdown rate of individual machines at base ,
µ , the repair rate of individual machines at base (or at the depot),
R , the number of repairmen at base (or at the depot),
p , the probability that a machine can be repaired at base ,
γ , the transportation rate to base ,
c , the costs to store an item at base (or at the depot),
C, the available budget for storing items.
Note that the the maximal stock sizes S0 , S1 , . . . , SL and the total availability Atot
are not given but computed by either the brute force approach (Atot,bf and S,bf )
or by Heuristic 6 (Atot,greedy and S,greedy ). The numerical results indicate that
the greedy approach yields good results.
Appendix
In this appendix, numerical results are given for various parameter settings in our
model. In most cases, the availability is high as desired in practical situations. In
Table 5 and Table 6 the focus is on the single base model. Multiple base models
are considered in Table 7 and Table 8.
Table 5. Results for the simple single base model, p1 = 0.5, λ1 = 1, µ0 = J1 , µ1 = J1
Table 6. Results for the simple single base model, p1 = 0.25, λ1 = 1, µ0 = 2J1 , µ1 = J1
Table 7. Parameter settings for test problems multi-base model with transportation (2)
1 Introduction
This paper addresses the problem of determining optimal inventory and production
control decisions for an integrated production-inventory (PI) system in which mul-
tiple products are made-to-stock through a functionally oriented shop. As can be
seen in Figure 1, inventory is carried to service customer demand. The customers
require that their orders are serviced with a target fill rate. Unsatisfied demand is
backordered. Customers arrive according to a renewal process that is characterised
by interarrival times with a known mean and squared coefficient of variation (scv).
The authors would like to thank two anonymous referees and the editor for many valuable
suggestions.
254 P.L.M. Van Nyen et al.
The inventory points generate replenishment orders that are, in this integrated PI
system, equivalent to production orders. There is a fixed cost incurred every time a
production order is generated. In this paper, all inventory points are controlled using
periodic review, order-up-to (R, S) inventory policies. Other inventory policies can
be embedded in the framework, if desired. The production orders are manufactured
through the shop. We assume ample supply of raw material. The production system
consists of multiple functionally oriented workcenters through which a consider-
able number of different products can be produced. Each of the products can have
a specific serial sequence of production steps, which results in a job shop routing
structure. The production orders for different products compete for capacity at the
different workcenters, where they are processed in order of arrival (FCFS priority).
Before starting the production of an order, a set-up that takes a certain time and cost
is performed. The total production time of a production order depends on its size.
When the production of the entire order is finished, it is moved to the inventory point
where the products are temporarily stored until they are requested by a customer.
We consider situations in which the average demand for end products is relatively
high and stationary. Since the production system is characterized by considerable
set-up times and costs, the products are produced in batches. We assume that a
centralized decision maker controls both the inventory points and the production
system. Then, the objective of the centralized decision maker is to minimize the
sum of fixed set-up costs, work-in-process holding costs and final inventory holding
costs. Moreover, we impose that customer demand has to be serviced with a target
fill rate. The decision variables that can be influenced by the decision maker are the
review periods and the order-up-to levels of the products.
Typically, integrated PI systems with job shop routing structure and stochastic
demand interarrival times, set-up times and processing times can be found in metal
or woodworking companies, e.g. the suppliers of the automotive or aircraft building
industry. In particular, the advent of Vendor Managed Inventory (VMI) has forced
manufacturing companies to integrate production and inventory decisions.
It appears that it is impossible to analyse this integrated PI system exactly and
consequently, it is very difficult (if possible) to solve the optimization problem faced
Control of multi-product multi-machine production-inventory systems 255
the influence of the review periods on the capacity utilization of the workcenters and
on the order throughput times (thus dealing with the multi-product aspect of the PI
system under study). Similarly to [13, 15] and [16], we observe a convex relationship
between the review periods and the order throughput times. The control decisions
situated at the lower level of the hierarchical framework of Lambrecht et al. [16],
the determination of the release dates and the sequencing decisions, are dealt with
in a straightforward way. Firstly, the production orders are released immediately.
Secondly, the sequencing of all orders is done using the FCFS priority rule. The use
of the FCFS rule allows for the queueing network to be analyzed using standard
queueing network analysers. However, from production control theory it is known
that other sequencing policies (priority rules or scheduling methods) may lead to
substantial time and cost savings. An overview of priority rules and scheduling
methods can be found in [21]. We believe that sequencing policies, other than the
FCFS priority rule, can be embedded in the hierarchical control framework in the
same way as Lambrecht et al. [16] incorporate the shifted bottleneck procedure [1]
for the scheduling of production orders. Observe that by choosing reasonable – not
necessarily optimal – review periods or lot sizes, the sequencing decisions situated
lower in the control hierarchy become significantly easier to make.
To the best of our knowledge, this paper is the only research that studies the
specific problem of minimizing the total relevant costs in integrated multi-product
multi-machine PI systems with job shop routing structure and stochastic demand
interarrival times, set-up times and processing times. For related problems, however,
solution methods have been developed. First, for the deterministic version of this
problem, Ouenniche et al. [19, 20] present heuristics that are based on cyclical
production plans. The cost-optimal cyclical plans are generated using mathematical
programming techniques. We believe that it is not possible to apply the results
of Ouenniche et al. directly in our setting because the influence of variability in
the demand and manufacturing processes makes the proposed production plans
infeasible.
A second relevant contribution is the literature review on the stochastic lot
scheduling problem (SLSP) by Sox et al. [25]. This paper gives an extensive
overview of current research on the control of multi-product PI systems in which
the production system consists of a single workcenter. The critical assumption in
this body of research is that the performance of the production system is deter-
mined by a single bottleneck process. Although this assumption may be valid in
some situations, it is certainly not in others. The heuristic proposed in this paper
explicitly considers situations in which the production system consists of multiple
workcenters.
Thirdly, Benjaafar et al. [4, 5] study managerial issues related to PI systems,
such as the effect of product variety and the benefits of pooling. Benjaafar et al.
[4] proposes a method to jointly optimize batch sizes and base-stock levels for
a multi-product single workcenter integrated PI system controlled by continuous
review (s, Q) policies.
Fourthly, Amin and Altiok [3] and Altiok and Shiue [2], study production
allocation issues in multi-product PI systems. More specifically, they address such
issues as “when to switch production to a new product” and “what product to switch
Control of multi-product multi-machine production-inventory systems 257
to”. They propose to handle the first issue with a continuous review inventory
control policy. The second issue is resolved by using switching policies that are
based on priority structures. Amin and Altiok [3] use simulation to compare a
number of manufacturing strategies and switching policies for a serial multi-stage
production system with finite interstage buffer space. Altiok and Shiue [2] develop
an approximate analytical model to compute the average inventory levels under
different priority schemes for a single machine production system.
Fifthly, Rubio and Wein [22] study a PI system where the production system
is modeled as a Jackson queueing network. Under this assumption, they derive a
formula characterizing the optimal base stock level. The main difference with our
approach is that they do not have batching issues in their model because of the
absence of set-up times and set-up costs. Note that it is precisely the batching issue
that makes the problem very hard to solve.
As a sixth contribution, we mention the work of Lambrecht et al. [16] who
study the control of a stochastic make-to-order job shop. They describe a lot sizing
procedure that minimizes the expected lead times and thus, the expected work-in-
process costs. To this end, they model the production environment as a general
open queueing network. Vandaele et al. [28] successfully implemented the method
described in Lambrecht et al. [16] to solve lot sizing problems with the medium-
sized metal working company Spicer Off-Highway. Their research shows that the lot
sizing procedure is capable of solving realistic, large-scale problems. Our research
builds on the work of Lambrecht et al. We extend their make-to-order model by
including inventory points so that we obtain an integrated PI system in which
products are made-to-stock instead of made-to-order.
Seventhly, Bowman and Muckstadt [7] present a production control approach
for cyclically scheduled production systems that are characterized by demand and
process variability. The production management can delay the release of material
to the floor and increase production in a cycle to anticipate on demand and to avoid
overtime in future cycles. The control approach uses estimates for the cycle time
and the task criticality that are obtained from a Markov chain model. Using these
estimates, the control approach seeks a trade-off between inventory holding costs
and overtime costs.
Finally, Liberopoulos and Dallery [18] propose a unified modeling framework
for describing and comparing several production-inventory control policies with lot
sizing. The control method used in this paper is based on (R, S) inventory policies
and is comparable to the Reorder Point Policies (RPPs) described in [18]. More
insights on RPPs can be found in their paper. Our work differs from their work
in several aspects. First, they study a N-stage serial PI system through which a
single product is manufactured while we focus on a single stage PI system in which
multiple products are produced. Secondly, Liberopoulos and Dallery use queueing
network representations to define (not analyze or optimize) several production-
inventory control policies that decide when to place and release replenishment
orders at each stage. Our work focuses on a (R, S) inventory rule based control
policy for which we not only define the control policy in place, but also present a
heuristic to make the production and inventory control decisions that minimize the
relevant costs.
258 P.L.M. Van Nyen et al.
The remainder of this paper is organized as follows: Section 2 presents the for-
mal problem statement; Section 3 proposes a heuristic to determine review periods
and order-up-to levels that minimize the relevant costs; in Section 4 the perfor-
mance of the heuristic is tested in an extensive simulation study; in Section 5, the
results of the simulation study are discussed; and finally, Section 6 summarizes the
major findings of our paper.
First, we introduce the notation used in this paper. Then, we derive formulas for
the different cost components. After this, a formal problem definition is given.
2.1 Notation
– P : number of products;
– M : number of workcenters in the production system;
– ADi : interarrival times of demand for product i (stochastic variable);
– αi∗ : target fill rate for product i;
Mathematical operators:
The final inventory cost for product i is given by the formula below [24]. The term
between brackets gives the average amount of final inventory at inventory point i,
which consists of half the average order quantity plus the safety stock.
Ri
F ICi = + ssi vi r.
2E AD i
The total cost for product i is simply the sum of its components. Clearly, the total
cost T C for the whole PI system is given by the sum over all products of the total
costs for each product:
P
TC = (SCi + W IP Ci + F ICi ) (1)
i=1
In this system, we have one centralized decision maker who wants to minimize the
total costs of the PI system. As stated in the introduction, the total costs consist
of set-up costs, final inventory holding costs and work-in-process holding costs.
The decision maker has to ensure that the target fill rates are satisfied and that
the review periods are positive. Consequently, the mathematical formulation of the
optimization problem can be stated as:
⎡
⎤
P M
⎣ci Ri + E [T ij ] R i
min −1
vij r + + ssi vi r⎦ (2)
Ri
i=1 j=1
E AD i 2E AD i
subject to:
1. αi ≥ αi∗ for i = 1, ..., P
2. Ri > 0 for i = 1, ..., P
Observe that one can easily compute most of the cost components if the review
periods for all the products are given. However, two variables – the throughput time
260 P.L.M. Van Nyen et al.
in the workcenters Tij and the safety stock ssi – cannot be computed analytically.
In order to find an expression for the throughput times Tij in the production system,
a general open queueing network should be solved. Unfortunately, no exact results
for the throughput times in such a queueing system are known. Consequently, it
is also impossible to find an exact expression for the safety stock ssi since ssi
depends on the average and the variance of the throughput times. In conclusion, it
is impossible to derive exact expressions for these variables and this implies that
our objective function is analytically intractable. Therefore, we have to rely on
estimates to evaluate the cost of a given solution.
In the first phase of the heuristic, near-optimal tactical inventory and production
control decisions are determined. The optimization tool consists of two main ele-
ments: (i) an analytical model that approximates the relevant costs given a vector
of review periods and (ii) a greedy search method that finds the vector of review
periods that minimizes the relevant costs. In the subsections below, both elements
are described in more detail.
The approximate analytical model follows the same basic idea as Zipkin [34].
First, we use a standard inventory model to represent the (R, S) inventory policy
262 P.L.M. Van Nyen et al.
of every product, see e.g. [24]. Next, we use standard results on open queueing
networks to represent the production system. Our open queueing network is solved
using the Queueing Network Analyser developed by Whitt [32]. Similarly to [16],
the expressions for the performance measures of the queueing network are written
as a function of the production lot size. Finally, we link both submodels together
using concepts from renewal theory; see [8]. The resulting analytical model is an
approximation to the real PI system. Similarly to the approach proposed by Zipkin,
we sacrifice accuracy for the sake of simplicity and computational tractability.
are independent and identically distributed (i.i.d.) and independent of the set-up
time. Then, the variance of the net processing times, excluding set-up time, equals
the variance of the sum of a variable number of variable processing times, which
can be computed with a formula given by e.g. [24]. Consequently, the variance of
the processing times of the orders of product i at machine j is given by:
σ 2 PijP = E [Ni ] σ 2 [Pij ] + E 2 [Pij ] σ 2 [Ni ] + σ 2 [Lij ]
2 D −1 D (3)
= Ri E −1 AD 2
i σ [Pij ] + E [Pij ] Ri c Ai E
2
Ai + σ 2 [Lij ]
where ki is the so-called safety factor for product i that depends on the target fill
rate αi∗ . Given a target fill rate, [24] presents a very accurate rational approximation
for ki in the case of normally distributed demand. Finally, the safety stock ssi for
product i can be computed as: ssi = ki σ[DiTi +Ri ]. This step results in a vector of
order-up-to levels S = (S1 , ..., Si , ..., SP ) that correspond to the vector of review
periods R = (R1 , ..., Ri , ..., RP ).
Step 4: Estimate costs
In the previous steps, we presented an approach to approximate the expected
throughput times E[Tij ] and the safety stocks ssi , given a vector R =
(R1 , ..., Ri , ..., RP ). Using the expressions for the different cost components given
in Section 2.1 we can now compute the total costs T C corresponding to a solution R.
In the first phase of the heuristic, we use an approximate analytical model to de-
termine the near-optimal review periods R∗ and initial settings for the order-up-to
levels S . Because of the use of approximations, the realized fill rates may be lower
Control of multi-product multi-machine production-inventory systems 265
than the target fill rates. Given the constraints on the fill rate in the formal problem
statement this may result in the infeasibility of the solution. Also, it may happen
that the realized fill rates are higher than the target fill rates. Obviously, this leads
to a solution that is unnecessarily expensive. For these reasons, we add a second
step to the heuristic in which the order-up-levels are fine-tuned.
In this second step of the heuristic, we use a procedure proposed by Gudum and
de Kok [10]. Their procedure builds on the following observation. Given that the
inventory points are controlled by periodic review (R, S) policies with full backo-
rdering, the size and the timing of replenishment orders is determined by the review
periods only. In an integrated PI system, this implies that the arrivals of production
orders to the production system, as well as the processing times of the production
orders at the different workcenters are determined by the review intervals and not
by the order-up-to-levels. Therefore, the throughput times of the replenishment
orders are completely determined by the review intervals. This also implies that
a change in the order-up-to levels only influences the customer service levels. In
conclusion, this observation states that a given selection of the review periods fully
determines the stochastic behavior of the PI system and that the order-up-to levels
can be adjusted to achieve a certain customer service level without affecting the
behavior of the PI system. For more details on this observation, see [10].
In our heuristic, the fine-tuning phase starts by simulating the initial solution
to estimate the realized fill rates corresponding to the solution (R*, S). Based on
a trace of the inventory levels in the first simulation run, a procedure developed in
[10] is used to fine-tune the order-up-to levels. Note that we do not adjust the review
periods. The procedure makes use of the observation above, stating that changes in
the order-up-to levels only influence the fill rates. This procedure allows us to set the
order-up-to levels to the lowest possible levels that satisfy the fill rate constraints,
denoted as S∗ . This results in the solution (R∗ , S∗ ). From a computational point
of view, the procedure developed in [10] is attractive because it uses only one
simulation run instead of iterative simulation runs.
In the third and final phase of the heuristic, a simulation experiment is used to
estimate the costs and operational performance characteristics corresponding to
the solution (R∗ , S∗ ) . Simulation is used because of its estimation accuracy. All
the costs that are listed in Section 2.2 are estimated. The operational performance
characteristics that are estimated include the fill rates, the throughput times at the
different workcenters and the utilization of the workcenters.
The heuristic uses an approximate analytical model to determine the tactical pro-
duction and inventory decisions, which may result in suboptimal decisions. In this
section we test the performance of our heuristic in an extensive simulation study.
A specific problem instance is studied in detail in order to gain understanding of
266 P.L.M. Van Nyen et al.
In this section, we discuss how the approximate analytical model uses the review
periods to minimize the total relevant costs. The mechanisms behind the selection
of the review periods are illustrated by comparing the detailed output of the opti-
mization tool with the output of a simple heuristic method for setting the review
periods. More specifically, we use the economic order quantity (EOQ) expressed
as a time supply to set the review periods, see e.g. [24]:
1
2ci E AD i
Ri = (5)
vi r
The EOQ formula completely ignores the impact of the lot sizing decision on the
production system. Hence, it does not take into account the costs that are related
to capacity utilization and throughput times, i.e. work-in-process and safety stock
costs. We use the uncapacitated EOQ approach to solve one random set of 48
problem instances of the experimental design. Below, the results for all 48 problem
instances are summarized, but first we study in detail one specific problem instance.
This problem instance is selected because it clearly demonstrates how our heuristic
works. In this way, the reader can gain understanding of the mechanisms that
are embedded in the heuristic. The selected problem instance is characterized by:
ρnet = 0.85; Lij ∈ [90, 180] min .; ci ∈ [6.67, 13.33] ¼ ; αi∗ = 0.98.
In Tables 1 up to 3, we present the detailed output of the simulation of the
heuristic (HEU ) and the uncapacitated EOQ method for this problem instance.
From the analysis of the decision variables and the corresponding performance
measures, we learn how the optimization tool works and how it tries to achieve the
minimal total relevant costs. Table 1 displays the decision variables (review periods
and order-up-to levels) and the resulting throughput times, characterized by their
expectation E[T] and standard deviation σ[T ]. It can be seen from Table 1 that
the heuristic proposes considerably smaller review periods than the uncapacitated
EOQ method. The order-up-to levels are lowered accordingly. Remark that the
review periods of the different products are decreased in a non-proportional way
in order to account for the specific processing characteristics of every product. The
impact of the smaller review periods on the expectation and standard deviation of
the throughput times is high: the expected throughput times decrease by 36.5% on
average while the standard deviations of the throughput times decrease by 42.4%
on average.
Table 2 shows the impact of the changes in the decision variables on the relevant
costs. Since the solution of the heuristic uses considerably smaller review periods,
the set-up costs are substantially higher compared to the uncapacitated EOQ solu-
tion (+58.4%). However, since the review periods chosen by the heuristic lead to
shorter and more reliable throughput times, the work-in-process costs and the final
inventory holding costs decline significantly (−35.8% and −36.5%). Overall, this
leads to a cost decrease realized by the heuristic versus the uncapacitated EOQ
approach of 9.0%.
Finally, Table 3 gives insight into the mechanisms embedded in the optimization
tool. We use elementary insights from queueing theory to illustrate the trade-offs
268 P.L.M. Van Nyen et al.
Table 1. Decision variables and throughput times for one problem instance: heuristic vs.
uncapacitated EOQ method
Table 2. Cost components for one problem instance: heuristic vs. uncapacitated EOQ
method
that are made by the heuristic. From elementary queueing theoretical results for
the GI/G/1 queue, e.g. Hopp and Spearman [12], we learn that there are four main
elements affecting the expectation of the throughput times E[T] on the machines:
(i) utilization ρ; (ii) variation of the arrivals c2a ; (iii) average processing time tp ;
Control of multi-product multi-machine production-inventory systems 269
Table 3. Operational characteristics of production system for one problem instance: heuristic
vs. uncapacitated EOQ method
(iv) variation of the processing times c2p . This insight is based on the Kingman
approximation for the expectation of the throughput times in a GI/G/1 queue [14]:
c2a + c2p ρ
E [T ] = tp + t p (6)
2 1−ρ
From Table 3, it can be seen that the optimization tool adapts the review periods
so that the variation of the arrivals and the expectation and the variation of the
processing times are reduced. However, this happens at the expense of increased
utilization levels. We may conclude that the optimization tool ‘harmonizes’ the
review periods of the different products so as to obtain the best balance between
utilization and variability.
In the job shop production system under study, the departure process of a ma-
chine is the arrival process to the next machine in the routing of a product. An
elementary approximation, due to Hopp and Spearman [12], for the scv of the
departure process leaving a queue is:
c2d = ρ2 c2p + 1 − ρ2 c2a (7)
From this approximation, it can be observed that when the utilization of the ma-
chines is high, it is important to achieve low variation in the processing times in
order to obtain an arrival process with low variability to the next machine. Table
3 shows that the heuristic realizes a low variation in the processing times, while
the utilization levels are high (around 90%). From Table 1, it can be seen that the
actions taken by the heuristic, based on the mechanisms presented above, result
in shorter and less variable throughput times. Note that the mechanisms described
above are embedded in the proposed approximate analytical model using advanced
queueing theoretical results developed in [32, 33].
Now, we briefly present the results for the 48 instances that were solved using
the uncapacitated EOQ method. In 14 out of 48 problem instances, the unca-
pacitated EOQ approach resulted into a solution that is infeasible with respect to
production capacity. For the 34 feasible instances, the uncapacitated EOQ solution
is on average 5.2% more expensive than the solution of the heuristic. The maxi-
mum cost increase reported on this set of experiments is 10.1%. The conclusion of
270 P.L.M. Van Nyen et al.
these experiments is that the uncapacitated EOQ method may work relatively well
compared to the heuristic, but since the uncapacitated EOQ approach does not take
into account capacity issues, it may result in unnecessarily expensive solutions or
in solutions that are infeasible with respect to production capacity (and that require
capacity expansion in the form of overwork, outsourcing, etc.).
In order to avoid that infeasible solutions are obtained, one can add capac-
ity restrictions to the uncapacitated EOQ method. Doing so, we obtain the fol-
lowing mathematical programming problem, which we call the capacitated EOQ
approach:
P
# %
ci vi r
min + Ri
Ri
i=1
Ri 2E AD i
subject to :
P
# % (8)
E [Pij ] E [Lij ]
1. D + ≤ ρmax
j f or j = 1, . . . , M
i=1
E Ai R i
2. Ri > 0 f or i = 1, . . . , P
The objective function of this mathematical program is identical to the cost function
of the uncapacitated EOQ method. The first set of constraints imposes that the
machine utilization must be lower than a maximum allowable utilization level
ρmax
j . The second set of constraints states that the review periods should be strictly
positive. Note that the objective function and the constraints are convex in the review
periods Ri . This convex programming problem can easily be solved to optimality
using the commercially available CONOPT algorithm. The CONOPT algorithm
attempts to find a local optimum satisfying the Karush-Kuhn-Tucker conditions. It
is well known that for convex programming problems a local optimum is also the
global optimum (see e.g.[11]).
The main difficulty that arises with this capacitated EOQ method is the choice
of the maximum allowable utilization level ρmax j . For deterministic problems ρmaxj
is usually chosen so that all production capacity is utilized, i.e. ρmax
j = 1. Clearly,
in stochastic settings ρmax
j should be lower than 1 for reasons of stability. It is,
however, not obvious how the precise value of ρmax j should be chosen. If ρmax
j is
chosen too low, this leads to long review periods (in order to reduce the capacity
utilization due to set-ups) and thus to high cycle stocks. On the contrary, if ρmax
j is
chosen too high, this results in high congestion, large amounts of work-in-process,
long throughput times and high safety stocks. A priori, the EOQ model is not able
to predict which value of ρmax
j leads to the lowest total relevant costs. Therefore,
in our experiments we vary ρmax j over a range of reasonable values and observe the
resulting total relevant costs.
We use the capacitated EOQ method to solve the 48 problem instances that
were also solved using the uncapacitated EOQ method. The value of ρmax j is set
to 0.90, 0.95 and 0.99. Let us define the deviation in the total costs between the
capacitated EOQ method and our heuristic as:
T C eoq − T C heu
∆= × 100%.
T C heu
Control of multi-product multi-machine production-inventory systems 271
Table 4. Summary statistics for ∆, the relative deviation in total costs between the capacitated
EOQ method and the proposed heuristic (instances with set-up costs > 0)
ρmax
j = 0.90 ρmax
j = 0.95 ρmax
j = 0.99
Table 5. Summary statistics for ∆, the relative deviation in total costs between the capacitated
EOQ method and the proposed heuristic (instances with set-up costs = 0)
ρmax
j = 0.90 ρmax
j = 0.95 ρmax
j = 0.99
Tables 4 and 5 give the minimum, average and maximum of ∆, respectively for the
instances with set-up costs > 0 and the instances with set-up costs = 0. The results
are shown for the different levels of the net utilization of the machines ρnet .
The results in Tables 4 and 5 show that the proposed heuristic always outper-
forms the capacitated EOQ method. The capacitated EOQ approach may work
reasonably well, provided that a good choice is made for ρmax j : the lowest ∆ ob-
served in this set of instances is 1.7%. However, one can also observe that an
inappropriate choice of ρmax
j may result in a very poor performance: the maximum
of ∆ in this set of instances is 224.5%. As mentioned before, the EOQ approach
does not provide any guideline for choosing the value of ρmax j .
For the instances with set-up costs > 0 and ρmax j = 0.99, the performance of
the capacitated EOQ method seems reasonable: the average of ∆ is 5.1% with a
272 P.L.M. Van Nyen et al.
maximum of 10.1%. It appears that in the majority of these instances the capacity
constraints are non-binding so that the solution of the capacitated EOQ method is
identical to the solution of the uncapacitated EOQ method. When ρmax j is lowered,
the performance of the capacitated EOQ method degrades for the instances with
high ρnet . When ρnet = 0.85 and ρmax j = 0.90, the average of ∆ is 25.2% with
a maximum of 80.2%. For the majority of the instances with low and moderate
ρnet , the capacity constraints are also non-binding when ρmax j = 0.90 and 0.95.
Therefore, in these instances the performance of the capacitated EOQ method is
similar to that of the uncapacitated EOQ method and the capacitated EOQ method
with ρmax
j = 0.99.
For the instances with set-up costs = 0, it seems to be even more important to se-
lect the appropriate value of ρmax
j than in the case of set-up costs > 0. For example,
when ρnet = 0.65, the capacitated EOQ method works well when ρmax = 0.90:
the average of ∆ is 3.8% with a maximum of 4.6%. If ρmax is chosen too high, the
performance of the capacitated EOQ method degrades strongly: for ρmax = 0.99,
the average of ∆ is 157.6% with a maximum of 224.5%. Similar results hold for
the instances with ρnet = 0.75. For the instances with ρnet = 0.85, the capacitated
EOQ approach performs rather poorly for all choices of ρmax : the minimum of ∆
reported on this set of instances is 11.7%.
The main conclusion from these experiments is that the capacitated EOQ ap-
proach is very sensitive to the choice of ρmax . Since the appropriate value of ρmax
depends on the specific characteristics of the problem instance, it is difficult to
develop a general rule of thumb for selecting ρmax . The heuristic proposed in this
paper does not suffer from this problem. The approximate analytical model embed-
ded in the heuristic explicitly models the impact of the review periods on capacity
utilization and on congestion phenomena, taking into consideration the specific
characteristics of the problem instance. Therefore, our heuristic is a more robust
and reliable method to set the decision variables.
In this section, we test the optimization quality of the heuristic. The methodology
used for this test warrants some discussion. First, note that the optimal solution
for the problem under study is unknown. Furthermore, no high-quality bounds on
the optimal costs are available, mainly due to the difficulty to find bounds on the
waiting times in the production system. Moreover, to the best of our knowledge, no
other control approaches have been developed for the integrated PI system with job
shop routings and stochastic arrival, processing and set-up times. Consequently, the
heuristic solution cannot be compared to the true optimum, nor to a good bound,
nor to another control approach reported in the literature. In short, there exists no
good benchmark to test the quality of our heuristic. Therefore, we constructed our
own benchmarks in order to test the performance of the heuristic.
First, we test the prediction quality of the approximate analytical model. If the
prediction quality of the approximate model is satisfactory, then one may expect
that the optimization quality of the tool is good. However, when the approximate
analytical model wrongly estimates the absolute value of the costs, but correctly
Control of multi-product multi-machine production-inventory systems 273
Fig. 3. Frequency diagram for relative difference between total costs approximate analytical
model (AAM) and simulation (SIM)
captures the relative behavior of the costs in function of the review periods, the
optimization process may still perform well. In Figure 3, we present the relative
difference between the cost estimates of the approximate analytical model (AAM)
and simulation (SIM) for the solutions proposed by the heuristic for the 240 in-
stances. Since for optimization purposes mainly the absolute value of the relative
deviation is relevant, we group the positive and negative intervals with the same
absolute value. The frequency of negative and positive values of the relative devi-
ation is shown in distinctive colors. From Figure 3, it can be seen that the relative
approximation error is rather small, lying in the range of −17% to 12% on this set
of 240 experiments. From this figure, it also can be seen that for the vast majority
of the instances (more than 92%) the absolute relative approximation error is lower
than 10%. Negative relative differences occur, but in more than 70% of the cases
the relative error is positive. From these results, we conclude that the estimation
quality of the approximate analytical model is satisfactory.
Secondly, in order to test the optimization quality of the heuristic we develop
two simulation based optimization methods to solve several instances of our opti-
mization problem. Simulation based optimization is known to be an accurate but
time-consuming optimization method, see e.g. Law and Kelton [17]. The perfor-
mance of the simulation based optimization methods is compared to the perfor-
mance of our heuristic in terms of the optimization quality as well as the required
computation time.
We use two different simulation based optimization methods:
– A modification of the greedy search algorithm presented in Section 3.1.2. We
use three different step sizes to perform the search along the axes. For each of
the step sizes, we use the greedy search algorithm. The solution of one phase
is used as an input for the next phase. The step sizes for the review periods are
2500, 500 and 100 minutes.
– OptQuest, a commercially available software package developed by Glover
et al. OptQuest combines elements of scatter search, taboo search and neural
networks to find solutions for non-convex optimization problems [17]. We limit
the review periods to the interval [0.5, 2] times the review periods R∗ proposed
274 P.L.M. Van Nyen et al.
by our heuristic. Moreover, we use a step size of 100 minutes for the review
periods.
Both optimization techniques suggest new vectors R that need to be evaluated
using simulation. In our research, we use the second and the third phase of our
heuristic to evaluate the vector R . The evaluation of a vector R consists of a
tuning phase, see Section 3.2, in which the correct order-up-to levels are computed
based on a simulation run. Next, the total cost of the solution R is estimated using
a second simulation run in the evaluation phase presented in Section 3.3.
In order to reduce the computation time required, we use the solution of the
heuristic as initial solution. The simulation based optimization methods are then
used to improve this initial solution. Furthermore, we limit the number of simulation
subruns to 5. However, even with these measures, the simulation based optimization
techniques take very large amounts of computation time. For this reason, it is
impossible to use the simulation based optimization techniques for all 240 instances
in the experimental design. Instead, we select 15 worst-case instances for which
we apply simulation based optimization techniques.
The selection of the 15 worst-case instances warrants some discussion. Let
us first introduce a lower bound for the total costs in the PI system under study.
The lower bound neglects the impact of the variability in the system as well as
the interaction between different products. More details on the lower bound can
be found in Appendix 2. The computation of the lower bound is only possible for
the 180 instances with set-up costs ci > 0. We compute the relative deviation e1
between the total cost of our heuristic and the lower bound. On the set of 180
instances, e1 usually lies in the interval 10–30% with an average of 20%. In this
paper, we use e1 as an indicator for the potential improvement that can be realised
when simulation based optimization techniques are used. We select 10 instances
with the largest e1 to be optimized using simulation based optimization techniques.
Since these instances have the largest potential for improvement, we call them
worst-case instances. We ensure that only one of the five random instances of the
same combination of levels of the four factors is selected as a worst-case instance.
For the case of set-up costs = 0, we cannot compute the lower bound presented above.
Therefore, we use another criterion to select the instances. Since the optimization
quality of the heuristic depends on the accuracy of the approximate analytical model
presented in Section 3.1.1, we select the 5 instances with the largest deviation
between cost estimate of the approximate model and the total costs estimated with
simulation for the optimal solution R∗ proposed by the heuristic. The selected
instances are worst-case instances on the criterion of approximation performance.
Again, we ensure that only one of the five random instances of the same combination
of levels is chosen. In this way, 15 worst-case instances are selected.
Table 6 summarizes the findings of the simulation based optimization exper-
iments for the 15 worst-case instances. It presents the optimal total costs of our
heuristic T C heu , the greedy search algorithm T C gs and the OptQuest algorithm
T C optq . Also the 90% confidence interval on the total costs is given. Finally, Table
6 presents the relative improvement of the greedy search and OptQuest algorithm
with regard to the solution of the heuristic, defined as:
Control of multi-product multi-machine production-inventory systems
Table 6. Results of simulation based optimization experiments
Exp. nr. T C heu CI (90%) T C gs CI (90%) i1 (%) T C optq CI (90%) i2 (%) max(i1 ,i2 ) (%)
275
276 P.L.M. Van Nyen et al.
277
278 P.L.M. Van Nyen et al.
results in Table 7. The relative improvement does not seem to be directly related
to the relative difference in the average review periods and order-up-to levels. Take
e.g. instance 4 where the OptQuest algorithm increases the average review peri-
ods by 10.1%, while the greedy search algorithm proposes an increase of 4.6%.
For this instance the relative improvement realized by the OptQuest algorithm is
1.8% while the relative improvement for the greedy search algorithm is 1.0%. One
may conclude that larger differences in the review periods lead to larger cost im-
provements. However, this conclusion is contradicted by e.g. instances 9 and 15.
In instance 9, the greedy search algorithm increases the review periods by 9.5%
leading to a cost saving of 0.8%, while the OptQuest algorithm increases the lot
sizes by only 5.3% leading to a larger cost saving of 2.1%. In instance 15, the
greedy search algorithm obtains a cost improvement of 6.1%, the largest observed
on this set of experiments. In order to achieve this cost improvement, the review
periods are reduced by 2.4% on average. Furthermore, it can be observed that even
if the utilization remains almost unchanged, still a substantial cost improvement
can be realized by the harmonization of the review periods. This can e.g. be seen
from instance nr. 10 where the utilization does not change for the greedy search
algorithm, but the costs improve by 2.9%.
From this analysis, we conclude that there is no direct relation between the
average review periods, order-up-to levels, machine utilization and the relative cost
improvement that can be obtained from the use of simulation based optimization
techniques. This conclusion leads to the obvious question how the relative cost
improvement is then realized by the simulation based optimization techniques. We
believe that the answer lies in the mechanisms that were described in Section 4.2.
Similarly to the approximate analytical model, the simulation based optimization
techniques ‘harmonize’ the review periods of the different products so as to obtain
the best balance between utilization and variability. Both simulation based opti-
mization techniques seek the best possible trade-off between the utilization of the
machines, variation of the arrivals, average processing times and variation of the
processing times. In this way, it is possible that relatively small differences in the
average review periods can lead to relatively large cost savings.
In this section we analyse and interpret the results of the heuristic for the 240
instances in the simulation study. This allows us to verify the soundness of the
solutions proposed by the heuristic.
Figures 4–6 show the impact of the four factors on the average costs, review periods
and order-up-to levels. The impact of every factor is discussed successively.
The impact of increases in the net utilization on the total costs is about 17 % and
19 % for changes from 0.65 to 0.75 and from 0.75 to 0.85 respectively. The increase
of the net utilization from 0.65 to 0.75 causes the review periods to decrease, while
Control of multi-product multi-machine production-inventory systems 279
a b
c d
Fig. 4a–d. Impact of factor on average costs: a net utilization – b fill rate – c avg. set-up
times – d avg. set-up costs
a b
c d
Fig. 5a–d. Impact of factor on average review periods: a net utilization – b fill rate – c avg.
set-up times – d avg. set-up costs
they increase when the net utilization is increased from 0.75 to 0.85. The reason for
this pattern is to be found in the mechanisms embedded in the heuristic, as described
in Section 4.2. When the net utilization goes from 0.65 to 0.75, the review periods
are decreased in order to reduce throughput times, work-in-process costs and final
inventory holding costs. However, a further increase in the net utilization requires
that the review periods be slightly increased in order to reduce the impact of the
280 P.L.M. Van Nyen et al.
a b
c d
Fig. 6a–d. Impact of factor on average order-up-to levels: a net utilization – b fill rate – c
avg. set-up times – d avg. set-up costs
set-up times on the capacity utilization. The order-up-to levels on the contrary,
increase steadily when the utilization is increased. This can be explained by two
effects. First, in the experiments, the rise in the net utilization is caused by increases
in the demand rate for the products. The increased demand rate results in higher
demand during a review period and increased cycle stock. Secondly, the rise in
the net utilization increases the congestion in the system, leading to longer order
throughput times and higher safety stocks.
When the fill rates increase from 90 % to 98 %, the average total costs increase
with 11%. In order to account for the increase in the fill rate, the order-up-to levels
are increased. However, the increase in the order-up-to levels is fairly small. This
is due to the fact that the heuristic proposes smaller review periods, which lead to
lower cycle stock. Moreover, the order throughput times are reduced so that less
safety stock is required.
The increase of the average set-up times from 45 to 135 leads to an increase
in the total costs of 12%. The review periods are raised to limit the impact on the
capacity utilization. The increased review periods lead to lower set-up costs, but
also to longer throughput times and higher work-in-process costs. Because of the
increases in the review periods and the throughput times, the order-up-to levels and
the final inventory costs increase.
The set-up costs appear to be the dominant factor in our experimental design:
when the average set-up costs increase from 0 to 10, the average total cost rises with
156%. Further increases in the average set-up costs result in cost increases of 64 %
and 70 %. The review periods and the order-up-to levels increase in a similar fashion.
From these results, it appears that efforts to cut set-up costs (as advocated by e.g.
the Just-In-Time philosophy) may effectively result in large savings in the overall
costs. Figure 4-d presents the division of the total costs (TC) over the different
Control of multi-product multi-machine production-inventory systems 281
Fig. 7. Frequency diagram of allocation of free capacity for set-ups (instances with set-up
costs = 0)
Fig. 8. Frequency diagram of allocation of free capacity for set-ups (instances with set-up
costs > 0)
cost components. For the instances with set-up costs = 0, the heuristic proposes
solutions in which the final inventory costs (FIC) and the work-in-process costs
(WIPC) are almost balanced, the former being slightly dominant. For the instances
with set-up costs > 0, the heuristic seeks a balance between the fixed set-up costs
(SC) and the final inventory and work-in-process holding costs. Remarkably, this
result is similar to the well-known Economic Order Quantity model for which the
holding costs and fixed ordering costs are the same if the economic order quantity
is ordered. Similarly to the instances with set-up cost = 0, the final inventory costs
dominate the work-in-process costs.
Now we take a look at the fraction of the free capacity, computed as 1 − ρnet , that is
allocated for performing set-ups. From Figure 7 we see that for the instances with
set-up costs = 0, the allocated fraction of free capacity lies in the interval 60–74%,
the average being 65%. As a rule of thumb, it appears that in the case of set-up costs
= 0 about 2/3 of the free capacity should be allocated for set-up times. However,
from the analysis of results of simulation based optimization techniques in Table 7
it appears that it is not only important to select the right level of capacity utilization.
282 P.L.M. Van Nyen et al.
The numerical results in Table 7 indicate that substantial savings can be realised by
adjusting the review periods, but keeping the capacity utilization more or less at the
same level. Indeed, the harmonization of the review periods is of high importance
in the multi-product system under study. Figure 8 shows the case of non-zero set-up
costs for which a wide range of allocation of free capacity is observed (3–69%).
In general, the allocated fraction of free capacity for set-ups is lower than in the
case of zero set-up costs. Moreover, our experiments indicate that the fraction of
allocated capacity decreases when the set-up costs increase. This sound behavior
is due to increases in the review periods because of rising set-up costs. Obviously,
the increases in the review periods lead to lower capacity allocation for set-ups.
Next, we turn our attention to the behavior of the average review periods, order-up-
to levels and capacity utilization when the set-up times are tripled from [30, 60] to
[90, 180]. The results are displayed in Figures 9 and 10. When set-up costs are zero
and the set-up times are tripled, our heuristic increases the review periods and order-
up-to levels with almost the same factor as the set-up times ([2.7, 3.2] in our set of
experiments). In this way, the fraction of free capacity allocated for set-ups remains
almost unchanged. For the instances with set-up costs > 0, a substantial change in
set-up times has virtually no impact on the average review periods and order-up-to
levels. In the main part of the instances, the proportion lies in the interval [0.9, 1.3].
Therefore, the fraction of free capacity used for set-up times increases with the
same factor as the increase in set-up times. From this observation, we conclude that
the cost aspect dominates the capacity aspect in the optimization process when the
set-up costs > 0. Finally, we observe that these results fade when set-up costs are
relatively small, i.e. Lij ∈ [6.67, 13.33]. Especially when the net utilization is high
(0.85), an increase in the set-up times leads to rises in the review periods and the
Fig. 9. Frequency diagram of proportion of average review periods for set-up times [90, 180]
over [30, 60] (instances with set-up costs = 0)
Control of multi-product multi-machine production-inventory systems 283
Fig. 10. Frequency diagram of proportion of review periods for set-up times [90, 180] over
[30, 60] (instances with set-up costs > 0)
corresponding order-up-to levels. Again, this indicates that our optimization tool
works soundly.
From the results in this subsection and the previous subsection, it appears that
behaviour of the optimized control parameters is rather different in the case where
set-up costs are equal to zero and the case where the set-up costs are larger than
zero. In the case that set-up costs are zero, the optimization tool focuses more on
the capacity utilization aspect whereas when the set-up costs are larger than zero,
the tool is mainly concerned with the cost aspect. The lesson that can be learned
from these observations is that it is very important to take into account both cost and
capacity issues when making production and inventory control decisions. Decision
support systems that focus solely on one of these issues are cursed to make errors
that can result in substantial cost increases. Unfortunately it is the case that most
decision support systems do focus either on the capacity aspect or on the cost aspect.
Finally, these observations illustrate the importance of a good knowledge of the cost
structure of the PI system, so that the application of sound management accounting
techniques should be given high priority.
6 Conclusions
In this appendix we propose a lower bound for the total costs in a production-
inventory system with job shop routings and stochastic arrival and processing times.
The lower bound neglects the impact of the variability in the system as well as the
interaction between different products. The lower bound consists of three terms:
1. final inventory costs. The lower bound is based on an inventory model charac-
terized by deterministic demand and zero replenishment lead times;
2. work-in-process costs due to processing times and set-up times (waiting times
are excluded);
3. fixed set-up costs (which are known exactly).
286 P.L.M. Van Nyen et al.
Total 1 1 1 1 1
We formulate a lower bound for the total costs for product i for a given review
period Ri :
M
vij r Ri E [Pij ]
−1
LBT Ci (Ri ) = ci Ri + + E [Lij ] (9)
j=1
E AD i E AD i
(αi∗ ) Ri vi r
2
+
2E AD i
Based on the formula for LBT Ci (Ri ), the review period Ri∗ that minimizes
LBT Ci equals:
2
3
3 2E 2 AD
i ci
∗ 3
Ri = 3 M (10)
4 2 E [P ] v r + (α∗ )2 E AD v r
ij ij i i i
j=1
References
1. Adams J, Balas E, Zawack D (1988) The shifting bottleneck procedure for job-shop
scheduling. Management Science 34: 391–401
2. Altiok T, Shiue GA (2000) Pull-type manufacturing systems with multiple product
types. IIE Transactions 32: 115–124
3. Amin M, Altiok T (1997) Control policies for multi-product multi-stage manufacturing
systems: an experimental approach. International Journal of Production Research 35:
201–223
4. Benjaafar S, Kim JS, Vishwanadham N (2004) On the effect of product variety in
production-inventory systems. Annals of Operations Research 126: 71–101
Control of multi-product multi-machine production-inventory systems 287
28. Vandaele NJ, Lambrecht MR, De Schuyter N, Cremmery R (2000) Spicer Off-Highway
Products Division-Brugge improves its lead-time and scheduling performance. Inter-
faces 30: 83–95
29. Van Nyen PLM, Van Ooijen HPG, Bertrand JWM (2004) Simulation results on the
performance of Albin and Whitt’s estimation method for waiting times in integrated
production-inventory systems. International Journal of Production Economics 90: 237–
249
30. Van Nyen PLM, Bertrand JWM, Van Ooijen HPG, Vandaele NJ (2004) The control
of an integrated multi-product multi-machine production-inventory system. Working
paper, BETA Research School, The Netherlands
31. Van Nyen PLM (2005) The integrated control of production inventory systems. Ph.D.
thesis, Department of Technology Management, Technische Universiteit Eindhoven,
The Netherlands (to appear)
32. Whitt W (1983) The queueing network analyzer. Bell System Technical Journal 62:
2779–2815
33. Whitt W (1994) Towards better multi-class parametric-decomposition approximations
for open queueing networks. Annals of Operations Research 48: 221–224
34. Zipkin P (1986) Models for design and control of stochastic multi-item batch production
systems. Operations Research 34: 91–104
Performance analysis of parallel identical machines
with a generalized shortest queue arrival mechanism
G.J. van Houtum1 , I.J.B.F. Adan2 , J. Wessels2 , and W.H.M. Zijm1,3
1
Faculty of Technology Management, Eindhoven University of Technology, P.O. Box 513,
5600 MB Eindhoven, The Netherlands (e-mail: [email protected])
2
Faculty of Mathematics and Computing Science, Eindhoven University of Technology,
Eindhoven, The Netherlands
3
Faculty of Applied Mathematics, University of Twente, Twente, The Netherlands
1 Introduction
Fig. 1. A flexible assembly system consisting of three parallel insertion machines, on which
three types of PCBs are made
2 Model
This section consists of three subsections. In the first subsection, we describe the
GSQS. In Subsection 2.2 we present a simple condition that is necessary and suf-
292 G.J. van Houtum et al.
ficient for ergodicity. In the last subsection, we present a related condition under
which the GSQS is said to be balanced and we briefly discuss symmetric systems.
The GSQS consists of c ≥ 2 parallel servers serving multiple job types. Each server
has its own queue and is capable of serving a restricted set of job types only. All
service times are exponentially distributed with the same parameter µ > 0. The
arrival stream of each job type is Poisson and an arriving job joins the shortest
queue among all queues capable of serving that job (ties are broken with equal
probabilities). Figure 2 shows a GSQS with c = 2 servers and three job types: type
A, B and C jobs arrive with intensity λA , λB and λC , respectively. The A jobs can
be served by both servers, the B jobs can only be served by server 1, and the C
jobs must be served by server 2.
We introduce the following notations. The servers are numbered from 1, . . . , c
and the set I is defined by I = {1, . . . , c}. The set of all job types is denoted
by J.
The arrival intensity of type j ∈ J jobs is given by λj ≥ 0, and λ = j∈J λj is
the total arrival intensity. For each j ∈ J, I(j) denotes the set of servers that can
serve the jobs of type j. We assume that each job type can be served by at least one
server and each server can handle at least one job type; so, I(j) = ∅ for all j ∈ J,
and ∪j∈J I(j) = I. Without loss of generality, we set µ = 1. Then the average
workload per server is given by ρ = λ/c. Obviously, the requirement ρ < 1 is
necessary for ergodicity.
The behavior of the GSQS is described by a continuous-time Markov process
with states (m1 , . . . , mc ), where mi denotes the length of the queue at server i,
i ∈ I (jobs in service are included). So, the state space is equal to
M = {m | m = (m1 , . . . , mc ) with mi ∈ IN0 for all i ∈ I} . (1)
We assume that j∈J λj 1{i∈I(j)} > 0 for all servers i ∈ I (here, 1{G} is the
indicator function, which is 1 if G is true and 0 otherwise), i.e., that all servers
have a positive potential arrival rate. This guarantees that the Markov process is
Performance analysis of parallel identical machines 293
irreducible. The transition rates are denoted by qm,n . Figure 3 shows the transition
rates for the GSQS in Figure 2.
The relevant performance measures are the mean waiting times W (j) for each
of job type j ∈ J, and the overall mean waiting time W , which is equal to
λj
W = W (j) . (2)
λ
j∈J
2.2 Ergodicity
By studying the job routing, we obtain a simple, necessary condition for the ergod-
each subset J ⊂ J, J = ∅, jobs of type j ∈ J arrive with
icity of the GSQS. For
an intensity equal to j∈J λj and they must be served by the servers ∪j∈J I(j).
This immediately leads to the following lemma.
Lemma 1 The GSQS can only be ergodic if
λj < | ∪j∈J I(j)| for all J ⊂ J, J = ∅. (4)
j∈J
294 G.J. van Houtum et al.
Note that for J = J, this inequality is equivalent to ρ < 1. For the GSQS in
Figure 2, condition (4) states that for ergodicity it is necessary that the inequalities
λB < 1, λC < 1 and λ < 2 (or, equivalently, ρ < 1) are satisfied. It appears that
condition (4) is also sufficient for ergodicity. To show this, we consider so-called
corresponding static systems.
A corresponding static system is a system that is identical to the GSQS, but
with static (random) routing instead of dynamic shortest queue routing. The static
(j)
routing is described by discrete distributions {xi }i∈I(j) , j ∈ J, where for each
(j)
j ∈ J and i ∈ I(j), the variable xi denotes the probability that an arriving job
of type j is sent to server i. Under static routing, it holds for each j ∈ J that the
Poisson stream of arriving type j jobs is split up into Poisson streams with intensities
(j)
xj,i = λj xi , i ∈ I(j), for type j arrivals joining server i. Hence the queues i ∈ I
constitute independent M/M/1 queues with identical mean service times equal to
µ = 1 and arrival intensities j∈J(i) xj,i , where J(i) = {j ∈ J|i ∈ I(j)}. As a
result, we obtain a simple necessary and sufficient condition for the ergodicity of a
corresponding static system, viz.
xj,i < 1 for all i ∈ I.
j∈J(i)
Lemma 2 For a GSQS, there exists a corresponding static system that is ergodic,
if and only if condition (4) is satisfied.
Proof. There exists a corresponding static system that is ergodic if and only if there
exists a nonnegative solution {xj,i }(j,i)∈A , with A = {(j, i) | j ∈ J, i ∈ I and i ∈
I(j)}, of the following equations and inequalities:
xj,i = λj for all j ∈ J, xj,i < 1 for all i ∈ I; (5)
i∈I(j) j∈J(i)
the equalities in (5) guarantee that the solution {xj,i }(j,i)∈A corresponds to discrete
(j)
distributions {xi }i∈I(j) which describe a static routing, and the inequalities in (5)
must be satisfied for ergodicity. It is easily seen that (5) has no solution if condition
(4) is not satisfied.
Now, assume that condition (4) is satisfied. To prove that there exists a nonnega-
tive solution {xj,i }(j,i)∈A of (5), we consider a transportation problem with supply
nodes V̂1 = J ∪ {0}, demand nodes V̂2 = I, and arcs  = A ∪ {(0, i)|i ∈ I}
(supply node 0 denotes an extra type of jobs, which can be served by all servers).
Define the supplies âj by âj = λj for all j ∈ V̂1 \ {0} and â0 = c − λ − c, where
| ∪j∈J I(j)| − j∈J λj
:= min
J ⊂J
| ∪j∈J I(j)|
J =∅
(from (4), it follows that > 0, and â0 ≥ 0 since by taking J = J we obtain
the inequality ≤ (c − λ)/c ). Further, we define the demands b̂i by b̂i = 1 −
for all i ∈ V̂2 ; note that j∈V̂1 âj = i∈V̂2 b̂i . It may be verified that this trans-
portation problem satisfies a necessary and sufficient condition for the existence of
a feasible flow; see Lemma 5.4 of [14] and its proof is based on a transformation
Performance analysis of parallel identical machines 295
It is easily seen that then the solution {xj,i }(j,i)∈A defined by xj,i = x̂j,i for all
(j, i) ∈ A, is a nonnegative solution of (5), which completes the proof.
In situations with many job types shortest queue routing will balance the queue
lengths more than any static routing. So if there is a corresponding static system
that is ergodic, then the GSQS will also be ergodic. Together with Lemma 2, this
informally shows that the following theorem holds.
Theorem 1 The GSQS is ergodic if and only if condition (4) is satisfied.
For a formal proof of this theorem, the reader is referred to Foss and Chernova [6]
or Foley and McDonald [5]. In the latter paper, a generalization of condition (4)
is proved to be necessary and sufficient for the (more general) model with differ-
ent service rates. Their proof also exploits the connection with a corresponding
static system. Foss and Chernova [6] use a fluid approximation approach to derive
necessary conditions for a model with general arrivals and general service times.
It is desirable that the shortest queue routing, as reflected by the sets I(j), balances
the workload among the servers. Formally, we say that a GSQS is balanced if there
exists a corresponding static system for which all queues have the same workload.
(j)
This means that there must exist discrete distributions {xi }i∈I(j) such that for
each server i ∈ I, the arrival intensity j∈J, (j,i)∈A xj,i is equal to λ/c = ρ, where
the xj,i and the set A are defined as before. Such discrete distributions exist if and
only if there exists a nonnegative solution {xj,i }(j,i)∈A of the equations
λ
xj,i = λj for all j ∈ J, xj,i = for all i ∈ I. (6)
i∈I j∈J
c
(j,i)∈A (j,i)∈A
These equations are precisely the equations which must be satisfied by a feasible
flow for the transportation problem with supply nodes V1 = J, demand nodes
V2 = I, arcs A, supplies aj = λj for all j ∈ V1 and demands bi = λ/c for all
i ∈ V2 . Applying the necessary and sufficient condition for the existence of such a
feasible flow (see [14]) leads to the following lemma.
Lemma 3 A GSQS is balanced if and only if
λ
λj ≤ | ∪j∈J I(j)| for all J ⊂ J. (7)
c
j∈J
296 G.J. van Houtum et al.
So, a GSQS is symmetric, if for all subsets I ⊂ I with the same number of servers
|I |, the arrival intensity λ(I ) for the jobs which can be served by precisely the
servers of I , is the same. The GSQS in Figure 2 is symmetric if λB = λC .
For a symmetric GSQS, all queue lengths have the same distribution, which
implies that all servers have equal workloads. For such a system, it follows from
Sparaggis et al. [13], that the shortest queue routing minimizes the total number of
jobs in the system and hence the overall mean waiting time W . In particular, this
implies that the overall mean waiting time in a symmetric GSQS is less than in the
corresponding system consisting of N independent M/M/1 queues with workload
ρ.
In this section we construct two truncation models which are much easier to solve
than the original model. One truncation model produces lower bounds for the mean
waiting times, and the other one upper bounds. At the end of this section we describe
a numerical method for the computation of the mean waiting times within a given,
desired accuracy.
The truncation models exploit the property that the shortest queue routing causes
a drift towards states with equal queue lengths. The state space M of the two models
is obtained by truncating the original state space M around the diagonal, i.e.,
M = {m ∈ M | m = (m1 , . . . , mc ) and mi ≤ min(m) + Ti for all i ∈ I} ,(9)
Performance analysis of parallel identical machines 297
In the lower (upper) bound model, the transitions to states n outside M are redi-
rected to states n with less (more) jobs inside M .
In the lower bound model, the transition in (i) is redirected to n = m − ek −
i∈I ei ∈ M . This means that the departure of a job at a non-empty shortest
queue is accompanied by killing one job at each of the queues i ∈ I , which are
already Ti greater than the shortest queue. The transition in (ii) is redirected to m
itself, i.e., a new job arriving at one of the servers i ∈ I is rejected. The lower
bound model is therefore called the Threshold Killing and Rejection (TKR) model.
In the upper bound model, the transition in (i) is redirected to m itself. This
means that if at least one queue is already Ti greater than the shortest queue, the
finished job in the shortest queue is not allowed to depart, but is served once more;
this is equivalent to saying that the servers atthe shortest queues are blocked.
Transition (ii) is redirected to n = m + ei + k∈Isq ek ∈ M , with Isq = {k ∈
I|mk = min(m)}. This means that an arrival of a new job at one of the queues
which is already Ti greater than the shortest queue, is accompanied by the addition
of one extra job at each of the shortest queues. The upper bound model is therefore
called the Threshold Blocking and Addition (TBA) model. Note that this model
may be non-ergodic while the original model is ergodic. However, the larger the
values of the thresholds Ti the more unlikely this situation. In Figure 4, we show
the redirected transitions in the lower and upper bound model for the GSQS of
Figure 3.
It is intuitively clear that the queues in the TKR model are stochastically smaller
than the queues in the original model. Hence, for each j ∈ J, the TKR model yields
a lower bound for the mean length of the shortest queue among the queues i ∈ I(j),
and thus also for the mean waiting time of type j jobs (cf. (3)). Denote the steady-
state probabilities in the TKR model by πT KR (m1 , . . . , mc ) and let
(j)
WT KR (T̂ ) = min mi πT KR (m1 , . . . , mc ) , j ∈ J.
i∈I(j)
(m1 ,...,mc )∈M
298 G.J. van Houtum et al.
Fig. 4. The redirected transitions in the TKR and TBA model for the GSQS depicted in
Figure 2. For both models, T̂ = (T1 , T2 ) = (3, 3)
(j)
Then we have for each j ∈ J that WT KR (T̂ ) ≤ W (j) , and thus (cf. (2))
λj (j)
WT KR (T̂ ) = WT KR (T̂ )
λ
j∈J
yields a lower bound for the overall mean waiting time W . The lower bounds
(j)
WT KR (T̂ ) monotonically increase as the thresholds T1 , . . . , Tc increase. Similarly
(j)
the TBA model produces monotonically decreasing upper bounds WT BA (T̂ ), j ∈
J, and WT BA (T̂ ). The bounds and the monotonicity properties can be rigorously
proved by using the precedence relation method, see [14]. This method is based on
Markov reward theory and it has been developed in [14, 15].
The truncation models can be solved efficiently by using the matrix-geometric
approach described in [10]. Since the truncation models exploit the property that
shortest queue routing tries to balance the queues, one may expect that the bounds
are tight for already moderate values of the thresholds T1 , . . . , Tc .
We will now formulate a numerical method to determine the mean waiting
times with an absolute accuracy abs . The method repeatedly solves the TKR and
TBA model for increasing threshold vectors T̂ = (T1 , . . . , Tc ). For each vec-
(j) (j)
tor T̂ we use (WT KR (T̂ ) + WT BA (T̂ ))/2 as an approximation for W (j) and
(j) (j)
∆(j) (T̂ ) = (WT BA (T̂ ) − WT KR (T̂ ))/2 as an upper bound for the error; we sim-
ilarly approximate W by (WT KR (T̂ ) + WT BA (T̂ ))/2 where the error is at most
∆(T̂ ) = (WT BA (T̂ ) − WT KR (T̂ ))/2. The approximations and error bounds are
set equal to ∞ if the TBA model is not ergodic (which may be the case for small
thresholds). The computation procedure stops when all error bounds are less than
or equal to abs ; otherwise at least one of the thresholds is increased by 1 and new
approximations are computed. The decision to increase a threshold Ti is based on
the rate of redirections rrd (i). This is explained in the next paragraph.
Performance analysis of parallel identical machines 299
The variable rrd (i), i ∈ I, denotes the rate at which redirections occur in the
boundary states m = (m1 , . . . , mc ) with mi = min(m) + Ti of the truncated state
space. If for given T̂ only the TKR model is ergodic, then rrd (i) denotes the rate for
the TKR model, otherwise rrd (i) denotes the sum of the rate for the TKR and TBA
model. The rates rrd (i) can be computed directly from the steady-state distributions
of the bound models. The higher the rate rrd (i), the higher the expected impact
of increasing Ti . The computation procedure increases all thresholds Ti for which
rrd (i) = maxk∈I rrd (k). The numerical method is summarized below.
Algorithm (to determine the mean waiting times for the GSQS)
Input: The data of an ergodic instance of the GSQS, i.e.,
c, J, I(j) for all j ∈ J, and λj for all j ∈ J;
the absolute accuracy abs ;
the initial threshold vector T̂ = (T1 , . . . , Tc ).
(j) (j)
Step 1. Determine WT KR (T̂ ), WT BA (T̂ ) and ∆(j) (T̂ ) for all j ∈ J,
and WT KR (T̂ ), WT BA (T̂ ) and ∆(T̂ ),
and rrd (i) for all i ∈ I.
Step 2. If ∆(j) (T̂ ) > abs for some j ∈ J or ∆(T̂ ) > abs ,
then Ti := Ti + 1 for all i ∈ I with rrd (i) = maxk∈I rrd (k),
and return to Step 1.
(j) (j)
Step 3. W (j) = (WT KR (T̂ ) + WT BA (T̂ ))/2 for all j ∈ J,
and W = (WT KR (T̂ ) + WT BA (T̂ ))/2.
Note that for a symmetric GSQS it is natural to start with a threshold vector T̂
with equal components. Then in each iteration all rates rrd (i) will be equal, and
hence each Ti will be increased by 1. So the components of T̂ will remain equal.
fraction of semi-common jobs (and no common jobs). In both cases the average
number of servers capable of serving an arbitrary job is the same. In Subsection
4.3 we evaluate a series of balanced, asymmetric systems. We investigate how the
mean waiting times deteriorate due to the asymmetry. Finally, in Subsection 4.4,
the main conclusions are summarized.
c=2 c=3
p W (1) W (c+1) W rr(p) W (1) W (c+1) W rr(p)
0.0 9.00 4.26 9.00 0.0 % 9.00 2.69 9.00 0.0 %
0.1 6.80 4.36 6.56 54.0 % 6.07 2.82 5.75 54.1 %
0.2 6.04 4.40 5.72 72.6 % 5.06 2.88 4.63 72.7 %
0.3 5.66 4.43 5.29 82.0 % 4.56 2.91 4.06 82.0 %
0.4 5.43 4.44 5.04 87.6 % 4.25 2.93 3.72 87.7 %
0.5 5.28 4.45 4.86 91.4 % 4.05 2.95 3.50 91.4 %
0.6 5.17 4.46 4.74 94.1 % 3.90 2.96 3.34 94.1 %
0.7 5.09 4.46 4.65 96.1 % 3.79 2.97 3.21 96.1 %
0.8 5.02 4.47 4.58 97.7 % 3.71 2.97 3.12 97.7 %
0.9 4.97 4.47 4.52 99.0 % 3.64 2.98 3.04 99.0 %
1.0 4.93 4.48 4.48 100.0 % 3.58 2.98 2.98 100.0 %
accurately in order to obtain sufficiently accurate estimates for rr(p). The results
in Table 2 show that for each combination of p and c, the values for W for varying
workloads ρ are not that far away from the values for WSSQS ; in particular, the
absolute differences are small for small workloads ρ and the relative differences are
small for high workloads ρ. The results also suggest that the rr(p) is insensitive to
the number of servers c. However, rr(p) strongly depends on ρ; it is rather small
for low workloads and large for high workloads (it seems that rr(p) ↑ 1 as ρ ↑ 1).
In Subsection 4.1 we distinguished two job types only, specialist and common
jobs. For GSQSs with more than two servers, one may also have jobs in between,
302 G.J. van Houtum et al.
c=2 c=3
p ρ WM/M/1 W WSSQS rr(p) W WSSQS rr(p)
0.25 0.2 0.25 0.19 0.07 32.1 % 0.18 0.02 30.8 %
0.4 0.67 0.51 0.26 39.6 % 0.46 0.13 38.6 %
0.6 1.50 1.10 0.68 49.2 % 0.97 0.42 48.9 %
0.8 4.00 2.67 1.96 64.8 % 2.24 1.29 64.8 %
0.9 9.00 5.47 4.47 77.9 % 4.31 2.98 78.0 %
0.95 19.00 10.69 9.49 87.3 % 7.94 6.33 87.4 %
0.98 49.00 25.86 24.49 94.4 % 18.17 16.35 94.4 %
0.50 0.2 0.25 0.14 0.07 58.7 % 0.12 0.02 57.2 %
0.4 0.67 0.40 0.26 66.4 % 0.32 0.13 65.5 %
0.6 1.50 0.89 0.68 74.5 % 0.70 0.42 74.3 %
0.8 4.00 2.27 1.96 84.7 % 1.70 1.29 84.8 %
0.9 9.00 4.86 4.47 91.4 % 3.50 2.98 91.4 %
0.95 19.00 9.93 9.49 95.4 % 6.92 6.33 95.4 %
0.98 49.00 24.97 24.49 98.1 % 16.98 16.35 98.1 %
i.e., jobs that can be served by two or more, but not all servers. In this subsection
we investigate which job types lead to the largest reduction of W : common or
semi-common jobs?
We consider a GSQS with c = 3 servers and a total arrival rate λ = 3ρ
with ρ ∈ (0, 1). The following two cases are distinguished for the detailed arrival
streams. For case I, we copy the situation in Subsection 4.1. In this case there are
4 job types. The type 4 jobs are common jobs; they arrive with intensity λ4 = pλ
with p ∈ [0, 0.5] (the reason why p may not exceed 0.5 follows below). Type j jobs,
j = 1, 2, 3 are specialist jobs which only can be served by server j; they arrive with
intensity λj = (1 − p)λ/3. So the mean number of servers capable of serving an
arbitrary job is equal to 1 + 2p. In case II we have 6 job types. The type j jobs,
j = 1, 2, 3, are again specialist jobs which can only be served by server j. The type
4, 5 and 6 jobs are semi-common jobs; the type 4 jobs can be served by the servers
1 and 2, the type 5 jobs by 1 and 3, and the type 6 jobs by 2 and 3. To guarantee
that the mean number of servers capable of serving an arbitrary job remains the
same (i.e., equal to 1 + 2p), the arrival intensity λj is set equal to λj = 2pλ/3 for
j = 4, 5, 6 and λj = (1 − 2p)λ/3 for j = 1, 2, 3 (to avoid negative intensities, p
must be less than or equal to 0.5).
Table 3 lists the overall mean waiting time W for different values of p and ρ.
The results for case I are copied from Table 2. We can conclude that the absolute
difference between the mean waiting time W in case I and II is rather small in each
situation. This suggests that W is mainly determined by the mean number of servers
capable of serving an arbitrary job; it does not matter whether this mean number is
realized by common or by (twice as many) semi-common jobs. Nevertheless, the
results in Table 3 also show that in each situation case II yields a smaller W than
Performance analysis of parallel identical machines 303
W Diff. (I−II)
p ρ Case I Case II Abs. Rel.
0.25 0.2 0.18 0.14 0.04 23.4 %
0.4 0.46 0.38 0.08 18.1 %
0.6 0.97 0.83 0.15 14.9 %
0.8 2.24 1.97 0.27 11.9 %
0.9 4.31 3.92 0.38 8.9 %
0.95 7.94 7.46 0.48 6.0 %
0.98 18.17 17.62 0.55 3.0 %
0.50 0.2 0.12 0.05 0.07 55.1 %
0.4 0.32 0.22 0.10 32.5 %
0.6 0.70 0.56 0.14 20.1 %
0.8 1.70 1.51 0.19 11.2 %
0.9 3.50 3.27 0.22 6.4 %
0.95 6.92 6.67 0.25 3.6 %
0.98 16.98 16.72 0.26 1.5 %
case I. This may be explained as follows. Let us consider the situation with p = 0.5.
In case I, λ1 = λ2 = λ3 = λ/6 and λ4 = λ/2. Hence, for each group of 6 arriving
jobs, on average 4 jobs join the shortest queue, 1 job joins the shortest but one
queue, and 1 job joins the longest queue. In case II, however, λ1 = λ2 = λ3 = 0
and λ4 = λ5 = λ6 = λ/3. Thus for each group of 6 arriving jobs, on average 4
jobs join the shortest queue and 2 jobs joins the shortest but one queue. So in case
II the balancing of queues will be slightly stronger, and thus W will be slightly
smaller.
In this subsection we study the GSQS with c = 2 servers and three job types as
depicted in Figure 2. The parameters are chosen as follows: ρ = 0.9, λ = 2ρ = 1.8,
λA = λ/2 = 0.9, λB = p̂λ/2 = 0.9p̂, λC = (1 − p̂)λ/2 = 0.9(1 − p̂) where
p̂ ∈ [0, 0.5]. So one half of the jobs are common (type A) jobs and the other half
are specialist (type B and C) jobs. But the specialist jobs are not equally divided
over the servers. The fraction p̂ of specialist jobs which must be served by server
1 (i.e., the type B jobs) is less than or equal to the fraction 1 − p̂ of specialist jobs
which must be served by server 2 (i.e. the type C jobs). Only for p̂ = 0.5 we have
a symmetric system. For all p̂ ∈ [0, 0.5) we have an asymmetric, but balanced
system; a static system with equal workloads for both servers is obtained when a
fraction 1 − p̂ of the type A jobs is sent to server 1 and a fraction p̂ to server 2.
Table 4 shows the mean waiting times W (A) , W (B) , W (C) for each job type and
the overall mean waiting time W for p̂ = 0, 0.1, . . . , 0.5. These waiting times have
again been computed with an absolute accuracy abs = 0.005. In the last column of
304 G.J. van Houtum et al.
Table 4 we list the realized reduction rr(p̂) defined by (10), where WM/M/1 = 9
and WSSQS = 4.475 for ρ = 0.9. The results in Table 4 show that W (A) is fairly
constant for all values of p̂. As expected, W (B) decreases and W (C) increases as
p̂ decreases. A striking observation is that W (C) sharply increases for p̂ close to
0; and thus also W = (W (A) + p̂W (B) + (1 − p̂)W (C) )/2. For p̂ = 0 we have
λA = λC = 0.9 and λB = 0, and the overall mean waiting time W is equal to
8.66. This is close to WM/M/1 = 9, which is realized when all type A jobs would
be sent to server 1.
4.4 Conclusion
The main conclusion from the numerical experiments is that the overall mean wait-
ing time may already be reduced significantly by creating a little bit of (semi-)
common work. Furthermore, this reduction is mainly determined by the amount
of overlap, i.e., the mean number of servers capable of handling an arbitrary job.
Finally, the beneficial effect of (semi-)common jobs may vanish for highly asym-
metric situations.
References
1. Adan IJBF, Van Houtum GJ, Van der Wal J (1994) Upper and lower bounds for the
waiting time in the symmetric shortest queue system. Annals of Operations Research
48: 197–217
2. Adan IJBF, Wessels J, Zijm WHM (1989) Queueing analysis in a flexible assembly
system with a job-dependent parallel structure. In: Operations Research Proceedings,
pp 551–558. Springer, Berlin Heidelberg New York
3. Adan IJBF, Wessels J, Zijm, WHM (1990) Analysis of the symmetric shortest queue
problem. Stochastic Models 6: 691–713
4. Ahuja RK, Magnanti TL, Orlin JB (1993) Network flows: theory, algorithms, and
applications. Prentice-Hall, Englewood Cliffs, NJ
5. Foley RD, McDonald DR (2000) Join the shortest queue: Stability and exact asymp-
totics. The Annals of Applied Probability (to appear)
6. Foss S, Chernova N (1998) On the stability of a partially accessible multi-station queue
with state-dependent routing. Queueing Systems 29: 55–73
Performance analysis of parallel identical machines 305
7. Green L (1985) A queueing system with general-use and limited-use servers. Operations
Research 33: 168–182
8. Hassin R, Haviv M (1994) Equilibrium strategies and the value of information in a two
line queueing system with threshold jockeying. Stochastic Models 10: 415–435
9. Latouche G, Ramaswami V (1993) A logarithmic reduction algorithm for quasi-birth-
death processes. Journal of Applied Probability 30: 650–674
10. Neuts MF (1981) Matrix-geometric solutions in stochastic models. Johns Hopkins
University Press, Baltimore
11. Roque DR (1980) A note on “Queueing models with lane selection”. Operations Re-
search 28: 419–420
12. Schwartz BL (1974) Queueing models with lane selection: a new class of problems.
Operations Research 22: 331–339
13. Sparaggis PD, Cassandras CG, Towsley D (1993) Optimal control of multiclass par-
allel service systems with and without state information. In Proceedings of the 32nd
Conference on Decision and Control, San Antonio, pp 1686–1691
14. Van Houtum GJ (1995) New approaches for multi-dimensional queueing systems. Ph.D.
Thesis, Eindhoven University of Technology, Eindhoven
15. Van Houtum GJ, Zijm WHM, Adan IJBF, Wessels J (1998) Bounds for performance
characteristics: A systematic approach via cost structures. Stochastic Models 14: 205–
224 (Special issue in honor of M.F. Neuts)
16. Zijm WHM (1991) Operational control of automated PCB assembly lines. In: Fandel
G, Zaepfel G (eds) Modern production concepts: theory and applications, pp 146–164.
Springer, Berlin Heidelberg New York
A review and comparison
of hybrid and pull-type production control strategies
John Geraghty1 and Cathal Heavey2
1
School of Mechanical and Manufacturing Engineering, Dublin City University, Glasnevin,
Dublin 9, Ireland (e-mail: [email protected])
2
Department of Manufacturing and Operations Engineering, University of Limerick,
Limerick, Ireland (e-mail: [email protected])
1 Introduction
The Kanban Control Strategy, KCS, developed by Toyota allows part flow in a Just-
In-Time, JIT, line to be controlled by basing production authorisations on end-item
demands. KCS is often referred to as a ‘Pull’ production control strategy since part
demands travel upstream and pull products down the line by authorising production
based on the presence of Kanban cards which are limited in number and circulated
between production stages. KCS has been the focus of considerable research effort
since the early 1980’s. In particular, optimising the number and distribution of
A review and comparison of hybrid and pull-type production control strategies 309
Once this level of inventory has been achieved, inventory may not enter the system
until a demand event removes a corresponding amount of inventory from the line.
With only one parameter to optimise, i.e. the WIP Cap, CONWIP is very simple to
implement and maintain. The main reason why CONWIP lines outperform KCS
lines is that demand information is instantly communicated to the initial stage and
the release rate is adjusted to match the demand rate. In a KCS line, as has been stated
earlier, demand information has to travel upstream from the end-item inventory
point to the initial stage. The longer the line and the more delays encountered
at individual production stages (e.g. processing time, breakdown/repair time, set-
up time etc.) the longer the information delay encountered. A disadvantage of
CONWIP is that inventory levels are not controlled at the individual stages, which
can result in high inventory levels building up in front of bottleneck stages.
Chang and Yih [6] introduced a Pull PCS named Generic Kanban System (GKS)
applicable to dynamic, multi-product, non-repetitive manufacturing environments.
KCS requires inventories of semi-finished products of each product type to be
maintained at each production stage. In multi-product environments the amount of
semi-finished inventory maintained in the line could be prohibitively large [6]. GKS
operates by providing a fixed number of Kanbans at each workstation that can be
acquired by any part. A part/job can only enter the system if it acquires a Kanban
from each of the workstations in the system. GKS reduces to CONWIP if an equal
number of Kanbans are distributed to all workstations. Comparison with a push-type
production control strategy was favourable with GKS shown to be less susceptible to
the position of the bottleneck. GKS outperformed KCS in terms of WIP required to
achieve a desired Cycle Time. Comparison with CONWIP was favourable and GKS
was shown to be more flexible in that by manipulating the number of Kanbans at
each workstation the performance of GKS could be improved beyond that achieved
by CONWIP. Chang and Yih [5] presented a simulated annealing algorithm for
determining the optimal Kanban distribution for a GKS line.
In order to overcome the disadvantages of loose coordination between pro-
duction stages in a CONWIP line Bonvik and Gershwin [3] and Bonvik et al.
[2] proposed an alternative strategy, hybrid Kanban-CONWIP. In hybrid Kanban-
CONWIP, as in CONWIP, an overall cap is placed on the amount of inventory
allowed in the production system. In addition, inventory is controlled using Kan-
bans in all stages except the last stage. CONWIP can be considered as special
case of hybrid Kanban-CONWIP in which there is an infinite number of Kanbans
distributed to each production stage [2]. A comparison of KCS, minimal blocking
KCS, BSCS, CONWIP and hybrid Kanban-CONWIP was presented in [2]. The
different PCS were compared in a four-stage tandem production line using simula-
tion. Each of the PCS were compared using constant demand and demand that had a
stepped increase/decrease. It was found that the hybrid Kanban-CONWIP strategy
decreased inventories by 10% to 20% over KCS while main taining the same service
levels (percentage of demands instantaneously matched with a finished product).
The performance of basestock and CONWIP strategies fell between those of KCS
and hybrid Kanban-CONWIP.
Two papers that generalize hybrid Kanban-CONWIP are [14] and [13]. These
papers propose a generic pull model that, as well as encapsulating the three basic
A review and comparison of hybrid and pull-type production control strategies 311
pull control strategies, KCS, CONWIP and BSCS, also allows customized pull
control strategies to be developed. Simulation and an evolutionary algorithm were
used to study the generic model. Details of the evolutionary algorithm are given in
[14] while results on extensive experimentation on the effect of factors (i.e., line im-
balance, machine reliability) on the proposed generic pull model are given in [13].
Gaury and Kleijnen [12] noted that Operations Research has traditionally con-
centrated on optimisation whereas practitioners find the robustness of a proposed
solution more important. A methodology was presented in [12] that was a stagewise
combination of four techniques: (i) simulation, (ii) optimization, (iii) risk or un-
certainty analysis, and (iv) bootstrapping. Gaury and Kleijnen [12] illustrated their
methodology through a production-control study for the four-stage, single product
production line utilised by [2]. Robustness was defined in [12] as the capability
to maintain short-term service, in a variety of environments; i.e. the probability of
the short-term fill-rate (service level) remaining within a pre-specified range. Be-
sides satisfying this probabilistic constraint, the system minimised expected long-
term WIP. Four systems were compared in [12], namely Kanban, CONWIP, hybrid
Kanban-CONWIP, and Generic. The optimal parameters found in [2] were used
for KCS, CONWIP and hybrid Kanban-CONWIP. Gaury and Kleijnen [12] used
a Genetic Algorithm to determine the optimal parameters for the Generic pull sys-
tem. For the risk analysis step, seventeen inputs were considered; the mean and
variance of the processing time for each of the four production stages, mean time
between failures and mean time to repair per production stage, and the demand rate.
The inputs were varied over a range of ±5% around their base values. Gaury and
Kleijnen [12] concluded that in this particular example, hybrid Kanban-CONWIP
was best when risk was not ignored; otherwise Generic was best and therefore, risk
considerations can influence the selection of a PCS.
Each of the pull-type production control strategies discussed above, with the
exception of GKCS, have one important advantage over KCS that ensures that they
are more readily applicable to non-repetitive manufacturing environments. That
advantage stems from the manner in which demand information is communicated
in comparison to KCS. In KCS, demand information is not communicated directly to
production stages that release parts/jobs into the system. Rather it is communicated
sequentially up the line from the finished goods buffer as withdrawals are made by
customer demands. This communication delay means that the pace of the production
line is not adjusted automatically to account for changes in the demand rate. The
arrival of demand information to the initial stages in a GKCS line might be delayed
if the demand cards at a production stage in the line are not instantaneously matched
with Kanban cards. BSCS, EKCS, CONWIP, GKS and hybrid Kanban-CONWIP
all, however, communicate the demand information instantaneously to the initial
stages allowing the release rate to be paced to the actual demand rate. For instance,
Bonvik et al. [2] showed that if the demand rate decreases unexpectedly the impact
on a CONWIP strategy and hybrid Kanban-CONWIP strategy would be for the
finished-goods buffer to increase toward the WIP Cap with all intermediate buffers
tending toward empty. The impact, however, on a KCS line would be that all
the intermediate buffers would increase toward their maximum permissible limits.
312 J. Geraghty and C. Heavey
Therefore, the KCS line would have semi-finished inventory distributed throughout
the line.
Hybrid control strategies can be classified into two categories: vertically integrated
hybrid systems (VIHS) or horizontally integrated hybrid systems (HIHS) [8]. VIHS
consist of two levels, usually an upper level push-type PCS and a lower level pull-
type PCS. For example, Synchro MRP utilises MRP for long range planning and
KCS for shop floor execution [17]. The main disadvantage of VIHS is that MRP
calculations must be performed for each stage in the production system. This makes
VIHS complex to implement and maintain and accounts for their relative lack of
use in industry [20]. HIHS consist of one level where some production stages are
controlled by push-type PCS and other stages by pull-type PCS. Only HIHS are
considered in the discussion that follows.
Hodgson and Wang [20, 21] developed a Markov Decision Process (MDP)
model for HIHS. The model was solved using both dynamic programming and
simulation for several production strategies, including pure push and pure pull
production strategies and strategies based on the integration of push and pull control.
In this push/pull integration strategy each individual stage may push or pull. This
type of control strategy is denoted as Hybrid Push/Pull in [20, 21]. Initially in [20],
the research was applied to a four-stage semi-continuous production iron and steel
works (see Fig. 1), with the first two stages in parallel and the remaining stages as
serial production stages. In order to simplify the analysis the model assumes that
the production process is a discrete time process and that demand per period and
the amount of inventory are both integer multiples of a unit size. The research was
later extended to a five-stage production system [21]. For both the four and five
stage production systems, a strategy where production stages 1 and 2 (P1 and P2
in Fig. 1) push and all other stages pull was demonstrated to result in the lowest
average gain (average system cost). Hodgson and Wang [21] stated that they had
observed similar results for an eight-stage system and concluded that this strategy
would be the optimal hybrid integration strategy for a J-stage system. Subsequent
papers that use the model in [20, 21] or extensions of it are [11], [28] and [35].
Deleersnyder et al. [11] considered that the complexity of the control structure
required for the successful implementation of Synchro MRP resulted in it being
largely ignored by industry. Synchro MRP requires MRP control to be linked into
every stage in the production line while utilising local kanban control to authorise
production at each stage. Deleersnyder et al. [11] developed a hybrid production
strategy that limited the number of stages into which MRP type information is
added in order to reduce the complexity of the hybrid strategy in comparison to
Synchro MRP, while realizing the benefits of integrating push and pull type control
strategies. The model developed in [11] is similar to that presented in [20, 21] and
comparable results were obtained for a serial production line.
Pandey and Khokhajaikiat [28] extended the model in [20, 21] to allow for
the inclusion of raw material constraints at each stage. The modified model also
allowed for a stage to require more than one item of inventory and/or more than one
A review and comparison of hybrid and pull-type production control strategies 313
Fig. 1. Parallel/Serial four stage production system modelled by Hodgson and Wang [20]
item of raw material to produce a part. Pandey and Khokhajaikiat [28] presented
results from two sets of experiments. In the first set they modelled a four-stage
parallel/serial production line similar to the system shown in Figure 1. The initial
production stages (P1 and P2 in Fig. 1) operated under raw material availability
constraints, had different order purchasing and delivery distributions but had iden-
tical production unreliability. Sixteen integration strategies were considered. In the
second experimental set the authors applied the raw material availability constraint
to all stages of the production line. The authors concluded that the hybrid strategy
in which the initial stages (P1 and P2 ) operate under push control and the remaining
stages operate under pull control is the best strategy when raw material constraints
apply only to the initial stages. When the raw material availability constraint is
applied to all stages the push strategy becomes the optimal control strategy. For
systems with large variability in demand none of the strategies dominated.
Wang and Xu [35] presented an approach that facilitated the evaluation of
a wide range of topologies that utilize hybrid push/pull. They used a structure
model to describe a manufacturing system’s topology. Their methodology was
used to investigate four 45-stage manufacturing systems: (i) A single-material serial
processing system; (ii) A multi-material serial processing system; (iii) A multi-part
processing and assembly system, and (iv) A multi-part multi-component processing
and assembly system. Wang and Xu [35] compared pure pull and push strategies
against the optimal hybrid strategy found in [20, 21], where the initial stages push
and all other stages pull. Their results suggest that the optimal hybrid strategy
out-performs pure push or pull strategies.
Other models that implement hybrid push/pull control strategies similar to [20,
21] have been developed. Takahashi et al. [32] defined push/pull integration as a
system in which there is a single junction point between push stages and pull stages.
In Takahashi et al. [32] a model was presented to evaluate this control strategy. Two
314 J. Geraghty and C. Heavey
subsequent papers, [33] and [34] further developed and experimented with this
model. Hirakawa et al. [19] and Hirakawa [18] developed a mathematical model
for a hybrid push/pull control strategy that allows each production stage to switch
between push and pull control depending on whether demand can be forecasted
reliably or not. Cochran and Kim [8] presents a HIHS with a movable junction point
between a push sub-system and a pull sub-system. The control strategy presented
had three decision variables: (i) the junction point, i.e., the last push stage in the
HIHS; (ii) the safety stock level at the junction point; (iii) the number of kanbans for
each stage in the pull sub-system. Simulation combined with simulated annealing
was used to find the optimal decision variables for the control strategy.
Several comparisons of Pull-PCS have been reported in the literature, for example
[2, 12, 13, 25]. There has also been several comparisons between HIHS and KCS,
for example [8, 27, 32, 33, 34, 20, 21, 11, 28, 35]. Comparisons between HIHS and
other Pull-Type PCS are rare in the literature. Orth and Coskunoglu [27] included
CONWIP, in addition to KCS, in the comparison analysis. In a previous paper [15]
we demonstrated that the optimal HIHS selected by Hodgson and Wang [20, 21]
where initial stages employ push control and all other stages employ pull control, is
equivalent to a Pull-Type PCS, namely hybrid Kanban-CONWIP [3, 2]. As well as
considering several alternative integration strategies, Hodgson and Wang [20, 21]
also included a Push-Type and a Pull-Type PCS in their analysis, which they referred
to as ‘Pure Push’ and ‘Pure Pull’ PCS. After examining the equations used in [20, 21]
to model the ‘Pure Push’ PCS we felt that there were similarities to the control
structure implemented by the Pull-Type PCS know as EKCS [9, 10]. Therefore in
this section we explore these comparisons. The notation used in the remainder of
this paper is shown in Table 1.
In Hodgson and Wang’s ‘Pure-Push’ PCS, production is authorised when (i) suf-
ficient space exists in the output buffer of the stage, (ii) sufficient inventory exists in
the input buffer of the stage, (iii) sufficient production capacity exists at the stage,
and (iv) downstream inventory levels have decreased below forecasted require-
ments necessary to meet expected demand. The MDP model presented in [20, 21]
required the evaluation of two equations (i.e. the production trigger, Aj (n), and
the production objective, P Oj (n) in order to determine production authorisations
for a stage in period n,). For the purposes of the discussion presented here we have
combined these equations to form a single equation for the number of production
authorisations, P Aj (n), available to a stage in period n. This was achieved without
making any simplifying assumptions. P Aj (n) for a system controlled by Hodgson
and Wang’s ‘Pure-Push’ PCS can be modelled by Eq. (1) where 1 ≤ j ≤ J − 1
and by Eq. (2) for the final production stage.
A review and comparison of hybrid and pull-type production control strategies 315
Notation Description
P Aj (n) = min Ijmax − Ij (n − 1), max Pjmin , SS + (N Sj + 1) × D (n)
⎛ ⎞⎤ ⎫
J ⎬
−⎝ Ii (n − 1) − cj (n − 1)⎠⎦, {Bj (n − 1)} , Pjmax , (1)
⎭
i=j
∀j ≤ J − 1
P AJ (n) = min IJmax − IJ (n − 1) + D (n) , max PJmin , SS + D (n)
−IJ (n − 1)], {BJ (n − 1)} , PJmax } (2)
It is possible for the term IJ (n − 1) to become negative. This occurs in the event
of a shortage in period n − 1, i.e. a failure to satisfy demand. Therefore the term
316 J. Geraghty and C. Heavey
IJ (n − 1) is not only used to record the inventory in the finished goods buffer in
period n − 1 but also the backlog in period n − 1. If a backlog occurs, i.e. IJ (n − 1)
is negative, Eq. (2) would effectively result in a temporary increase of the maximum
capacity of the finished goods inventory buffer. Equation (3) models the number
of production authorisations available to the final stage where shortages are not
permitted to temporarily increase the maximum capacity of the finished goods
inventory buffer.
P AJ (n) = min IJmax − max [0, IJ (n − 1)] + D (n) , max PJmin , SS
+D (n) − IJ (n − 1)], {BJ (n − 1)} , PJmax } (3)
Notice that Hodgson and Wang’s model of Push control includes a limit on inventory
in the output buffer of a stage, Ijmax . Since a production stage does not become
aware of a change in state of the buffer until the subsequent production period, n+1,
and with the exception of the final stage does not attempt to predict the removal
of inventory from the output buffer by immediate successors, this limit behaves
similarly to Kanban control. However, the Push strategy modelled by [20, 21] is
not equivalent to KCS since each stage has information J about the status of line
downstream from its output buffer, through the term i=j Ii (n − 1) − cj (n − 1) in
Eq. (1) above. Since only a demand event can change the state of the downstream
section of the line, in terms of total WIP, this is equivalent to demand cards being
passed to each stage in the line. Therefore, it would appear that there is some
similarity between the ‘Pure-Push’PCS modelled by [20, 21] and EKCS.
In an EKCS system, a production stage has authorisation to produce a part
when: (i) inventory is available in its buffer; (ii) a Kanban card is available and (iii)
a demand card is available. The number of demand cards available to a production
stage in period n is given by Eq. (4). In period n the number of production au-
thorisations at stage j, P Aj (n), is given by Eq. (5). For an EKCS system where
the introduction of temporary Kanbans in the event of shortages is not permitted
P Aj (n) is modelled by Eq. (6) for the final production stage.
DCj (n) = DCj (n − 1) − Pj (n − 1) + d (n) , ∀j ≤ J (4)
P Aj (n) = min Kj − Ij (n − 1) , max Pjmin , DCj (n) ,
{Bj (n − 1)} , Pjmax , ∀j ≤ J − 1 (5)
P AJ (n) = min KJ − max [0, IJ (n − 1)] , max PJmin , DCJ (n) ,
{BJ (n − 1)} , PJmax } (6)
Let us assume a serial production line with J stages. The state transition equa-
tions for the inventory levels of the buffers, for either model, can be determined
from Eqs. (7) and (8):
From examining the equations for the two models, for both dynamic and static
Kanban distributions, it is clear that in order for the PCS to be equivalent three
conditions must be satisfied: (i) the inventory level of a buffer in the ‘Pure-Push’
A review and comparison of hybrid and pull-type production control strategies 317
model must equate to the inventory level of the same buffer in the EKCS model in
a given production period n. (ii) the Kanban distribution in EKCS must be equal
to the buffer capacity limits in the ‘Pure-Push’ model.
Kj = Ijmax ∀j ≤J −1 (9)
KJ = IJmax + D (n) (10)
and (iii) the following two equalities must hold:
DCJ (n) = SS + D (n) − IJ (n − 1) (11)
⎡ ⎛ ⎞⎤
J
DCj (n) = ⎣SS+ (N Sj +1) × D (n) − ⎝ Ii (n−1)−cj (n−1)⎠⎦
i=j
(12)
∀j ≤ J − 1
Substituting Eq. (11) into Eq. (4) yields:
SS + D (n) − IJ (n − 1) = SS + D (n) − IJ (n − 2)
−PJ (n − 1) + d (n) (13)
Re-writing Eq. (13) yields:
IJ (n − 1) = IJ (n − 2) + PJ (n − 1) − d (n) (14)
Equation (14) is not equivalent to Eq. (8) and therefore the two PCS are not equiv-
alent. The primary difference between the two PCS stems from the method in
which demand information is communicated to each stage. In EKCS demand in-
formation is instantly communicated to all stages. In the ‘Pure-Push’ PCS the
communication of demand information is delayed by one period. In fact, the
‘Pure-Push’ PCS described by [20, 21] could more accurately be described as
a Vertically Integrated Hybrid System, in which each production stage develops
a forecast of production requirements
J for the production period through the term
SS + (N Sj + 1) × D(n) − ( i=j Ii (n − 1) − cj (n − 1)) and utilises kanbans to
implement the production plan on the shop-floor plan. It would, therefore, be more
accurate to refer to this PCS as Synchro MRP than ‘Pure-Push’.
However, it is worth noting that the two PCS can be made equivalent if the
demand information is communicated instantly in both PCS or delayed for one
period in both PCS. For instance, communication of demand information in the
‘Pure-Push’ PCS can be made instantaneous by adjusting Eqs. (1), (2) and (3) by
including a component to adjust the downstream inventory levels by the demand
quantity d(n) as shown by Eqs. (15), (16) and (17) below.
P Aj (n) = min Ijmax − Ij (n − 1), max Pjmin , SS + (N Sj + 1) × D (n)
⎛ ⎞⎤ ⎫
J ⎬
−⎝ Ii (n−1)−cj (n−1)−d (n)⎠⎦, {Bj (n−1)} , Pjmax ,
⎭
i=j
∀j ≤ J − 1 (15)
P AJ (n) = min − IJ (n − 1) + D (n) , max PJmin , SS + D (n)
IJmax
318 J. Geraghty and C. Heavey
∀j ≤ J − 1 (19)
Substituting Eq. (18) into Eq. (4) yields:
SS + D (n) − (IJ (n − 1) − d (n)) = SS + D (n) − (IJ (n − 1) − d (n − 1))
−PJ (n − 1) + d (n) (20)
Re-writing Eq. (20) yields:
IJ (n − 1) = IJ (n − 2) + PJ (n − 1) − d (n − 1) (21)
Equation (21) is equivalent to Eq. (8) and therefore requires no further proof.
Substituting Eq. (19) into Eq. (4) yields:
⎡ ⎛ ⎞⎤
J
⎣SS + (N Sj + 1) × D (n) − ⎝ Ii (n − 1) − cj (n − 1) − d (n)⎠⎦ =
i=j
⎡ ⎛ ⎞⎤
J
⎣SS + (N Sj + 1) × D (n) − ⎝ Ii (n − 2) − cj (n − 2) − d (n − 1)⎠⎦
i=j
+Pj (n − 1) − d (n − 1) (23)
We can use the state transition equations (i.e. Eqs. (7) and (8)) to prove Eq. (23) as
shown below. Note that in a serial line the terms cj (n−1) and cj (n−2) are both zero.
A review and comparison of hybrid and pull-type production control strategies 319
Ij (n − 1) = Ij (n − 2) + Pj (n − 1) − Pj+1 (n − 1)
Ij+1 (n − 1) = Ij+1 (n − 2) + Pj+1 (n − 1) − Pj+2 (n − 1)
Ij+2 (n − 1) = Ij+2 (n − 2) + Pj+2 (n − 1) − Pj+3 (n − 1)
.. . .
. = .. + Pj+3 (n − 1) − ..
.. .. .. .. .. .. ..
. . . . . . .
.. . ..
. = .. +. −P (n − 1)
J−1
IJ−1 (n − 1) = IJ−1 (n − 2) + PJ−1 (n − 1) − PJ (n − 1)
IJ (n − 1) = IJ (n − 2) + PJ (n − 1) − d (n − 1)
J
J
Ii (n − 1) = Ii (n − 2) + Pj (n − 1) − d (n − 1)
i=j i=j
The initialisation stock levels for the buffers for both models can be determined
from Eqs. (4) and (19). For instance, assume that the initial number of demand
cards, DCj (0), the production in the previous period, Pj (0), and the demand in the
previous period d(0) are all zero. Therefore, from Eq. (4) the number of demand
cards available to each production stage in the first period, DCj (1), will be zero.
The initialisation stocks for both models can be calculated from Eq. (19) as follows:
⎡ ⎛ ⎞⎤
J
0 = ⎣SS + (N Sj + 1) × D (n) − ⎝ Ii (0) − cj (0)⎠⎦ (24)
i=j
J
Ii (0) − cj (0) = SS + (N SJ + 1) × D (n) (25)
i=j
Therefore, for the final production stage the initialisation stock level should be:
IJ (0) − cJ (0) = SS + (N Sj + 1) × D (n) (26)
Since it is the final production stage, the term cJ (0) on the LHS will equal zero and
on the RHS the term N SJ will equal zero. Therefore:
IJ (0) = SS + D (n) (27)
For stage J − 1 the initialisation stock level will be:
IJ (0) + IJ−1 (0) − cJ−1 (0) = SS + (N SJ−1 + 1) × D (n) (28)
Given that cJ−1 (0) = 0, N SJ−1 = 1 and substituting in Eq. 27, then:
SS + D (n) + IJ−1 (0) = SS + 2 × D (n) (29)
IJ−1 (0) = SS + 2 × D (n) − SS − D (n) (30)
IJ−1 (0) = D (n) (31)
In fact it can be shown that for 1 ≤ j ≤ J −1 the appropriate choice for initialisation
stock is Ij (0) = D(n). Of course, if the initial number of demand cards available
to production stages is not zero then appropriate initialisation stock levels for both
models can be determined in a similar manner from Eq. (19). Therefore, EKCS and
Hodgson and Wang’s ‘Pure-Push’ PCS are equivalent if:
320 J. Geraghty and C. Heavey
q 3 4 5
The remainder of this section details the models that were developed for each
PCS examined, the experiment design and the results from the experiment. The
A review and comparison of hybrid and pull-type production control strategies 321
models have been developed by the authors with reference to the notation and
methodologies employed by [20, 21], and [28].
The models just described for each PCS were translated into discrete event simu-
lation models in eM-Plant, an object-oriented simulation software tool developed
A review and comparison of hybrid and pull-type production control strategies 323
(1) The output of our simulation model of KCS was compared to the output of
our validated simulation model of Hodgson and Wang’s conceptual model of
KCS. The results were identical when the assumption that a backlog could not
temporarily increase the number of kanbans available to the final stage was
incorporated into our simulation model of Hodgson and Wang’s conceptual
model of KCS.
(2) In Geraghty and Heavey [15] we showed mathematically that the optimal HIHS
identified by [20, 21] is equivalent to hybrid Kanban-CONWIP, under certain
conditions. We also demonstrated this equivalence by comparing the outputs
of our simulation model of hybrid Kanban-CONWIP with our validated simu-
lation model of Hodgson and Wang’s optimal HIHS.
(3) In this paper we have demonstrated mathematically that EKCS and Hodgson
and Wang’s ‘Pure-Push’ PCS will give equivalent results if the occurrence of
demand events are communicated at the same time in both PCS and other con-
ditions detailed earlier are met. Results are presented in [16] that demonstrate
that our simulation model of EKCS achieves the same results as our validated
simulation model of Hodgson and Wang’s ‘Pure-Push’ PCS when all conditions
for equivalence are met.
(4) It was not possible to validate our models of CONWIP and BSCS. However,
these PCS are simplifications of hybrid Kanban-CONWIP and EKCS, respec-
tively, in which kanbans are not distributed. Therefore, since we have been able
to validate our simulation models of hybrid Kanban-CONWIP and EKCS, we
assume that our simulation models of CONWIP and BSCS are valid.
For the purposes of the experimental process the simulation run-time over which
statistics were collected was 10, 000 periods with a warm-up period of 1, 000 peri-
ods. Ten replications of each simulation were conducted. The PCS were compared
by conducting a partial enumeration of the solution space for their control parame-
ters. A detailed description of the solution spaces evaluated for each PCS for each
demand distribution is described below.
The comparison of the strategies was achieved by conducting a partial enumer-
ation of the control parameters of the five PCS examined. The minimum values for
the Kanban allocations for the KCS, EKCS and hybrid Kanban-CONWIP models
were eight for each stage. This was selected since preliminary work indicated that
values below this level significantly degraded the solution. For instance setting the
Kanban levels of the input stages (j = 1, 2) equal to 7 always resulted in a Service
324 J. Geraghty and C. Heavey
the EKCS design, i.e. CC − 121 . The Kanban allocation for EKCS therefore was
10, 10, 15, 9 and 13 for workstations 1 to 5 respectively. These values were also
the set as the maximum initialisation stock levels, Sjmax , for each workstation.
Strategy I1 I2 I3 I4 I5
KCS K1 K2 K3 K4 K5
CONWIP 0 0 0 0 CC
Hybrid Kanban-CONWIP 0 0 0 0 CC
BSCS S1 S2 S3 S4 S5
EKCS S1 S2 S3 S4 S5
Of the five PCS examined, KCS was consistently the worst performer in terms of
addressing the Service Level vs. WIP trade-off. Table 4 illustrates this by giving
the percentage reduction in minimum WIP required by each PCS to achieve a
targeted Service Level when compared to KCS. hybrid Kanban-CONWIP was
consistently the best performer, requiring 9% to 15.5% less WIP than KCS to
achieve a targeted Service Level. A paired-t test demonstrated that the performance
of hybrid Kanban-CONWIP was statistically significantly better than CONWIP at
both 95% and 99% significance levels. BSCS and EKCS required on average 8%
to 13.5% less WIP than KCS to achieve a targeted Service Level. A paired-t test
demonstrated that the performance of EKCS was statistically significantly better
than BSCS at both 95% and 99% significance levels for all targeted service levels
with the exception of a targeted Service Level of 96%. Tables 5 and 6 illustrate
the inventory placement patterns achieved by each PCS for targeted service levels
of 100% and 99.9% respectively. KCS required more semi-finished inventory than
the other four PCS and a similar amount of end-item inventory as CONWIP and
hybrid Kanban-CONWIP to achieve a targeted Service Level. While the differences
between the other four PCS in terms of total WIP was small, the inventory placement
patterns of the PCS were different. CONWIP and hybrid Kanban-CONWIP tended
to maintain less WIP in semi-finished inventory and more in the end-item buffer
than BSCS and EKCS.
6 Discussion
In the last two decades researchers have followed two approaches to developing pro-
duction control strategies to overcome the disadvantages of KCS in non-repetitive
1
CC is a component based inventory cap, therefore the internal inventory for a component
in this parallel/serial model in period n is I1 (n) + I3 (n) + I4 (n) or I2 (n) + I3 (n) + I4 (n)
and the value 12 is arrived at as S1min + S3min + S4min or S2min + S3min + S4min
326 J. Geraghty and C. Heavey
Table 4. Percentage reduction over KCS in minimum inventory required by each PCS to
achieve a targeted service level
Table 5. Inventory placements under optimal parameters for each PCS for targeted service
level of 100%
Table 6. Inventory placements under optimal parameters for each PCS for targeted service
level of 99.9%
manufacturing environments. The first approach has been to develop new or com-
bine existing Pull-type PCS while the second approach has been to develop hybrid
PCS based on combining elements of Push and Pull PCS. In a previous paper [15]
it was demonstrated that the optimal HIHS selected by Hodgson and Wang [20, 21]
where initial stages employ push control and all other stages employ pull control, is
equivalent to a Pull-type PCS, namely hybrid Kanban-CONWIP [3, 2]. Here it was
shown that the ‘Pure-Push’ PCS modelled by Hodgson and Wang [20, 21] would
A review and comparison of hybrid and pull-type production control strategies 327
meet new customer specifications. If this is the strategy of the organisation then
EKCS is the preferable PCS for the manufacturing system.
Finally, as has been stated, the experiment presented here to examine the com-
parative performance of various Pull-type PCS was for a manufacturing system with
moderate variability and a light-to-medium demand load. Future planned work is
to examine the comparative performance of the five PCS further by examining how
the PCS respond as the coefficient of variation of the demand distribution increases
and as the mean of the demand distribution approaches the maximum capacity of
the manufacturing system.
References
1. Berkley BJ (1992) A review of the kanban production control research literature.
Production and Operations Management 1(4): 393–411
2. Bonvik AM, Couch CE, Gershwin SB (1997) A comparison of production-line control
mechanisms. International Journal of Production Research 35(3): 789–804
3. Bonvik AM, Gershwin SB (1996) Beyond kanban – creating and analyzing lean shop
floor control policies. In: Proceedings of Manufacturing and Service Operations Man-
agement Conference, Dartmouth College, The Amos Tuck School, Hanover, NH, USA
4. Buzacott JA (1989) Queueing models of kanban and MRP controlled production sys-
tems. Engineering Cost and Production Economics 17: 3–20
5. Chang T-M, Yih Y (1994a) Determining the number of kanbans and lotsizes in a generic
kanban system: A simulated annealing approach. International Journal of Production
Research 32(8): 1991–2004
6. Chang T-M, Yih Y (1994b) Generic kanban systems for dynamic environments. Inter-
national Journal of Production Research 32(4): 889–902
7. Clark AJ, Scarf H (1960) Optimal policies for the multi-echelon inventory problem.
Management Science 6(4): 475–490
8. Cochran JK, Kim S-S (1998) Optimum junction point location and inventory levels
in serial hybrid push/pull production systems. International Journal of Production
Research 36(4): 1141–1155
9. Dallery Y, Liberopoulos G (1995) A new kanban-type pull control mechanism for multi-
stage manufacturing systems. In: Proceedings of the 3rd European Control Conference
vol 4(2), pp 3543–3548
10. Dallery Y, Liberopoulos G (2000) Extended kanban control system: combining kanban
and base stock. IIE Transactions 32(4): 369–386
11. Deleersnyder JL, Hodgson TJ, King RE, O’Grady PJ, Savva A (1992) Integrating
kanban type pull systems and MRP type push systems: Insights from a Markovian
model. IEE Transactions 24(3): 43–56
12. Gaury EGA, Kleijnen JPC (2003) Short-term robustness of production management
systems: A case study. European Journal of Operational Research 148: 452–465
13. Gaury EGA, Kleijnen JPC, Pierreval H (2001) A methodology to customize pull control
systems. Journal of the Operational Research Society 52(7): 789–799
14. Gaury EGA, Pierreval H, Kleijnen JPC (2000) An evolutionary approach to select a
pull system among kanban, CONWIP and hybrid. Journal of Intelligent Manufacturing
11(2): 157–167
15. Geraghty J, Heavey C (2004) A comparison of hybrid Push/Pull and CONWIP/Pull
production inventory control policies. International Journal of Production Economics
91(1): 75–90
A review and comparison of hybrid and pull-type production control strategies 329
1 Introduction
operations that minimize the total expected costs, i.e., in other words we are looking
for optimal safety times.
The considered problem with several stages is, in general, too difficult to be
solved exactly. We therefore suggest a heuristic that is based on successive appli-
cations of the solution of a simpler one-stage problem. Different versions of such
simpler problems with one or two stages have been studied in several papers before.
Examples of this research are Yano (1987a), Kumar (1989), Hopp and Spearman
(1993), Chu et al. (1993), and Shore (1995). Song et al. (2000) consider stochastic
operation times as well as stochastic demand. In a recent overview Song and Zipkin
(2003) consider a more general class of stochastic assembly problems. There are
also a number of papers dealing with similar problems for other types of systems.
Gong et al. (1994) consider a serial system and show that the problem of choos-
ing optimal lead-times is equivalent to the well-known model in Clark and Scarf
(1960). Yano (1987b) deals also with a serial system, while Yano (1987c) consid-
ers a distribution-type system. Examples of papers analyzing related problems for
single-stage systems are Buzacott and Shanthikumar (1994), Hariharan and Zipkin
(1995), Chen (2001), and Karaesmen et al. (2002).
The outline of this paper is as follows. We first give a detailed problem for-
mulation in Section 2. In Section 3 we consider the simpler single-stage system
that is the basis for our heuristic. The approximate procedure is then described in
Section 4. In Section 5 we apply our technique to two sets of sample problems, and
finally we give some concluding remarks in Section 6.
2 Problem formulation
We consider an assembly network (see Fig. 1). The arcs represent the operations.
The node where operation i starts is denoted node i. The operation times are in-
dependent random variables with continuous distributions. Our purpose is to plan
production so that the expected holding and delay costs are minimized.
Let us introduce the following notation:
ti = starting time for operation i,
τi = stochastic duration time of operation i,
fi (x) = density for τi ,
Fi (x) = cumulative distribution function for τi , (It is assumed that Fi (x) < 1 for
any finite x.)
td = requested delivery time for the assembly,
ei = positive echelon holding cost associated with operation i,
h = sum of all echelon holding costs, i.e., holding cost for all end items,
b = positive delay cost per time unit.
The delay costs at node 0 are obtained as b(t1 + τ1 − td )+ , where x+ =
max(x, 0). If there are several end items, the delay cost is based on the maximum
delay. There are no delay costs associated with other nodes. However, delays at
other nodes may affect the delay at node 0. Consider node i, which is the starting
point for operation i. Note first that we must have ti ≥ max(tj + τj , tk + τk ). After
starting operation i, the echelon holding cost ei is incurred until the final delivery,
Planning order releases for an assembly system with random operation times 335
m j
ti
n i
τi
k
1 τ1 0
t1
which will take place at the requested delivery time td , or later in case of a delay.
However, we disregard the holding costs during the operations, because they are
not affected by when the operations are carried out. It is assumed that raw material
can be obtained instantaneously from an outside supplier. This means that initial
operations like l, m, and n in Figure 1 can start at any time.
3 Single-stage system
1
.
i
.
0
.
.
N
Fig. 2. Single-stage system
apply a simple search procedure. If we start with some initial values of the starting
times, e.g., ti = td − E(τi ), and carry out optimizations with respect to one ti at a
time the costs are nonincreasing and we will ultimately reach the optimal solution
due to the convexity.
Assume now that we have lower bounds for the starting times ti .
ti0 = lower bound for ti .
We can then obtain the optimal solution in almost the same way. We start with
feasible values of ti , e.g., ti = max(td −E(τi ), ti0 ). Then we optimize with respect
to one ti at a time. Consider a local optimization of ti . We first check whether ti = ti0
is optimal. This is the case if ∂C/∂ti > 0 for ti = ti0 . Otherwise we determine
the ti giving ∂C/∂ti = 0 as before. Again the costs are nonincreasing and we will
ultimately reach the optimal solution due to the convexity.
Consider now a general assembly system. We shall describe our approximate plan-
ning procedure. Let
P(i) = set of operations that are immediate predecessors of node i.
Consider first the immediate predecessors of the final node, i.e., P(0). In Figure 1
we have a single predecessor, but in a more general case we may have multiple
predecessors like in Figure 2. Assume that there are N operations in P(0). Assume
also that the operations that must precede these operations are finished, i.e., the N
operations in P(0) are ready to start. Consider first all operations, which do not
belong to P(0). The holding costs associated with these operations before time td
cannot be affected by the starting times for the operations in P(0). So we disregard
these holding costs in the first step. There are also holding costs associated with
these operations during a possible delay, which should be included because they
are affected by the starting times for the operations in P(0). For the operations in
P(0) both the holding costs from their respective starting times to td , and during a
possible delay are affected by the starting times for the operations in P(0). There are
also delay costs associated with a delay. This leads to the following cost expression
for the operations in P(0).
⎛ ⎞
N $ ∞ -N
C(0) = ej (td − tj − E(τj )) + (h + b) ⎝1 − Fj (x + td − tj )⎠ dx.(8)
j=1 0 j=1
Note that the only difference compared to (5) is that h includes also holding costs
during a possible delay for operations preceding P(0). Using the algorithm in
Section 3 (without lower bounds for the starting times) we optimize (8) in our first
step and get the corresponding starting times t∗i for the operations in P(0).
Assume then that i is one of the operations in P(0), and consider its immediate
predecessors, i.e., the operations in P(i). We shall now consider the single stage
system consisting of these operations and thereby interpret t∗i as a requested delivery
time. For an operation in P(i) the starting time will affect the holding costs before
338 S. Axsäter
t∗i . Let us also consider a delay cost b̂ that replaces h + b in (8). The resulting
problem is to minimize
N $ ∞ -
N
C(i) = ej (t∗i − tj − E(τj )) + b̂ (1 − Fj (x + t∗i − tj ))dx. (9)
j=1 0 j=1
Note that although we, for simplicity, are using a similar notation in (8) and (9), the
considered operations are not the same. So the number of operations N, the holding
costs ej , and the distribution functions Fj are normally different.
It remains to determine b̂. If there is a delay, this delay will affect the starting
times of the operations in P(0). This will increase the costs C(0) in (8). Consider
some given starting times for the operations in P(i) and let the corresponding
stochastic delay relative to t∗i be δ. A reasonable approximate delay cost is
8 9
dC(0) ∗ ∗ ∗
b̂ = Eδ>0 (t1 + δ, t2 + δ, ..., tN + δ)
dδ
8 9
dC(0) ∗ ∗ ∗
= Eδ (t1 + δ, t2 + δ, ..., tN + δ) / Pr(δ > 0). (10)
dδ
The second equality in (10) follows because dC(0)/dδ = 0 for δ = 0 due to the
optimality of t∗i . Because of our assumption concerning the distributions of the
operation times we know that Pr(δ > 0) > 0. In (10) it is implicitly assumed
that all operations in P(0) are started δ time units later compared to the optimal
solution, i.e., not only operation i. This is a reasonable assumption and will also
simplify the computations. Using that
d - d -
N N
Fj (x + td − tj − δ) = − Fj (x + td − tj − δ), (11)
dδ dx
j=1 j=1
certain b̂in from the considered interval. There are now two possibilities. If b̂in is
sufficiently small there is no finite optimum of (9). The resulting b̂out from (10)
will approach the upper bound. If, on the other hand, there is a finite solution we
know that b̂out is between the lower and upper bounds. Clearly, b̂out is a continuous
function of b̂in . Consequently, it follows from Brouwer’s fixed point theorem that
there exists a fixed point b̂out = b̂in , i.e., a solution corresponding to the delay cost
(10). It is easy to find such a fixed point by a one-dimensional search.
Remark. Normally b̂out is a decreasing function of b̂in . In that case it is very easy
to find the unique fixed point.
We can then handle the predecessors of the operations in P(i) in the same way,
etc. Let j be one of the operations in P(i). When dealing with P(j) we let t∗j be the
requested delivery time for the single-stage system. In (10) we are using C(i) instead
of C(0). A difference here is that the upper bound for b̂ will not necessarily lead
to a finite solution. This means in that case that the delay δ will approach infinity
and, as a consequence, also the operations in P(0) will be delayed. Consequently
it is reasonable to use the costs in the preceding step, i.e., in this case C(0) instead
of C(i). If necessary we can go one step further, and so on. This will always work
because C(0) will provide an upper bound leading to a finite optimal solution.
We will end up with starting times t∗i for all operations and delay costs for all
single-stage systems. When implementing the solution we will stick to the obtained
starting times as long as they are possible to follow. However, delays may enforce
changes. Consider, for example, operation j in Figure 1. Assume that operations
l, m, and n are finished at some time tj0 > t∗j . We then derive a new solution for
the operations in P(i). (The delay cost at node i is not changed.) In the solution we
apply the constraint tj ≥ tj0 . Starting times that have already been implemented are
regarded as given. If operation k has not yet started, its starting time may increase
but cannot decrease. To see this consider (6) and note that ∂C/∂ti is nonincreasing
if some other tj is increasing, i.e., t∗i is nondecreasing.
Let us summarize our approximate procedure:
1. Optimize C(0) as given by (8). Let K be the number of stages. Set k = −1.
2. Set k = k + 1.
3. For all operations i with k succeeding operations, optimize (9) for the operations
of P(i) under the constraint (10) with the cost function C(i). If k > 0 it may
occur that no finite optimum exists. If this is the case use the cost function for
the successor of i. If necessary go to the successor of the successor, etc.
4. If k < K − 1, then goto 2.
5. Implement the solution. In case of a delay reoptimize free starting times without
changing the delay cost.
340 S. Axsäter
5 Numerical results
To evaluate the suggested approximate procedure we used two sets of sample prob-
lems.
Problem set 1
4 1
td = 0
5 0
2
Fig. 3. Network for Problem set 1
The requested delivery time is td = 0. All operation times have the same
distribution. We denote this stochastic operation time by τ . This means that it is
relatively easy to determine also the exact solution for comparison. By symmetry
operations 3, 4, and 5 should have the same starting time ts = t3 = t4 = t5 . As
before consider first the single-stage network to the right. Let t∗1 = t∗2 be the optimal
solution of (8). Given ts there is a certain stochastic delay d at node 1 relative to t∗1 .
If d = 0 it is optimal to apply t∗1 and t∗2 . If d > 0 it is optimal to use the solution
obtained with the constraint t1 ≥ t10 = t∗1 +d. Recall that this leads to t2 ≥ t∗2 . Let
c(d) be the corresponding expected costs for the single-stage network according to
(8). We obtain the total costs as
5
C(ts ) = ei (td − ts − E(τ ))+Ed {c(d)} . (13)
i=3
This is the case both for the exact and the approximate solution. The only differ-
ence between the exact and approximate solution is the determination of ts . In the
approximate procedure we use the procedure described in Section 4. In the optimal
solution we optimize (13) with respect to ts .
All echelon holding costs are kept equal, ei = 1, while we considered three
different delay costs b = 5, 25, and 50. Furthermore the expected operation time
E(τ ) = 1 in all considered cases. Two different types of distributions for the
operation time were considered. For each distribution we considered the standard
deviations σ = 0.2, 0.5, and 1. Both distributions are constructed as α + (1 − α)X,
where α is a constant between 0 and 1 and X is a stochastic variable with its mean
equal to 1. Distribution 1 is obtained by letting X have an exponential distribution
with mean (and standard deviation) equal to 1. Distribution 2 is similarly obtained
Planning order releases for an assembly system with random operation times 341
Problem set 2
Our second problem set concerns a more complicated network with three stages
(see Fig. 4). Operations 1, 2, 3, 5, and 7 have the same stochastic operation time τ ,
while the times of operations 4 and 6 are 2τ . All times are independent. The time
τ has the same distribution as distribution 1 in Problem set 1 with E(τ ) = 1 and
the standard deviations σ = 0.2, 0.5, and 1. Furthermore, as for Problem set 1 all
342 S. Axsäter
2
5 td = 0
1 00
6
3
7
0.2 5 −4.3 −3.1 −2.3 3.68 −4.20 −3.20 −2.26 3.94 7.1
0.2 25 −5.0 −3.5 −2.5 6.46 −4.57 −3.53 −2.53 6.86 6.2
0.2 50 −5.3 −3.6 −2.7 7.82 −4.76 −3.70 −2.68 8.69 11.1
0.5 5 −4.6 −3.2 −2.6 9.18 −4.49 −3.49 −2.66 9.86 7.4
0.5 25 −5.8 −3.9 −3.4 16.16 −5.42 −4.32 −3.33 17.57 8.7
0.5 50 −7.1 −4.4 −3.8 19.12 −5.90 −4.75 −3.70 22.29 16.6
1 5 −5.2 −3.3 −3.3 17.89 −4.98 −3.99 −3.31 20.04 12.0
1 25 −8.3 −4.7 −4.6 31.85 −6.84 −5.64 −4.67 34.41 8.0
1 50 −9.9 −5.5 −5.4 38.79 −7.80 −6.49 −5.41 43.69 12.6
echelon holding costs are equal, ei = 1, and we consider the delay costs b = 5, 25,
and 50. Table 2 shows the results.
The approximate policy in Table 2 is obtained as described in Section 4. The
costs for the approximate policy are obtained by simulation. The standard deviation
is less than 0.02. The time ts1 is the starting time of the longer operations 4 and 6,
and ts2 is the starting time of operations 5 and 7. The time tm is the starting time
for operations 2 and 3 if both operations are ready to start.
The optimum by simulation is obtained by a combination of simulation and ana-
lytical techniques. We omit the details. All times, both starting times and stochastic
times were for simplicity rounded to multiples of 0.1. This does not affect the re-
sults much. More important is that in the determination of the optimal policy we
carried out minimizations over several simulated costs. This means that the costs
are somewhat underestimated. By considering the costs for the starting times sug-
gested by the approximate policy we could, however, conclude that the error is not
very significant. The cost increase of the approximate policy in Table 2 may be
overestimated by 1−2 % but hardly more.
We must conclude that the approximation errors for Problem set 2 are much
larger than for Problem set 1. The average cost increase is nearly 10 %. Note
that the intermediate time tm is very accurate also in Table 2. We note that our
Planning order releases for an assembly system with random operation times 343
approximation underestimates the needed safety times for the long operations 4
and 6 while it overestimates the safety times for the shorter operations 5 and 7. To
explain the large difference in errors, consider first the network for Problem set 1 in
Figure 3. The delay at node 1 will determine the starting times t1 and t2 because we
can initiate operation 2 at any time. A linear delay cost is then reasonable. Consider
then the network for Problem set 2 in Figure 4 and the delay at node 2. A small
delay may then not be that serious because there may any way be a delay at node 3.
A long delay may, however, cause a long delay also for operation 3. This indicates
that our linear delay cost may be less appropriate in this case. It also explains why
the longer more stochastic operations are starting so early in the optimal solution,
i.e., we wish to avoid long delays. A general, not very surprising, conclusion can
be that our approximation works better for networks with a single “critical path”.
6 Conclusions
We have considered the problem of minimizing expected delay and holding costs for
a complex assembly network where the operation times are independent stochastic
variables. An approximate decomposition technique for solving the problem has
been suggested. The technique means repeated applications of the solution of a
simpler single-stage problem. The approximate method has been evaluated for two
problem sets. The results are very good for the first set of two-stage problems and
the relative cost increase due to the approximation is only about 1 %. For the second
set of three-stage problems the errors are about 10 % and cannot be disregarded.
Although the numerical results show some promise, further research is needed for
evaluation of the applicability of the suggested technique in more general settings.
Because it is difficult to derive exact solutions for problems of realistic size it may
be most fruitful to compare different heuristics for larger problems.
References
Buzacott JA, Shanthikumar JG (1994) Safety stock versus safety time in MRP controlled
production systems. Management Science 40: 1678–1689
Chen F (2001) Market segmentation, advanced demand information, and supply chain per-
formance. Manufacturing & Service Operations Management 3: 53–67
Chu C, Proth JM, Xie X (1993) Supply management in assembly systems. Naval Research
Logistics 40: 933–950
Clark AJ, Scarf H (1960) Optimal policies for a multi-echelon inventory problem. Manage-
ment Science 6: 475–490
Gong L, de Kok T, Ding J (1994) Optimal leadtimes planning in a serial production system.
Management Science 40: 629–632
Hariharan R, Zipkin P (1995) Customer-order information, leadtimes, and inventories. Man-
agement Science 41: 1599–1607
Hopp W, Spearman M (1993) Setting safety leadtimes for purchased components in assembly
systems. IIE Transactions 25: 2–11
Karaesmen F, Buzacott JA, Dallery Y (2002) Integrating advance order information in make-
to-stock production systems. IIE Transactions 34: 649–662
344 S. Axsäter
Kumar A (1989) Component inventory costs in an assembly problem with uncertain supplier
lead-times. IIE Transactions 21: 112–121
Shore H (1995) Setting safety lead-times for purchased components in assembly systems: a
general solution procedure. IIE Transactions 27: 634–637
Song J-S, Yano CA, Lerssrisuriya P (2000) Contract assembly: Dealing with combined
supply lead time and demand quantity uncertainty. Manufacturing & Service Operations
Management 2: 287–296
Song J-S, Zipkin P (2003) Supply chain operations: Assemble-to-order systems, Ch. 11.
In: Graves SC, De Kok T (eds) Handbooks in operations research and management
science, Vol 11. Supply chain management: design, coordination and operation. Elsevier,
Amsterdam
Yano C (1987a) Stochastic leadtimes in two-level assembly systems. IIE Transactions 19:
371–378
Yano C (1987b) Setting planned leadtimes in serial production systems with tardiness costs.
Management Science 33: 95–106
Yano C (1987c) Stochastic leadtimes in two-level distribution-type networks. Naval Research
Logistics 34: 831–843
A multiperiod stochastic production planning
and sourcing problem with service level constraints
Işıl Yıldırım1 , Barış Tan2 , and Fikri Karaesmen1
1
Department of Industrial Engineering, Koç University, Rumeli Feneri Yolu,
Sariyer, Istanbul, Turkey (e-mail: [email protected];[email protected])
2
Graduate School of Business, Koç University, Rumeli Feneri Yolu, Sariyer,
Istanbul, Turkey (e-mail: [email protected])
The authors are grateful to Yves Dallery for his ideas, comments and suggestions on the
earlier versions of this paper.
Correspondence to: F. Karaesmen
346 I. Yıldırım et al.
Plant 1
Plant
Distribution Center
Retailer
Subcontractor Inventory
1
Product M
N
Decision and Control
Production
Fig. 1. A manufacturer with multiple plants that sells multiple products to a retailer
A multiperiod stochastic production planning and sourcing problem 347
difficult dynamic problem where decisions can be updated over time. Since the
equivalent deterministic problem is a well-structured mathematical programming
problem, the proposed methodology can easily be integrated with the Advanced
Planning and Optimization tools, such as the products of i2, Manugistics, etc., that
are commonly used in practice.
The organization of the remaining parts of the paper is as follows: In Section 2,
we review the literature on mathematical-programming-based stochastic produc-
tion planning methodologies. The particular stochastic production planning and
sourcing problem we investigate is introduced in Section 3. Section 4 presents the
proposed solution methodology that is based on solving the deterministic equiva-
lent problem at each time step on a rolling horizon basis. The performance of the
rolling horizon approach is evaluated by considering a number of special cases in
Section 5. Finally, conclusions are presented in Section 6.
2 Literature review
jective is to find the optimal length of production cycles that minimizes the sum of
set-up costs and inventory holding costs per unit of time and satisfies the demand
of products at the required service levels.
Clay and Grossman (1997) focus on a two-stage fixed-recourse problem with
stochastic Right-Hand-Side terms and stochastic cost coefficients and propose a
sensitivity-based successive disaggregation algorithm.
Sox and Muckstadt (1996) present a model for the finite-horizon, discrete-time,
capacitated production planning problem with random demand for multiple prod-
ucts. The proposed model includes backorder cost in the objective function rather
than enforcing service level constraints. A subgradient optimization algorithm is
developed for the solution of the proposed model by using Lagrangian relaxation
and some computational results are provided.
Beyer and Ward (2000) report a production and inventory problem of Hewlett-
Packard’s Network Server Division. The authors propose a method to incorporate
the uncertainties in demand in an Advanced Planning System utilized by Hewlett-
Packard.
Albritton, Shapiro and Spearman (2000) study a production planning problem
with random demand and limited information and propose a simulation based op-
timization method. Qui and Burch (1997) study a hierarchical production planning
and scheduling problem motivated by the fibre industry and propose an optimization
model that uses logic of expert systems.
Van Delft and Vial (2003) consider multiperiod supply chain contracts with
options. In order to analyze the contracts, they propose a methodology to formulate
the deterministic equivalent problem from the base deterministic model and from
an event tree representation of the stochastic process and solve the stochastic linear
program by discretizing demand under the backlog assumption.
For the textile-apparel-retail problem discussed in Abernathy et al. (2000), a
simulation model has also been developed (Yang et al., 1997). Then a simulation-
based optimization technique that is referred as ordinal optimization, has been used
to determine the parameters of a production and inventory control policy that gives
a good-enough solution approximately (Yang et al., 1997; Lee, 1997). However,
one needs to set a specific production and inventory control policy in the simulation.
In addition to the difficulty of setting a plausible policy in a complicated case, as the
number of sources and products increase, the number of parameters to be optimized
also increases. As a result, finding an approximate solution requires a considerable
time.
Simplified versions of the sourcing problem studied in this paper have been
investigated in the past by using stochastic optimal control (Bradley, 2002; Tan and
Gershwin, 2004; Tan, 2001). Bradley (2002) considers a system with a producer
and a subcontractor and discrete flow of goods. In an M/M/1 setting without the
service level requirements, he proves that the optimal control policy structure is a
dual-base stock policy. In this policy when the number of customers in the queue
reaches a certain level, then new incoming customers are sent to the subcontractor.
When there are no customers waiting in the queue, then the producer continues
production until a certain threshold is reached.
A multiperiod stochastic production planning and sourcing problem 349
In Tan (2001) and Tan and Gershwin (2004), a producer with a single sub-
contractor is formulated with continuous flow of goods without the service level
requirements. They also show that a threshold-type policy is optimal to decide when
and how to use a subcontractor. In the threshold policy, the subcontractor is used
when the inventory or the backlog is below a certain threshold level.
Our paper uses the idea of incorporating randomness in a deterministic math-
ematical program that is used in many of the above studies in different formats.
We utilize the approach proposed by Bitran and Yanasee (1984) that shows the
equivalence for the static problem. In contrast to this study where the main objec-
tive is determining error bounds for the optimal cost in the non-sequential case,
our main focus is generating a production and sourcing plan, i.e. determining the
values of the decision variables in the sequential (dynamic) problem where sourc-
ing decisions are made (or updated) dynamically over time. We also compare the
approximate solution of the dynamic problem with certain benchmark policies.
Since the exact optimal solution of the dynamic problem is not known, we use
two different benchmarks. It is proven that for a single source with lead time, the
proposed approach yields the same production policy as the optimal base stock
policy. For a dual-source, e.g. a producer with a subcontractor, a threshold-type
subcontracting policy suggested by Bradley (2002), Tan (2001), Tan and Gershwin
(2004) is utilized as a benchmark. After adopting the threshold policy to a more
generalized case with lead time and service-level requirements, it is observed that
the proposed approach yields very similar results to the threshold-based benchmark
in the numerical examples considered.
Assume that there is a single product and N different production sources (plants and
subcontractors). The demand for this specific product at time t, dt is random. The
main decision variables are the production quantities at each production source at
time t, Xi,t , i = 1, . . ., N . The inventory level at the end of time period t is denoted
by It . The number of periods in the planning horizon is T . The inventory holding
cost per unit per unit time is ht and the production cost at production source i at
time t is ci,t .
Constraints on the performance (related to backorders) of the system are im-
posed by requiring service levels. The frequently used Type 1 Service Level is
defined to be the fraction of periods in which there is no stock out. It can be viewed
as the plant’s no-stock-out frequency. This service level measures whether or not a
backorder occurs but is not concerned with the size of the backorder. In this study,
we consider a Modified Type 1 Service Level requirement. The Modified Type 1
Service Level forces the probability of having no stock out to be greater than or
equal to a service level requirement in each period. The service level requirement
in period t is denoted by αt .
The Stochastic Production Planning and Sourcing Problem (SP) is defined as:
# T
N
%
∗ +
Z (SP ) = MinE ht (It ) + ci,t Xi,t
t=1 i=1
350 I. Yıldırım et al.
subject to
N
It = It−1 + Xi,t − dt , t = 1, ..., T ; (1)
i=1
P {It ≥ 0} ≥ αt , t = 1, ..., T. (2)
Xi,t ≥ 0, i = 1, ..., N t = 1, ..., T. (3)
where (It )+ = Max {0, It } , t = 1, ..., T .
The objective of the problem is to minimize the total expected cost, which is
the expected value of the sum of the inventory holding and production costs in the
planning horizon. The first constraint set defines the inventory balance equations
for each time period. The next constraint imposes the service level requirement for
each period. Finally, the last constraint states that the production quantities cannot
be negative.
This formulation can easily be extended to multiple products and production
sources with lead times. Moreover different service level definitions can also be
considered by following the same approach.
The solution of the above problem at time 0 for the planning horizon [0, T ] is
referred as the static solution. The static solution is obtained by using the available
information about the distribution of demand in the future periods and the initial
inventory. A policy that sets (or updates) the future production quantities Xi,t at
time t based on the information available at that time, e.g., demand realizations,
demand distributions in the future periods, and current inventory levels, is referred
to as the dynamic solution.
In theory, the optimal policy which determines production quantities based on
actual state information may be obtained by solving the stochastic dynamic pro-
gram associated with this problem. In practice, however, there are several problems
with the stochastic dynamic programming solution. First, the well-known curse
of dimensionality makes numerical solutions challenging even for relatively small
problems. Second, it is difficult to integrate constraints on the trajectory of the
underlying stochastic processes such as service level requirements in inventory
models. Therefore, we propose a rolling-horizon approach that is based on solving
the static problem at each time period based on the available information. This, how-
ever, requires solving the static problem repeatedly which requires a transformation
explained below.
Although obtaining the optimal dynamic solution is, in general, not tractable, the
static solution can relatively easily be obtained by using deterministic mathematical
programming as suggested by Bitran and Yanasse (1984).
A multiperiod stochastic production planning and sourcing problem 351
In particular, Bitran and Yanasse show that the (Modified Type 1) service level
constraint can be transformed into a deterministic equivalent constraint by specify-
ing certain minimum cumulative production quantities that depend on the service
level requirements.
To summarize this approach, let lt denote the (deterministic equivalent) mini-
mum cumulative production quantity in period t which is calculated by solving the
probabilistic inequality:
t
P dτ ≤ lt = αt , t = 1, ..., T for lt (t = 1, ..., T )
τ =1
that yields
lt = Ft−1 (αt ), t = 1, ..., T
t
where Ft (.) is the cumulative distribution function of the random sum τ =1 dτ .
Then the probabilistic constraint P {It ≥ 0} ≥ αt , t = 1, ..., T can be expressed
equivalently by:
t
N
Xi,τ + I0 ≥ lt , t = 1, ..., T (4)
τ =1 i=1
Now, the deterministic equivalent problem with service level constraints that has
been mentioned in the previous sections can be modeled as below (Bitran and
Yanasse, 1984):
Deterministic Equivalent Problem (DEP):
T
t N N
∗
Z (DEP ) = Min ht (I0 + Xi,τ ) + ci,t Xi,t
t=1 τ =1 i=1 i=1
subject to
t
N
Xi,τ + I0 ≥ lt , t = 1, ..., T (5)
τ =1 i=1
Xi,t ≥ 0, i = 1, ..., N t = 1, ..., T. (6)
The optimal decision variable values in DEP are the same as the ones in the
solution of SP at time 0.
The static solution is obtained by transforming the stochastic problem into a
deterministic one and then solving the resulting mathematical program. The rolling
horizon approach repeats this procedure by using the available information at each
time period until time T .
It is known that the rolling-horizon approach yields good results for a number of
dynamic optimization problems. In some special cases, the rolling horizon method
may even yield the optimal solution. In this section, we evaluate the performance
of the proposed method by comparing it to certain benchmark policies in two
commonly encountered special cases in production planning.
352 I. Yıldırım et al.
We start with the special case of a single production source. When there is only one
source, the objective function includes only the holding cost (since the expected
total production costs must equal the total expected demand over the planning
horizon). In this case, we use the base stock policy as the benchmark policy. The
base stock policy is widely known and utilized in many applications. In addition,
it is known to be optimal in a number of related inventory problems. It, therefore,
constitutes a natural benchmark for comparison. The base stock policy has a single
parameter which is a reorder level and a base lot size of one unit. It aims to maintain
a pre-specified target inventory level. Under this policy, the sequence of events is
as follows: the system starts with a pre-specified base stock level in the finished
goods inventory. The arrival of the customer demand triggers the consumption of an
end-item from the inventory and issuing of a replenishment order to the production
facility. Using this policy, an order is placed (or the manufacturing facility operates)
if and only if the inventory level drops below the base stock level. The comparison
of these two models is performed for two cases with and without a lead time.
required service levels and the two policies are equivalent, we can say that the base
stock level S1 = l1 ensures that the resulting production plan satisfies the required
service levels. Note that S1 = l1 must be optimal because decreasing the base stock
level from l1 leads to an infeasible solution and increasing it above l1 would lead
to higher average inventory costs and therefore cannot be optimal.
Even though a formal proof is lacking, it is highly likely that the base stock
policy (with a stationary base stock level) is optimal for the single-plant single-
product problem in an infinite horizon setting. Theorem 1 and Corollary 1 establish
that for this problem, the rolling horizon approach yields the same solutions as the
optimal base stock policy leading us to conclude that the rolling horizon procedure
performs optimally in this case.
subject to
t
LT
Xτ −LT + SRτ + I0 ≥ lt , t = (LT + 1), ..., T ; (7)
τ =LT +1 τ =1
Xt ≥ 0, t = 1, ..., T. (8)
Our main result is as follows:
Proposition 2. When the production facility has a non-negative lead time LT, the
demand is stationary and there are no scheduled receipts initially, using a base stock
policy is equivalent to solving the deterministic equivalent model with service level
constraints on a rolling horizon basis in the following manner: assume that the base
stock level in the base stock policy including lead time equals I0 (BSLT) = S2
and the initial inventory level in the deterministic equivalent model including lead
time equals I0 (DEPLT) = lLT +1 . If S2 = lLT +1 , then the equivalent base stock
policy gives the same total expected cost value, yields the same production plan
and results in the same service level with the deterministic equivalent model with
service level constraints solved on a rolling horizon basis.
Proof. The proof of Proposition 2 is given in the Appendix.
354 I. Yıldırım et al.
Since the optimal solution of our dynamic problem is not known, a plausible bench-
mark is used to evaluate the performance of the proposed approach. We propose
a threshold subcontracting model suggested in a number of studies in the litera-
ture (Bradley, 2002; Tan, 2001; Tan and Gershwin, 2004). Although the threshold
policy is only shown to be optimal under specific assumptions including zero lead
time, stationary demand, no service level requirements, etc., we think that it is a
reasonable benchmark policy for our problem.
Fig. 2. Sample realization of dt , X1,t , X2,t and It under the threshold policy S = 15,
Z = 7, C = 8
in the planning horizon and the same 5000 sample demand streams are utilized.
The service level requirement is relaxed with the one-sided 95% confidence interval
of the simulation result. That is whenever upper confidence level of the observed
service level reaches the desired one, this case is accepted as satisfying the service
level requirement. The underlying reasoning behind making this modification in
service levels is that, the sample size we utilize might not be sufficient enough to
make the realized service level equal exactly to the required one. Among the base
356 I. Yıldırım et al.
4 16 8 1 4 0.8
4 16 12 1 4 1.2
4 16 20 1 4 2
6 1 8 1.5 0.25 0.8
6 1 12 1.5 0.25 1.2
6 1 20 1.5 0.25 2
6 4 8 1.5 1 0.8
6 4 12 1.5 1 1.2
6 4 20 1.5 1 2
stock and threshold levels that satisfy the relevant service level requirements, the
model aims to find the one with minimum total cost. The calculations are performed
for periods between 451 and 550.
For the numerical examples reported below, the order arrivals are governed by a
Poisson process with rate 10 products per period. The production cost is assumed to
be $4 per product for the in-house facility. The initial inventory level of the specific
product is set to be zero. The service level requirement is set to be 95%.
The comparison between the deterministic equivalent model and the threshold
subcontracting model is performed for nine combinations of subcontracting cost
to in-house production cost, holding cost to in-house production cost and capacity
to mean demand ratios. The combinations of subcontracting costs, holding costs
and the in-house production capacities and therefore, the combinations of relevant
subcontracting cost to in-house production cost, holding cost to in-house production
cost and capacity to mean demand ratios for which the comparisons are made can be
observed in Table 1. For each of the problem settings, the base stock and threshold
levels observed in the threshold subcontracting model are reported in Table 2.
Note that, in some of the cases, the base stock and threshold pairs are observed
to be the same. The reasoning behind this is, these pairs lead to the same average
inventory levels and minimum cost values in these settings.
While comparing the two models, total expected cost, average production cost,
average inventory holding cost values and the assignment of production to the
plants (in percentages) are the key elements we focus on. Table 3 summarizes the
total expected cost values of the deterministic equivalent model (DEM) and the
threshold subcontracting model (TSM) for the nine different scenarios for each
modified service level type.
The below tables display that the deterministic equivalent model gives very close
solutions when compared with the threshold subcontracting model for both types
of the modified levels. The deterministic equivalent model results in total expected
cost values equal to or a little bit larger than those of the threshold subcontracting
A multiperiod stochastic production planning and sourcing problem 357
4 16 8 15 7
4 16 12 15 3
4 16 20 15 −∞
6 1 8 17 7
6 1 12 16 0
6 1 20 15 −∞
6 4 8 15 7
6 4 12 15 3
6 4 20 15 −∞
Table 3. The comparison of total expected cost values observed in each scenario
model. For our set of numerical experiments, the deterministic equivalent model
gives close results to the threshold subcontracting model when the service level
requirement is of Modified Type 1.
Tables 4 and 5 display the comparison of average production and holding cost
values. As can be seen, the deterministic equivalent model gives similar results to
the threshold subcontracting model.
Table 6 summarizes the percentage of production assigned to the in-house pro-
duction facility for both the deterministic equivalent model and the threshold sub-
contracting model. The results suggest that the production assignments of the de-
terministic model follow a similar pattern with the benchmark chosen.
Based on these figures, we can conclude that the proposed deterministic equiv-
alent model solved on a rolling horizon basis performs as well as the threshold
subcontracting model solved on a simulation-based optimization technique for the
358 I. Yıldırım et al.
Table 4. The comparison of average production cost values observed in each scenario
Table 5. The comparison of average holding cost values observed in each scenario
Modified Type 1 service level. The total expected cost values of deterministic equiv-
alent models for all nine different cases are equal to or a little bit larger than those of
the threshold subcontracting model. However, we cannot reach the same conclusion
for the average production and holding cost values. The deterministic equivalent
model performs either worse for some cases or better for some other cases when
the comparison is based on average production or holding cost values. However,
the sum of these two terms, the total expected cost, is equal to a little bit larger than
that of the threshold subcontracting model. Moreover, the proportion of production
assigned to the in-house facility in the deterministic equivalent model resembles
that in the simulation based threshold subcontracting model.
It is worth mentioning that the sample size utilized in the above numerical
comparisons, 5000, might not be large enough to satisfy the service level require-
ments in each time period that the modified service level definitions necessitate.
A multiperiod stochastic production planning and sourcing problem 359
Table 6. The percentage of production assignments to the in-house production facility ob-
served in each scenario
4 16 8 75.45 75.40
4 16 12 94.73 94.70
4 16 20 99.97 100.00
6 1 8 79.76 78.17
6 1 12 94.73 98.78
6 1 20 99.97 100.00
6 4 8 75.45 75.40
6 4 12 94.73 94.70
6 4 20 99.97 100.00
The coefficient of variation in the realized service level values might be larger than
expected. To handle this problematic issue, we introduced one-sided confidence
intervals. Although the threshold subcontracting model constitutes a lower bound
in terms of total expected cost values for our set of numerical examples, it can not
be generalized from our examples that the deterministic equivalent model always
gives solutions worse than those of the threshold subcontracting model. Never-
theless, the proposed approach seems to give extremely promising results in this
particular case as well.
6 Conclusions
case or a stylized model can yield insights regarding the interaction of demand
variability, cost, and the lead time.
Appendix
Proof of Proposition 2
In period (LT + 2), the base stock policy produces XLT +1 (BSLT) = dLT +2 .
For the deterministic equivalent approach, we know that XLT +2 (DEPLT) +
2LT +1 LT +1
τ =LT +2 SRτ (DEPLT)ILT +1 (DEPLT) = XLT +2 (DEPLT)+ Xτ +
LT LT +1 τ =2
ILT +1 (DEPLT) = XLT +2 (DEPLT)+ τ =1 dτ + lLT +1 − τ =1 dτ ≥ lLT +1 ;
i.e. XLT +2 (DEPLT) ≥ dLT +1 and then, XLT +2 (DEPLT) = dLT +1 . After
the arrival of dLT +2 , the following end of period inventory levels are observed
LT +1
ILT +2 (BSLT) = ILT +1 (BSLT)+SRLT +2 (BSLT)−dLT +2 = S2 − τ =1 dτ +
LT +1 LT +2
X2 (BSLT) − dLT +2 = S2 − τ =1 dτ + d1 − dLT +2 = S2 − τ =2 dτ and
ILT +2 (DEPLT) = ILT +1 (DEPLT) + SRLT +2 (DEPLT) − dLT +2 = lLT +1 −
LT +1 LT +1
τLT =1 dτ +X2 (DEPLT)−dLT +2 = lLT +1 − τ =1 dτ +d1 −dLT +2 = lLT +1 −
+2
τ =2 dτ . Since we know that S2 = lLT +1 , ILT +2 (BSLT) = ILT +2 (DEPLT).
Now assume that at the end of any period t2 such that t2 ≥
t2
(LT + 1), It2 (BSLT)=S2 − τ =t2 −LT dτ , It2 (DEPLT) = lLT +1 −
t2
d
τ =t2 −LT τ and S 2 = l LT +1 . In period (t2 + 1), X t2 +1 (BSLT) =
dt2 and Xt2 +1 (DEPLT) is determined by the constraint Xt2 +1 (DEPLT) +
t2 +LT t2
τ =t2 +1 SRτ (DEPLT) + It2 (DEPLT) = Xt2 +1 (DEPLT) + τ =1 Xτ +
t2 −1 t2
It2 (DEPLT) = Xt2 +1 (DEPLT) + τ =1 dτ +lLT +1 − τ =1 dτ ≥ lLT +1 ;
Xt2 +1 (DEPLT) ≥ dt2 and since the model is of minimization type
Xt2 +1 (DEPLT)=dt2 . Next, a customer demand of dt2 +1 arrives. The end of pe-
riod inventory levels for both policies become It2 +1 (BSLT) = It2 (BSLT) +
t2
SRt2 +1 (BSLT) − dt2 +1 = S2 − τ =t2 −LT dτ + Xt2 +1−LT (BSLT) −
t2 t +1
dt2 +1 = S2 − τ =t2 −LT dτ + dt2 −LT − dt2 +1 = S2 − τ2=t2 +1−LT dτ
t2 It2 +1 (DEPLT) = It2 (DEPLT) + SRt2 +1 (DEPLT)
and t2 − dt2 +1 = S2 −
τ =t2 −LT d τ + X t +1−LT (DEPLT) − d t +1 = S 2 − τ =t2 −LT dτ + dt2 −LT −
t2 +12 2
References
Abernathy FH, Dunlop JT, Hammond JH, Weil D (1999) A stitch in time. Oxford University
Press, New York
Abernathy FH, Dunlop JT, Hammond JH, Weil D (2000) Control your inventory in a world
of lean retailing. Harvard Business Review Nov-Dec: 169–176
Albritton M, Shapiro A, Spearman M (2000) Finite capacity production planning with ran-
dom demand and limited information. Stochastic Programming E-Print Series
Beyer RD, Ward J (2000) Network server supply chain at HP: a case study. HP Labs Tech
Report 2000-84
Bitran GR, Yanasse HH (1984) Deterministic approximations to stochastic production prob-
lems. Operations Research 32: 999–1018
Bitran GR, Haas EA, Maatsudo (1986) Production planning of style goods with high setup
costs and forecast revisions. Operations Research 34(2): 226–236
Bradley JR (2002) Optimal control of a dual service rate M/M/1 production-inventory model.
European Journal of Operational Research (2002) (forthcoming)
Candea D, Hax AC (1984) Production and inventory management. Prentice-Hall, New Jersey
Clay RL, Grossman IE (1997) A disaggregation algorithm for the optimization of stochastic
planning models. Computers and Chemical Engineering 21(7): 751–774
A multiperiod stochastic production planning and sourcing problem 363
Feiring BR, Sastri T (1989) A demand-driven method for scheduling optimal smooth pro-
duction levels. Annals of Operations Research 17: 199–216
Kelle P, Clendenen G, Dardeau P (1994) Economic lot scheduling heuristic for random
demands. International Journal of Production Economics 35: 337–342
Lee LH (1997) Ordinal optimization and its application in apparel manufacturing systems.
Ph.D. Thesis, Harvard University, Cambridge, MA
Qiu MM, Burch EE (1997) Hierarchical production planning and scheduling in a multi-
product, multi-machine environment. International Journal of Production Research
35(11): 3023–3042
Sox CR, Muckstadt JA (1996) Multi-item, multi-period production planning with uncertain
demand. IIE Transactions 28: 891–900
Tan B, Gershwin SB (2004) Production and subcontracting strategies for manufacturers with
limited capacity and volatile demand. Annals of Operations Research (Special volume
on Stochastic Models of Production/Inventory Systems) 125: 205–232
Tan B (2002) Managing manufacturing risks by using capacity options. Journal of the Op-
erational Research Society 53(2): 232–242
Van Delft C, Vial J-PH (2003) A practical implementation of stochastic programming: an
application to the evaluation of option contracts in supply chains. Automatica (to appear)
Yang MS, Lee LH, Ho YC (1997) On stochastic optimization and its applications to manu-
facturing. Lectures in Applied Mathematics 33: 317–331
Yıldırım I (2004) Stochastic production planning and sourcing problems with service level
constraints. M.S. Thesis, Koç University, Industrial Engineering, Istanbul, Turkey
Zäpfel G (1996) Production planning in the case of uncertain individual demand extension
for an MRP II concept. International Journal of Production Economics 46–47: 153–164