Kis Etal BPMDS2017 DataDrivenFrameworkforMeasuringProcessPerformance
Kis Etal BPMDS2017 DataDrivenFrameworkforMeasuringProcessPerformance
net/publication/317235693
CITATIONS READS
17 1,961
4 authors, including:
Jan Mendling
Humboldt-Universität zu Berlin
635 PUBLICATIONS 25,860 CITATIONS
SEE PROFILE
All content following this page was uploaded by Claudio Di Ciccio on 19 December 2017.
Abstract. Studies have shown that the focus of Business Process Man-
agement (BPM) mainly lies on process discovery and process implemen-
tation & execution. In contrast, process analysis, i.e., the measurement
of process performance, has been mostly neglected in the field of pro-
cess science so far. However, in order to be viable in the long run, a
process’ performance has to be made evaluable. To enable this kind of
analysis, the suggested approach in this idea paper builds upon the well-
established notion of devil’s quadrangle. The quadrangle depicts the pro-
cess performance according to four dimensions (time, cost, quality and
flexibility), thus allowing for a meaningful assessment of the process. In
the course of this paper, a framework for the measurement of each di-
mension is proposed, based on the analysis of process execution data.
A trailing example is provided that reflects the expressed concepts on a
tangible realistic scenario.
1 Introduction
According to a survey conducted by Müller in 2010, a majority of the ques-
tioned companies saw a direct correlation between Business Process Manage-
ment (BPM) and corporate success [13]. To be able to conduct BPM success-
fully the performance of a process needs to be measured. Nonetheless, studies
have shown that business process analysis has long been neglected in the field of
BPM [16,19], as BPM devoted most of the research endeavours on the aspects
of process discovery, and process implementation and execution. To be able to
analyse a process properly, process performance has to be measured first. A
well-established paradigm in that sense is dictated by the so-called devil’s quad-
rangle [4,8]. It shows process performance based on four dimensions: time, cost,
quality and flexibility. Those four factors for performance measurement influ-
ence each other in a way that it is not possible to improve performance of one
dimension without affecting other dimensions, either positively or negatively [4].
An advantage of the approach of measuring process performance with the devil’s
quadrangle is the possibility to compare changes in the performance over time
as visually depicted in Fig. 1.
So far, process analysis has been a neglected field of process science and only
a few suggestions have been made on how process performance could be mea-
sured according to those four dimensions, such as in the case of [8,11,10,20].
Furthermore, to the best of our knowledge, no metrics have been proposed for
those dimensions that can be automatically measured over the log data of Busi-
ness Process Management Systems (BPMSs). However, in the long run process
analysis will be crucial for corporate success, thus demanding a framework that
allows for a meaningful assessment of a process.
With a focus on the service
sector, we propose a framework
that suggests how metrics for Time
2 Preliminaries
Figure 2 depicts an insurance claim process. The process starts with a claim
that is received and forwarded (activity A) to a specialist by the secretary of
the insurance company. The specialist then assesses the damage (activity B) and
writes a damage report (activity C). Subsequently, it has to be decided whether
Secretary
Secretary
Forward claim
Insurance claim
received
Insurance Company
Specialist
D
Transfer money
Money transferred
Specialist
Yes
B C
Write damage Claim approved?
Assess damage report
No
Inform insured
party
Insured party
informed
the claim is approved or not. In case of approval, the money to cover the damage
is transferred to the policyholder (activity D). If the claim is rejected, the insured
party is informed (activity E).
As it can be seen in Table 1, the average duration of the process is 940
minutes, which is equivalent to 15.66 hours. The total average duration consists
of the average duration of every single activity (except for activities D and E
who are counted as one, as one instance can only take one path, plus the time
an instance had to wait for further processing, i.e., the wait time). To sum it up,
the length of the observation period roughly corresponds to two working days,
assuming that one working day amounts to 7.5 hours. The reference period is
by default also set to two working days, i.e. 940 minutes.
The example process and its log data will be exploited in the remainder
of the paper to exemplify how the suggested metrics are measured. Before the
derivation of the metrics for each dimension though, we draw some preliminary
considerations about (i) the time span into which the process performance is
assessed, and (ii) the comparability of the measurements.
For what the first point is concerned, we are interested in the notions of
observation time and reference time. The observation time in terms of duration
is equivalent to the lead time of a process, namely the time it takes to handle
an entire case. To assess the performance of a process, it is crucial to know for
how long data on a process needs to be collected in order to allow for a sound
statement. Reference time, on the other hand, relates to the past performance
of a process or, more precisely, to the period of time for which former process
performance is observed. The setting of a reference time enables a comparison
of current process performance and past process performance, thus making it
possible to further enhance performance assessment. Therefore, the reference
period is used to compare the metrics measured during the observation period
with past process performance, so that conclusions about the development of
the process can be drawn.
To derive benefit from the devil’s quadrangle the observation period has
to be chosen carefully. This is because it is highly unlikely that a reasonable
conclusion about process performance can be drawn from the quadrangle if the
observation period is longer or shorter than the actual process time. Following
the cycle-time concept from Kanban literature [3], the average total duration of
the process is used as a basis of calculation – consequently, data on the process
has to be collected first. To ensure that the amount of data collected is sufficient,
we recommend that the process owners be consulted. They would know how
long the process lasts on average and can recommend an adequate period of
time for data collection. Once the average total duration of the process has been
determined, a safety margin in the form of the standard deviation will be added.
Of course the time frame for the observation period can be modified by the user.
The observation period calculated by the system is merely a default setting and
has to be seen as a recommendation for the user. To calculate the duration of
the observation period for the example process, the average total duration of
the process has to be determined. According to Table 1 the average observation
period should be set to 940 minutes. Please notice that for reasons of simplicity
no safety margin was added.
In Fig. 1 the reference period is represented by a dashed line. The reference
period depicts the past performance of a process for a predefined period of time.
By default the reference period comprises the same time frame as the observation
period. This setting can again be changed by the user according to the current
evaluation needs. The reference period for the example process is equal to the
observation period’s duration and amounts for 940 minutes. Both the current
quadrangle (i.e. observation period) and the original quadrangle (i.e. reference
period) are put on top of each other in order to make comparison possible.
To guarantee the comparability of the devil’s quadrangle’s four dimensions
their values should move within the same scale. In this paper it has been decided
upon a scale ranging from 0 to 100 percent. The more the value of one dimension
approaches 0, the worse the respective dimension performed.
3 Approach
Throughout this section, we define the metrics that we associate to each di-
mension of the devil’s quadrangle: Section 3.1 deals with the time dimension,
Section 3.2 is concerned with the cost dimension, Section 3.3 focuses on the
quality dimension, and Section 3.4 discusses the flexibility dimension.
Activity
Wait. Lead. Service Service/time
A B C D E time time time ratio
Run 1 10 240 60 20 360 690 330 47.83%
Run 2 5 300 120 20 240 685 445 64.96%
Run 3 20 360 120 10 240 750 510 68.00%
Run 4 10 240 120 10 180 560 380 67.86%
Run 5 10 180 60 20 360 630 270 42.86%
Run 6 25 300 240 10 360 935 575 61.50%
Run 7 10 300 120 30 300 760 460 60.51%
Run 8 10 240 60 10 60 380 320 84.21%
Run 9 5 360 180 20 360 925 565 61.08%
Run 10 5 360 180 10 360 915 555 60.66%
Total Runtime 110 2.880 1.260 100 60
actually handling a case [8]. The service time of a case is then compared with
the case’s lead time, thus indicating how much of the total time an instance
takes to finish is spent on actual work. The higher the service time ratio, the
better, as more time has been spent on actually handling a case and less time
was lost due to a process instance being at a resting stage.
As it is very likely that more than one case is examined during the obser-
vation period, we once again calculate the service time ratio for each process
instance, i.e., the comparison of a case’s service time with its respective lead
time, and calculate a median for all the single values. The resulting median is
then transferred to the time axis of the quadrangle.
In order to generate data for the calculations regarding the insurance claim
process, we ran an example of ten instances of a process (see Table 2). First,
the lead time of the process, consisting of the duration of each activity and the
wait time, i.e., the time an instance waited for further processing, was calculated.
After that, the service time (i.e. the time a process instance was actually handled)
was computed. These two steps were taken in order to be able to gather the
service time ratio, which indicates how much time of the process was spent on
actual work. In the end, the service time ratios for each run were sorted in
descending order to calculate the median for the time dimension. The resulting
median for our computation is 61.29%.
of each employee can be calculated. It is important to notice that for cost cal-
culation the actual personnel costs have to be used. The term actual personnel
costs refers to direct payments to employees increased by continued payment of
salaries (in case of holiday, sick leave and bank holidays), holiday pay, Christmas
bonuses, the employer’s social security contributions, overtime rates and other
personnel costs [6].
As the event log stores information about how long an employee has been
working on a task, a viable measurement of the process’ cost can be achieved by
multiplying the actual labour costs per hour by the processing time per task. The
costs will then be determined for the selected observation period. The result is
then compared with the total costs of the organisation for the same observation
period resulting in a percentage value that can be transferred to the cost axis
of the devil’s quadrangle. The higher the value the worse the process performed
with regard to the cost dimension (i.e. a high value means high personnel costs
compared to total costs). However, it has already been mentioned in Section 2,
we want all the axes to have a uniform meaning: The closer the value is to 100%
the better process performance is rated. This is why the value received from the
previous division has to be inverted before transferring it to the cost axis.
Considering the insurance claim process example, it is assumed that the costs
for the secretary amount to $20 per hour and that the specialist is paid $40 per
hour. Moreover, it is exactly known which activities are handled by whom. With
this in mind, the total duration of activity A is multiplied by the hourly costs
of the secretary, whereas the duration of the remaining activities is multiplied
by the hourly rate of the specialist (the total duration can also be extracted
from Table 2). The resulting sum is then compared with the total cost of the
company, which we estimated with $4.000 per working day. The value for the
cost dimension thus amounts to the inverted ratio: Its value, 27.42%, can be
depicted on the quadrangle.
We remark here that we assume a complete knowledge of the work of re-
sources, with information on the assigned task and duration of the carry-out
thereof. This is a reasonable assumption in case a BPMS is supporting the pro-
cess execution. Otherwise, the conduction of tasks in parallel, or the interrup-
tions during task handling, holidays, weekends, etc., need a substantial amount
of effort to be considered [1].
for the insurance company. At that stage, the number of positive end events
is compared with the total number of process instances within the observation
period. The four positive end events thus are divided by the total of ten, i.e., the
number of process instances in the observation period. The result is an estimated
outcome quality of 40%.
Technical Quality. To enable the assessment of a process’ technical quality
the number of incidents within a predefined observation period can be counted.
The more incidents registered for a specific period of time, the worse the process
performed in terms of technical quality. In the end, all incidents recorded in the
observation period are compared with the total number of elements in a process.
The following example should help to better illustrate this procedure. It is
assumed that a process instance records 20 events in the log. During the process
execution five incidents are thrown. The technical quality given by the inverted
ratio of incidents per process results in a value of 75% for that process instance.
Afterwards, a median is calculated for all the values of the separate process
instances. The resulting median is then transferred to the quadrangle’s quality
axis.
Table 4 summarises all the incidents that were registered during the run of
one process instance of the example insurance claim process. The number of
incidents is then compared with the number of elements that occurred within
the same process instance, resulting in a percentage value. The higher this value,
the higher the number of incidents within one process instance. To gather a value
that represents the technical quality of the whole observation period, the median
for all incident rates is calculated. Thus, the technical quality of the process
equals 87.5% We remark that the ratio is inverted so as to keep consistency and
comparability of the metrics: the more the values on the scale approach 100 %,
the better the process performed. In contrast, a higher incident rate means lower
technical quality, thus requiring for an inversion of the original result.
Combining Outcome and Technical Quality. To transfer a single value
to the respective axis of the quadrangle, we combine the aforementioned quality
measurements into a single one. This is achieved by assigning a weight, which can
be chosen according to the interest of an organisation, to each of the two quality
metrics, i.e., outcome quality and technical quality. Then the values resulting
from the measurement of each dimension are multiplied with their respective
weights in order to compute a single value that can be transferred to the quality
axis of the devil’s quadrangle.
For the example process, it was chosen to weigh both outcome and technical
quality with 50% which results in a combined value of 66.88%, which can be
transferred to the quadrangle’s quality axis.
taken in the initial phase of the process analysis we know the planned average
lead time of the process.
A first approach to the concept of flexibility is the calculation of an open-
closed-ratio (OCR) for a previously defined observation period. This ratio is
computed by comparing cases in progress (open cases) with completed cases
(closed cases). When there has been an increase in the workload and the values
for open and closed cases balance roughly, it can be presumed that there is a
constant handling of cases. If, on the other hand, there is no balance between
the two values, it can be concluded that the process lacks flexibility.
The only problem of measuring the flexibility dimension as suggested above is
that the results could be corrupted. This is owing to the fact that the workload
level is not taken into account. To put it differently, what would happen if
the workload does not change, hence remains at a constant level, and the OCR
indicates a constant handling of cases? It could be assumed that the organisation
is highly flexible even though the workload remained stable. However, this does
not correspond to the definition of flexibility, rooted in the ability to adapt to a
changed or new situation. For this reason an additional factor has to be included
in the measurement of flexibility: the workload itself.
As it can be seen in Fig. 3, two factors are taken into consideration for the
measurement of process flexibility: (i) the OCR, and (ii) the workload, i.e, the
cases in progress. The OCR shows the open and closed cases of an observation
period, i.e., the changing workload over time. If the workload increases (resp.
decreases) over time, the OCR has to rise (fall) too in order to be able to speak
of a highly flexible process. In contrast, if the OCR remains on the same level
this is a sign of lacking flexibility, because the process is apparently not able to
adapt to changing conditions.
To receive a percentage value for the volume flexibility, the coverages both of
the workload and the OCR curve have to be compared. The higher the coverage,
the more flexible the process is, as this indicates that the OCR is able to adapt
to the changing workload conditions. However, we remark that the OCR will
not rise immediately after an increase in the workload, because the cases take a
certain time to finish – at least the average lead time. It still has to be considered
that both the workload curve and the curve representing the OCR could exactly
coincide, even though the number of open and closed cases over time did not
increase recently, i.e., the backlog remained on the same level. Again we are
confronted with a case where an organisation faces steady workload, which does
not correspond to our definition of flexibility. Our understanding of flexibility is
that an organisation is able to adapt to changing conditions. But where there is
no change, there can be no reaction either. We thus integrate a warning signal
that indicates an increased (decreased) amount of cases in progress (or congruent
areas below the curves) but no significant rise (fall) of the backlog curve. The
user should then be enabled to switch to a more detailed view where both curves
are shown, as suggested in [14].
We report two examples showing different ways to measure the volume flexi-
bility, both complying with the described rationale yet tackling the computation
from two different perspectives: the first one measures volume flexibility in terms
of the total duration per case, the second one adopts a more global perspective
and focuses on the number of cases that were opened and closed within the
observation period. Both examples refer to the example insurance claim process.
For what the first computation strategy is concerned, Table 5 shows the
open cases within the observation period of 940 minutes. It is assumed that
the usual number of open cases is ten. It can be then recognised that there are
five additional cases to handle with respect to the expectations. This implies
that the workload has increased and the measurement of the process’ flexibility
can be started. To consider the process flexible, each case has to finish within
the average lead time of the process. In the “Lead time” column the actual
lead time of the respective instance is reported. The average lead time, based
on our calculation for the framework, is given in column “Target lead time”.
In another step, target lead time and actual lead time of every instance are
compared, showing that in total the 15 instances took 225 minutes longer than
planned to finish. Expressed as a percentage, the process took 1.60% longer
than initially planned, thus reducing volume flexibility to 98.40%. The second
strategy to measure the volume flexibility metric refers to open and closed cases
in the observation period. In Table 6 the open cases in the observation period are
reported. Under the assumption that the normal number of open cases is ten, it
can be recognised that there are additional five cases to handle which signals a
higher workload. In order to be considered flexible, all 15 cases have to be closed
within the average lead time of the process. However, as can be gathered from
Table 6, only five cases were closed, meaning that 66.67% of the cases are still
open for processing and thus reducing volume flexibility to 33.33%.
The purpose of the variation in the calculation of volume flexibility for the
example process is to show how different viewpoints influence the outcome of
metrics measurement. It can be recognised that the results for volume flexibility
4 Conclusion
Within this paper, suggestions for the measurement of metrics for the four di-
mensions of the devil’s quadrangle have been made, exclusively based on the
inspection of process data. All suggested calculations are made under the as-
sumption that the data are extracted from a BPMS event log running the pro-
cess. The log data used for metrics measurement has been chosen under the
condition that it can be extracted from almost any BPMS. For the calculation
of a value for the time dimension, the activity execution time is considered as
reported in the log. The activity execution time is needed again in order to mea-
sure the cost metric. The additional piece of information needed in that case
are the personnel expenses. Due to the amplitude of interpretations that can
be given to the quality and flexibility dimensions in particular, we identified a
combination of metrics, each singularly considering different aspects thereof. For
the assessment of the outcome quality, a comparison between the positive and
the negative terminations of the process instances is compared. For the measure-
ment of the technical quality, the number of incidents within a process instance is
considered. The quality dimension is ultimately assessed as a linear combination
of the aforementioned ones. Process flexibility is also assessed as a weighted sum
of two different components: the volume flexibility and the technical flexibility.
The former is measured on the basis of a comparison between open and closed
cases during the observation period. The latter is likewise based on the reaction
time to incidents.
References
1. van der Aalst, W.M.: Business process simulation revisited. In: EOMAS. pp. 1–14.
Springer (2010)
2. van der Aalst, W.M., Rosemann, M., Dumas, M.: Deadline-based escalation in
process-aware information systems. Decision Support Systems 43(2), 492–511
(2007)
3. Dickmann, P.: Schlanker Materialfluss : mit Lean Production, Kanban und Inno-
vationen. Springer Vieweg, Berlin, Heidelberg, 3rd edn. (2015)
4. Dumas, M., La Rosa, M., Mendling, J., Reijers, H.A.: Fundamentals of Business
Process Management. Springer (2013)
5. Gong, Y., Janssen, M.: Measuring process flexibility and agility. In: ICEGOV. pp.
173–182. ACM (2010)
6. Horsch, J.: Kostenrechnung: Klassische und neue Methoden in der Unternehmen-
spraxis. Springer-Verlag (2015)
7. Houy, C., Fettke, P., Loos, P., van der Aalst, W.M., Krogstie, J.: BPM-in-the-
large – towards a higher level of abstraction in business process management. In:
E-Government, E-Services and Global Processes, pp. 233–244. Springer (2010)
8. Jansen-Vullers, M., Loosschilder, M., Kleingeld, P., Reijers, H.: Performance mea-
sures to evaluate the impact of best practices. In: BPMDS workshop. vol. 1, pp.
359–368. Tapir Academic Press Trondheim (2007)
9. Knoll, K., Jarvenpaa, S.L.: Information technology alignment or “fit” in highly
turbulent environments: the concept of flexibility. In: SIGCPR. pp. 1–14. ACM
(1994)
10. Kronz, A.: Managing of process key performance indicators as part of the aris
methodology. In: Corporate performance management, pp. 31–44. Springer (2006)
11. Kueng, P.: Process performance measurement system: a tool to support process-
based organizations. Total Quality Management 11(1), 67–85 (2000)
12. Kumar, R.L., Stylianou, A.C.: A process model for analyzing and managing flex-
ibility in information systems. European Journal of Information Systems 23(2),
151–184 (2014)
13. Müller, T.: Zukunftsthema Geschäftsprozessmanagement. Tech. rep., Pricewater-
houseCoopers AG Wirtschaftsprüfungsgesellschaft (2011)
14. Nguyen, H., Dumas, M., ter Hofstede, A.H., La Rosa, M., Maggi, F.M.: Business
process performance mining with staged process flows. In: CAiSE. pp. 167–185.
Springer (2016)
15. Parasuraman, A., Zeithaml, V.A., Berry, L.L.: Servqual: A multiple-item scale for
measuring consumer perc. Journal of retailing 64(1), 12 (1988)
16. Recker, J., Mendling, J.: The state of the art of business process management
research as published in the BPM conference. Business & Information Systems
Engineering 58(1), 55–72 (2016)
17. Regev, G., Wegmann, A.: A regulation-based view on business process and sup-
porting system flexibility. In: CAiSE. vol. 5, pp. 91–98. Springer (2005)
18. Stephens, S.: Supply chain operations reference model version 5.0: a new tool to
improve supply chain efficiency and achieve best practice. Information Systems
Frontiers 3(4), 471–476 (2001)
19. Van Der Aalst, W.M.: Business process management: A comprehensive survey.
ISRN Software Engineering 2013 (2013)
20. Venkatraman, N., Ramanujam, V.: Measurement of business performance in strat-
egy research: A comparison of approaches. Academy of management review 11(4),
801–814 (1986)
References
Kis, Isabella, Stefan Bachhofner, Claudio Di Ciccio, and Jan Mendling (2017).
“Towards a Data-Driven Framework for Measuring Process Performance”.
In: BPMDS/EMMSAD. Vol. 287. Lecture Notes in Business Information
Processing. Springer, pp. 3–18. isbn: 978-3-319-59466-8. doi: 10.1007/
978-3-319-59466-8_1.
BibTeX
@InProceedings{ Kis.etal/BPMDS2017:DataDrivenFrameworkforMeasuringProcessPerformance,
author = {Kis, Isabella and Bachhofner, Stefan and Di Ciccio,
Claudio and Mendling, Jan},
title = {Towards a Data-Driven Framework for Measuring Process
Performance},
booktitle = {BPMDS/EMMSAD},
year = {2017},
pages = {3--18},
publisher = {Springer},
crossref = {BPMDS-EMMSAD2017},
doi = {10.1007/978-3-319-59466-8_1}
}
@Proceedings{ BPMDS-EMMSAD2017,
title = {Enterprise, Business-Process and Information Systems
Modeling: 18th International Conference, BPMDS 2017, 22nd
International Conference, EMMSAD 2017, Held at CAiSE 2017,
Essen, Germany, June 12-13, 2017, Proceedings},
year = {2017},
volume = {287},
series = {Lecture Notes in Business Information Processing},
publisher = {Springer International Publishing},
isbn = {978-3-319-59466-8},
doi = {10.1007/978-3-319-59466-8}
}