Envelopment Analysis (DEA) and Machine Learning
Envelopment Analysis (DEA) and Machine Learning
with Applications
PERGAMON Expert Systems with Applications 16 (1999) 283–296
Abstract
Data envelopment analysis (DEA), a non-parametric productivity analysis, has become an accepted approach for assessing efficiency in a
wide range of fields. Despite its extensive applications, some features of DEA remain unexploited. We aim to show that DEA can be used to
evaluate the efficiency of the system integration (SI) projects and suggest the methodology which overcomes the limitation of DEA through
hybrid analysis utilizing DEA along with machine learning. In this methodology, we generate the rules for classifying new decision-making
units (DMUs) into each tier and measure the degree of affecting the efficiencies of the DMUs. Finally, we determine the stepwise path for
improving the efficiency of each inefficient DMU. 䉷 1999 Elsevier Science Ltd. All rights reserved.
Keywords: Data envelopment analysis; System integration; Self-organized map; C4.5; Machine learning
of system integration (SI) projects, as considering these DMU in each time period as being independent from the
characteristics of DEA mentioned. Previous studies have performance in the previous period. Also, with this
been rarely applied to evaluate the performance of SI approach it is not feasible to ascertain trends in performance
projects except for some applications of DEA to measure or to observe the persistence of efficiency or inefficiency.
the productivity of software projects. By SI project we mean The window analysis approach corrects for some of these
all the activities that are necessary to build and maintain problems. For cross-section/time series/panel data one
various kinds of information systems in response to their could employ window analysis or a Malmquist index to
customers’ needs. SI companies mainly carry their works on examine changes across time periods.
a projects basis. Upon receiving the project request from a Window analysis, dubbed by Charnes et al. (1985) and
customer, the SI company organizes a team for the project. recently applied by Charnes et al. (1994) and Day et al.
But it is often over budget, late in delivering and does not (1994), 1995), uses DEA to analyze panel data by convert-
satisfy the requesting user. Precise evaluation of projects ing a panal into an overlapping sequence of windows, which
becomes an important issue for SI companies. Hence the are then treated as separate cross-sections. The Malmquist
quality of the company is determined by the efficiency of index, developed by Fae et al. (1995), uses DEA to analyze
projects. The evaluation results of the proposed projects are panel data by constructing a Malmquist (1953) type of index
considered critical in gaining a competitive strategy for the of productivity change. One advantage of this approach is
SI company. The project evaluation results also influence that it provides two types of information: efficiency evalua-
the level of incentives for the project members. tions for each DMU between each successive pair of
In this paper, we aim to show that DEA can be used to periods. Another advantage of this approach is that it
evaluate the efficiency of SI projects and suggest the meth- provides a decomposition of productivity change into two
odology to overcome the limitation of DEA. We present our mutually exclusive and exhaustive components: efficiency
research framework, which is divided into two phases. In the change and technical change.
first phase, we generate the rules for classifying new DMUs
into each tier, and discriminate among the input and output 2.2. Efficiency evaluation factors of SI projects
variables by the degree of affecting the efficiencies of the
DMUs (discriminant descriptor). In the second phase, we Within an efficiency measurement framework, one is
determine the stepwise path for improving the efficiency of more interested in assessing how well a DMU uses its
each inefficient DMU. resources to obtain a desired outcome outcome; alterna-
The remainder of the paper is structured as follows. tively, one may want to assess how good an outcome one
Section 2 presents a review of literature on DEA and SI is producing with the given resources. Thus, one is intui-
project measures. This is followed by a description of the tively interested in defining the main resources (inputs) and
research methodology in Section 3. Subsequently Section 4 the relevant products (outputs) of the process, and in finding
presents results. The concluding remarks are presented in appropriate measures for these attributes. But, measuring SI
Section 5. projects has not been easy, primarily because most research-
ers and practitioners have difficulty in agreeing on what to
measure and how to measure it.
2. Literature review Christopher et al. (1996) used software quality (Customer
Satisfaction Index) and meeting targets, which includes
2.1. DEA schedules and budgets, and rework after delivery as output
variables of software development projects.
DEA was developed in operational research and Software quality (CSI) covers the extent to which the
economic studies as a method for assessing the efficiency software system meets the actual needs of the intended
of activity units, making the minimum possible assumptions users. These needs can be diverse. For example, a major
regarding the functional form of the underlying production software developer conducted a customer satisfaction
function. DEA is a method based on linear programming survey that included seven aspects of software quality: relia-
that has been used extensively for assessing the relative bility, capability, usability, installability, maintainability,
efficiency of activity units of non-profit (e.g., schools, performance and documentation.
hospitals) and profit-making (e.g., banks, airlines) Meeting targets centers around the fact that, to be
organizations. successful, a project should be ‘‘on time and on budget’’.
In its extensive applications, longitudinal studies in DEA Projects that are behind schedule or over budget have a
are still very rare. Most DEA analyses are comparing the number of consequences. Anticipated benefits of the
performance of DMUs in the same time period. One completed project may be lost or delayed. People on the
approach to performing the longitudinal analysis is to project generally must stay with the project, instead of
compare cross-sectional runs across the time periods in moving on to a new one, which would likely delay the
the study. This approach introduces variability into the next project. A late project could lead to an embarrassing
analysis, however, because it treats the performance of a post hoc review of the original decision to start the project or
H.K. Hong et al. / Expert Systems with Applications 16 (1999) 283–296 285
to select a particular group of people to carry it out. An we determine the stepwise path for improving the efficiency
important issue is the level of realism in the initial schedule of each inefficient DMU (refer to Fig. 1).
and budget. These are often negotiated between the devel-
oper and user organizations. Since the users’ goal is to mini- 3.1. Classification rule generation for each tier
mize costs and completion time, and the developers’ goal is
to gain the users’ agreement to the project, there is a In the first phase, we evaluate the efficiencies of the
tendency to set unrealistically low targets. DMUs via a DEA and cluster the DMUs together through
Rework has been observed on many SI projects. Rework the tier analysis, which recursively applies the DEA analysis
may occur as a result of poor understanding of the require- to the remaining inefficient DMUs, and then generate the
ments or poor technical design. When not planned, as in a DMU classification rules using the C4.5, a decision tree
prototyping strategy, rework can play havoc with schedules classifier, with the DMU tiers identified by the tier analysis.
and budgets.
Also, Banker and Kemerer (1992) used Budget Perfor- 3.1.1. The efficiency evaluation of SI projects — input/
mance, Schedule Performance, User Satisfaction, and Main- output data set
tenance Complexity as output variables of Information In this paper, we propose an SI project management
Systems Development Projects. model with four inputs and four outputs as shown in Fig.
All the models have Labor, measured either in labor hours 2. The basic resources (inputs) used by each DMU are mate-
or cost, as their main input representing the effort. rial and equipment resources such as software and hardware
tools, and total labor hours. Total labor hours are the amount
of total person-month considering career. The outputs are
3. Methodology customer satisfaction index, schedule performance, budget
performance, and rework hours after delivery. These vari-
In this section we present our research framework which ables are summarized in Table 1.
is divided into two phases. In the first phase, we generate the There are other criteria of judging the performance of SI
rules for classifying new DMUs into each tier and determine projects that are not covered in this paper. One such criter-
the input and output variables that will discriminate best ion could be productivity such as source lines of code
between the tiers by the degree of affecting the efficiencies (SLOC) and function points (FP). These variables are not
of the DMUs (discriminant descriptor). In the second phase, included in this paper because it does not seem feasible to
reliably measure them across projects.
Table 1
Summary of variables
Variable Measurement
Input factors Labor hours A (La) Amount of total person-months, who have a career over 10 years
Labor hours B (Lb) Amount of total person-months, who have a career between 6 and 10 years
Labor hours C (Lc) Amount of total person-months, who have a career below 5 years
Material and equipment resource (Mr) Total monetary amount of hardware, software and other materials
Output factors Customer Satisfaction Index (CSI) Customer questionnaire
Schedule performance (Sp) Ratio of planned period to real period
Budget performance (Bp) Difference between real development cost and planned budget
Rework after delivery (Rew) 5 person-months minus person-month of rework (additional service)
non-parametric (mathematical programming) approaches efficiency score of each DMU, subject to the condition that
use all the information contained in the data. the set of weight. Obtained in this manner, DMU must also
In parametric analysis, the single optimized regression be feasible for all the other DMUs included in the calcula-
equation is assumed to apply to each DMU. DEA, in tion. For each inefficient DMU (one that lies below the
contrast, optimizes the performance measure of each DMU. frontier), DEA identifies the sources and level of ineffi-
This results in a revealed understanding about each ciency for each of the inputs and outputs. The level of inef-
DMU instead of the depiction of a mythical ‘‘average’’ ficiency is determined by comparison to a single referent
DMU. In other words, the focus of DEA is on the individual DMU or a convex combination of other referent DMUs
observations as represented by the n optimizations (one for located on the efficient frontier that utilize the same level
each observations) required in DEA analysis, in contrast to of inputs and produce the same or a high level of outputs.
the focus on the averages and estimation of parameters that Details of the methodology as well as descriptions of data
are associated with single-optimization statistical envelopment analysis can be found in Charnes et al. (1978).
approaches. DEA calculates a maximal performance
measure for each a DMU relative to all the DMUs in the
3.1.3. Clustering the DMUs through the tier analysis
observed population with the sole requirement that each
In preceding section, we used DEA to evaluate the effi-
DMU lie on or below the extreme frontier. Each DMU not
ciencies of SI projects. The DEA determines the most
on the frontier is scaled against a convex combination of the
productive group of the DMUs and the group of less-
DMUs on the frontier facet closest to it.
productive DMUs. That is, the DMUs are clustered into
The solid line in Fig. 3 represents a frontier derived by
an efficient group or an inefficient one by DEA. A similar
applying DEA to data on a population of DMUs, each utiliz-
approach to clustering DMUs by DEA was presented by
ing different amounts of a single input to produce various
Thanassoulis (1996). However, the clusters on that study
amounts of a single output. It is important to note that DEA
were not grouped by their efficiency levels but by the char-
calculations, because they are generated from actual
acteristics of the input resource mix. Tier analysis that we
observed data for each DMU, produce only relative effi-
propose is a kind of technique that can be used to cluster
ciency measures. The relative efficiency of each DMU is
DMUs together according to their efficiency levels.
calculated in relation to all the other DMUs, using the actual
In the first step of tier analysis, we obtain the efficiency
observed values for the outputs and inputs of each DMU.
scores of the entire set of DMUs. The result of the first step
The DEA calculations are designed to maximize the relative
should reveal the most efficient group of DMUs by indicat-
ing that their scores are equal to 1. We call this group ‘‘tier
1’’. In the second step, we proceed DEA again only with the
inefficient DMUs which are not part of tier 1. DMUs whose
efficiency scores in the second step are equal to 1 are tier 2.
The same procedure can be repeated while the number of
remaining DMUs is at least three times greater (8 × 3 24)
than that of inputs along with outputs (4 ⫹ 4 8), as Banker
et al. (1984) have proposed. This makes it possible to appro-
priately discriminate between efficient DMUs and ineffi-
cient DMUs. We call this procedure the tier analysis
because DMUs that belong to the efficient group in each
step form the efficient production frontier in each step as
shown in Fig. 4.
Fig. 4 shows that DMUs on tier 1 are superior to those in
Fig. 3. Comparison of DEA with regression analysis. tier 2 and DMUs in tier 2 are superior to those in tier 3. We
H.K. Hong et al. / Expert Systems with Applications 16 (1999) 283–296 287
use these DMU tiers as input data for C4.5 to generate DMU equation (Quinlan, 1993),
classification rules and also use them to determine the step-
gain ratio
X gain
X=split info
X
wise improvement path for any of inefficient DMUs.
where
3.1.4. Generating classification rules for each tier using X
n
兩Ti 兩 兩Ti 兩
C4.5 slit info
X ⫺ × log2 ;
A typical decision tree learning system, C4.5, which is i1 兩T兩 兩T兩
going to be used to generate the rule set for classifying SI
projects, adopts a supervised learning scheme that gain
X info
T ⫺ infoX
T
constructs decision trees from a set of examples. A decision and gain(X) measures the information that is gained by
tree is a directed graph showing the various possible partitioning T in accordance with the test X.
sequences of questions (tests), answers and classifications. We generate the rules for classifying new DMUs into
The method first chooses a subset of the training examples each tier and determine the input and output variables that
(window) to form a decision tree. If the tree does not give will discriminate best between the tiers by the degree of
the correct answer for all the objects, a selection of the affecting the efficiencies of the DMUs (discriminant
exceptions (incorrectly classified examples) is added to descriptor).
the window and the process continues until the correct deci-
sion set is found. The eventual outcome is a tree in which 3.2. Determination of the improvement path
each leaf carries a class name, and each interior node
specifies an attribute with a branch corresponding to each In the second phase of our analysis, we plan to use a self-
possible value of that attribute. organized map (SOM), which is one clustering tool, and the
C4.5 uses an information theoretic approach aiming at DMU tiers further to suggest improvement paths for ineffi-
minimizing the expected number of tests to classify an cient projects based on the features of the projects.
object. The attribute selection part of C4.5 is based on the
assumption that the complexity of the decision tree is 3.2.1. Clustering the DMUs using SOM
strongly related to the amount of information. An informa- DEA offers no guidelines where relatively inefficient
tion-based heuristic selects the attribute providing the high- DMUs improve since a reference set of an inefficient
est information gain ratio, i.e., the ratio of the total DMU consists of several efficient DMUs. Hence, we make
information gain due to a proposed split to the information an SOM to group similar DMUs according to the character-
gain attributable solely to the number of subsets created as istics of the inputs, for the inefficient DMU to select an
the criterion for evaluating proposed splits. efficient DMU in a reference set as a benchmarking target
The C4.5 system uses a information gain ratio as the (refer to Fig. 5).
evaluation function for classification, with the following The SOM uses an unsupervised learning scheme to train
288 H.K. Hong et al. / Expert Systems with Applications 16 (1999) 283–296
the neural network (Sestito and Dillon, 1994; Berry and statistical features, and model the probability distributions
Linoff, 1997). Unsupervised learning is comprised of that are present in the input data.
those techniques for which the resulting actions or desired The SOM uses competitive learning. When an input
outputs for the training sequences are not known. The pattern is imposed on the network, one output node is
network is only told the input vectors, and the network selected from among all the output nodes as having the
self-organizes these inputs into categories. smallest Euclidean distance between the presented input
Each link between a node in the input layer and a node in pattern vector and its weight vector. This output unit is
the output layer has an associated weight. The net input into declared the winner in the competition among all the
each node in the output layer is equal to the weighted sum of neurons in the output layer. Only the winning neuron gener-
the inputs. Learning proceeds by modifying these weights ates an output signal from the output layer. All the other
from an assumed initial distribution with the presentation of neurons have a zero output signal (see Fig. 6).
each input pattern vector. This process identifies groups of The input and weight vectors are usually normalized in an
nodes in the output layer that are close to each other and SOM so that they have values between 0 and 1 inclusive. If
respond in a similar manner. A particular group of units the dot products between the normalized input vector X^ and
together forms an output cluster. The topology-preserving a normalized set of weight vectors W^ j are determined, the
mappings from the inputs to the clusters reflect the existing neuron with the largest dot product (the one with the smal-
similarities in the inputs and capture any regularities and lest Euclidean distance) is declared to be the winner. Thus
the winner is the vector obtained from the expression: limit. Thus, during these stages, fewer neurons have
their weights adjusted closer to the input vector. Lateral
max
X^ t W^ j inhibition of weight vectors that are distant from a
j
particular input pattern may also be carried out as shown
in Fig. 7.
As learning involves adjustment of weight vectors, learn-
A general algorithm for the SOM may be summarized as
ing with this particular input pattern is restricted to lateral
follows.
interconnections with immediately neighboring units of the
winning neuron in the output layer. Adjusting their weights 1. Initialize weights to small random values and set the
closer to the input vector carries out learning for the nodes initial neighborhood to be large. One approach is to set
within the neighborhood. The size of the neighborhood is each weight vector equal to an input vector pattern when
initially chosen to be enough large to include all units in the there are more training input patterns than output units.
output layer. However, as learning proceeds, the size of the This approach performs best with very large networks
neighborhood is progressively reduced to a pre-defined and training sets.
2. Stimulate the net with a given input vector. 3.2.2. Determining the reference DMU of the inefficient
3. Calculate the Euclidean distance between the input and DMUs on each tier
each output node and select the output with the minimum Efficient DMUs in the upper tier become a reference set
distance. of inefficient DMUs in the lower tier. How can we select a
4. Update weights for the selected node and the nodes target reference DMU among DMUs in the reference set?
within its neighborhood. We use the SOM in advance to find a target reference DMU
5. Repeat from 2 until a stopping criterion is met. in the upper tier. DMUs, which have similar characteristics
Table 2
SI project efficiency ratings
SI project (DMU) Efficiency rating Reference set SI project (DMU) Efficiency rating Reference set
Table 5 Table 7
Clustering of SI projects by the tier analysis — tier 3 Training cases for C4.5
Group (tier) DMUs (SI projects) Reference set in tier 2 DMUs Variables
Table 6
Clustering of SI projects by the tier analysis — tier 4
5. Conclusion
Fig. 12. Rule induced from the decision tree. DEA is good at estimating the ‘‘relative’’ efficiency of a
294 H.K. Hong et al. / Expert Systems with Applications 16 (1999) 283–296
Table 8
The characteristics of each cluster
Cluster Count La (avg) Lb (avg) Lc (avg) Mr (avg) CSI (avg) Sp (avg) Bp (avg) Rew (avg)
DMU, it can tell us how well we are doing compared with DMUs can improve since a reference set of an inefficient
our peers but not compared with a ‘‘theoretical maximum’’. DMU consists of several efficient DMUs. Hence, we
Thus, to measure the efficiency of new DMU, we have to utilized a self-organizing map (SOM), which is one clus-
develop DEA previously with the same method with the tering tool for grouping similar DMUs according to
data of used DMU. Also, we cannot predict the efficient inputs, for the inefficient DMU to select one efficient
level of a new DMU. Second, because DMUs are directly DMU in a reference set as a benchmarking target. With
compared against a peer or combination of peers, DEA the tiers identified by the tier analysis, it could provide
offers no guide as to where a relatively inefficient DMU the guidelines for stepwise improvements of inefficient
improve. Third, DEA identifies peer DMUs and targets for DMUs.
inefficient DMUs, but it does not provide the stepwise path
In conventional DEA, it can only (1) identify inefficien-
for improving the efficiency of each inefficient DMU
cies, (2) identify comparable efficient units and (3) locate
considering the difference of efficiency. In order to over-
slack resources. But, we provide more information about
come this limitation of DEA, we suggest to a new
discriminant descriptors among input and output variables,
methodology that is found to be meaningful.
which affects the efficiency of DMUs, rules for classifying
The methodology we propose is a hybrid analysis and
new SI projects, and stepwise improvement path.
longitudinal analysis using machine learning, and it can be
We resolved the limitations of the DEA that are listed in
summarized as follows.
Section 1.
1. We apply a DEA to evaluate the efficiency of the DMUs 1. To evaluate a new SI project, the conventional DEA has
with their multidimensional inputs and outputs. After to be reapplied to all SI projects including the new one.
that, we clustered the DMUs together through the tier That is, because it measures the relative efficiencies of SI
analysis, which recursively apply the DEA to the remaining projects, after we group 50 SI projects together into four
inefficient DMUs, and then generated the DMU classifica- groups by the tier analysis, we can generate classification
tion rules using the C4.5, the decision tree classifier, with the rules using C4.5 in order to classify any new SI project
DMU tiers that had been identified by the tier analysis. without perturbing previously existing evaluation
2. DEA offers no guidelines where relatively inefficient structures.
Table 9
Input/output factors of DMUs on the improvement path
La Lb Lc Mr CSI Sp Bp Rew
2. The conventional DEA provide a reference set (multiple the Quality Control Group at Samsung Data System (SDS)
efficient SI projects) for each inefficient SI project. It for providing data and valuable advice.
cannot give a hint on where relatively inefficient
DMUs improve. But, since we utilize SOM as a tool
for clustering SI projects according to similarity of Appendix A. The DEA model
inputs, we can choose one reference project on the refer-
The CCR ratio model was proposed by Charnes et al.
ence set as a benchmarking target for each inefficient SI
(1978). In this model, the efficiency measure of any DMU
project.
is obtained as the maximum of a ratio of weighted outputs to
3. The conventional DEA cannot provide information about
weighted inputs subject to the condition that the similar
a continuous improvement path. It simply gives us infor-
ratios for every DMU be less than or equal to unity. That
mation about identification of inefficient SI projects and
is, the model is as follows:
slack variables via the reference set. We can resolve this
,
problem and provide information about continuous X X
improvement paths by using the SI project clusters by max e0 ur Yr0 vi Xi0
r僆R i僆I
SOM and reference project by tier analysis.
,
However, the present research has a number of limita- X X
tions. They can be also the topics for further research. s:t: ur Yrj vi Xrj ⱕ 1 ᭙ j 僆 N
i僆I i僆I
1. Obviously, environmental factors such as project ,
complexity, the quality of available hardware and soft- X
ware tools may also affect the efficiency of SI projects. ur vi Xi0 ⱖ e ᭙ r 僆 R
i僆I
Unfortunately, due to the unavailability of data, those
factors could not be included in this research. In future ,
X
DEA analyses, these factors may be incorporated into the vi vi Xi0 ⱖ e ᭙ i 僆 I
production model as exogenous (uncontrollable) inputs i僆I
or outputs, or as categorical variables.
where
2. The present model does not include project management
i index for input i, i 僆 I {1,2,…,I}
indexes (process indexes) such as the observance of
company-internal project procedure guidelines and of j index for DMU j, j 僆 N {1,2,…,n}
design review meetings, etc., which are determined in
r index for output r, r 僆 R {1,2,…,R}
the progress of the project. Experienced managers
knew that these variables are important to manage an vi virtual multiplier (weight) of ith input
SI project, yet they are not sure how these characteristics
vr virtual multiplier (weight) of rth output
should determine the overall quality of projects. In our
next study, we will apply the DEA technique to find how Xij the values (ⱖ 3 0) of input i for jth DMU
these process index variables influence performance (for j 1,…,n)
within the limited input resources.
Yrj the values (ⱖ 3 0) of output r for jth DMU
3. The data used in this study are from one large SI
(for j 1,…,n)
company in Korea. Caution should be taken in general-
izing the results of this study to other firms. e non-Archimedean infinitesimal
References
Acknowledgements Adolphson, et al. (1989). Railroad property valuation using data envelop-
ment analysis. Interfaces, 19, 18–26.
The authors would like to thank Dr Hyun Seok Jeong of Ahn, T. S. (1987). Efficiency and related issues in higher education: a data
296 H.K. Hong et al. / Expert Systems with Applications 16 (1999) 283–296
envelopment analysis approach. Ph.D. dissertation, The University of Fae, R., Grosskopf, S., Lindgen, B., & Roos, P. (1995). Productivity
Texas at Austin. developments in Swedish hospitals: a Malmquist output index
Athanassopoulos, A. D. (1997). Service quality and operating efficiency approach. In A. Charnes, W. W. Cooper, A. Y. Lewin, & L. M. Seiford
synergies for management control in the provision of financial services: (Eds.), Data envelopment analysis: theory, methodolgy, and applica-
evidence from Greek bank branches. European Journal of Operational tion. Boston: Kluwer Academic.
Research, 87, 300–313. Golany, B., & Roll, Y. (1989). An application procedure for DEA.
Banker, R. D., & Kemerer, C. F. (1992). Performance evaluation metrics OMEGA, 17 (3), 237–250.
for information systems development: a principal-agent model. Infor- Haag, S., Jaska, P., & Semple, J. (1992). Assessing the relative efficiency of
mation Systems Research, 379–398. agricultural production units in the Blackland Prairie, Texas. Applied
Banker, R. D., Charnes, A., & Cooper, W. W. (1984). Some models for Economics, 24, 559–565.
estimating technical and scale inefficiencies in data envelopment analy- Hjalmmarsson, J., & Odeck, J. (1996). Efficiency of trucks in road construc-
sis. Management Science, 30, 29–40. tion and maintenance: an evaluation with data envelopment analysis.
Beasley, J. E. (1990). Comparing university departments. OMEGA, 18 (2), Computers Operations Research, 23 (4), 393–404.
171–183. Lewin, A. Y., Morey, R., & Cook, T. (1982). Evaluating the administrative
Berry, M. J. A., & Linoff, G. (1997). Data mining techniques for marketing, efficiency of courts. OMEGA, 10, 404–411.
sales and customer support. New York: Wiley. Mahmood, M. A., Pettingell, K. J., & Shaskevich, A. I. (1996). Measuring
Brockett, P. L., Charnes, A., Cooper, W. W., & Sun, D. B. (1997). Data productivity of software projects: a data envelopment analysis
transformations in DEA cone ratio envelopment approaches for moni- approach. Decision Sciences, 27 (1), 56–77.
toring bank performances. European Journal of Operational Research, Malmquist, S. (1953). Index numbers and indifference surfaces. Trabajos
98, 250–268. de Estadistica, 4, 209–242.
Charnes, A., Cooper, W. W., & Rhodes, E. (1978). Measuring the efficiency Oral, M., Kettani, O., & Yolalan, R. (1992). An empirical study on analyz-
of decision making units. European Journal of Operational Research, ing productivity of bank branches. IIE Transactions, 24, 166–176.
2, 429–444. Pina, V., & Torres, L. (1992). Evaluating the efficiency of nonprofit orga-
Charnes, A., Clark, C. T., Cooper, W. W., & Golany, B. (1985). A devel- nizations: an application of data envelopment analysis to public health
opment study of data envelopment analysis in measuring the efficiency service. Financial Accountability and Management, 8, 213–224.
of maintenance units in the U.S. Air Force. Annals of Operations Quinlan, J. R. (1993). C4.5: programs for machine learning. San Mateo,
Research, 2, 95–112. CA: Morgan Kaufmann.
Charnes, A., Cooper, W. W., Golany, B., Phillips, F. Y., & Rousseau, J. J. Rutledge, R., Parsons, S., & Knaebel, R. (1995). Assessing hospital
(1994). A multi-period analysis of market segments and branch effi- efficiency over time: an empirical application of data envelopment
ciency in the competitive carbonated beverage industry. In A. Charnes, analysis. Journal of Information Technology Management, 6 (1), 13–
W. W. Cooper, A. Y. Lewin, & L. M. Seiford (Eds.), Data envelopment 23.
analysis: theory, methodolgy, and application. Boston: Kluwer Schaffnit, C., Rosen, D., & Paradi, J. C. (1997). Best practice analysis of
Academic. bank branches: an application of DEA in a large Canadian bank.
Christopher, et al. (1996). Software processes and project performance. European Journal of Operational Research, 87, 269–289.
Journal of Management Information System, 12 (3), 187–205. Schefczyk, M. (1993). Operational performance of airlines: an extension of
Clark, R. L. (1992). Evaluating USAF vehicle maintenance productivity traditional measurement paradigms. Strategic Management Journal,
over time: an application of data envelopment analysis. Decision 14, 301–317.
Sciences, 24, 376–384. Sestito, S., & Dillon, T. S. (1994). Automated knowledge acquisition.
Cook, W. D., Roll, Y., & Kazakov, A. (1990). A DEA model for measuring Prentice Hall.
the relative efficiency of highway maintenance patrols. INFOR, 28 (2), Sherman, H. D., & Ladino, G. (1995). Managing bank productivity using
113–124. data envelopment analysis (DEA). Interfaces, 25 (2), 60–73.
Cooper, W. W., Thomson, R. G., & Trall, R. M. (1996). Introduction: Thanassoulis, E. (1995). Assessing police forces in England and Wales
extensions and new developments in DEA. Annals of Operations using data envelopment analysis. European Journal of Operational
Research, 66, 3–45. Research, 87, 641–657.
Day, D. L., Lewin, A. Y., Salazar, R., & Li, H. (1994). Strategic leaders in Thanassoulis, E. (1996). A data envelopment analysis approach to cluster-
the U.S. brewing industry: a longitudinal analysis of outliers. In A. ing operating units for resource allocation purposes. Omega Interna-
Charnes, W. W. Cooper, A. Y. Lewin, & L. M. Seiford (Eds.), Data tional Journal of Management Science, 24 (4), 463–476.
envelopment analysis: theory, methodolgy, and application. Boston: Thompson, R. G., Dharmapala, P. S., & Rothenberg, L. J. (1996). DEA/AR
Kluwer Academic. efficiency and profitability of 14 major oil companies in U.S.
Day, D. L., Lewin, A. Y., & Li, H. (1995). Strategic leaders or strategic exploration and production. Computers Operations Research, 23 (4),
groups: a longitudinal data envelopment analysis of the U.S. brewing 357–373.
industry. European Journal of Operational Research, 80, 619–638. Thompson, R. G., Brinkmann, E. J., Dharmapala, P. S., & Gonzalez-Lima,
Drake, L., & Howcroft, B. (1994). Relative efficiency in the branch network M. D. (1997). DEA/AR profit ratio and sensitivity of 100 large U.S.
of a UK bank: an empirical study. OMEGA, 22 (1), 83–90. banks. European Journal of Operational Research, 98, 213–229.