Você está na página 1de 11

Data Envelopment Analysis, Operational Research and Uncertainty

Author(s): R. G. Dyson and E. A. Shale


Source: The Journal of the Operational Research Society, Vol. 61, No. 1, Part Special Issue
for 60th Anniversary of the Journal: Challenges for OR (Jan., 2010), pp. 25-34
Published by: Palgrave Macmillan Journals on behalf of the Operational Research Society
Stable URL: http://www.jstor.org/stable/40540225
Accessed: 06-07-2016 17:35 UTC

REFERENCES
Linked references are available on JSTOR for this article:
http://www.jstor.org/stable/40540225?seq=1&cid=pdf-reference#references_tab_contents
You may need to log in to JSTOR to access the linked references.

Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at
http://about.jstor.org/terms

JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted
digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about
JSTOR, please contact support@jstor.org.

Palgrave Macmillan Journals, Operational Research Society are collaborating with JSTOR to
digitize, preserve and extend access to The Journal of the Operational Research Society

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
Journal of the Operational Research Society (2010) 61, 25-34 2010 Operational Research Society Ltd. All rights reserved. 0160-5682/10 ^O

www.palgrave-journals.com/jors/

Data envelopment analysis, operational research and


uncertainty
RG Dyson* and EA Shale
University of Warwick, Coventry, UK

This paper discusses a number of applications of data envelopment analysis and the nature of uncertainty in
those applications. It then reviews the key approaches to handling uncertainty in data envelopment analysis
(DEA) (imprecise DEA, bootstrapping, Monte Carlo simulation and chance constrained DEA) and considers
their suitability for modelling the applications. The paper concludes with suggestions about the challenges
facing an operational research analyst in applying DEA in real- world situations.
Journal of the Operational Research Society (2010) 61, 25-34. doi:10.1057/jors.2009.145

Keywords: data envelopment analysis; uncertainty; Monte Carlo simulation

Introduction multiple inputs and outputs using linear programming and


called their approach data envelopment analysis - as the fron-
Data envelopment analysis (DEA) is a method for analysing
tier, formed by the efficient units, envelopes the inefficient
a set of organisational units in order to identify the efficient
units. In our work, we follow those approaches in the sense
units so that they can become benchmarks or peers for the
that we consider the situation of interest to be a set of existing
inefficient units in the set and in a cooperative system can
units and our concern is to identify the efficient ones, which
facilitate the spread of best practice. Examples of the units are
can then provide targets and benchmarks for the inefficient
schools, bank branches and retail outlets and they typically
units and can be sources of best practice. It is important to note
involve multiple inputs being converted into multiple outputs.
that not all efficient units will necessarily be selected as bench-
The DEA model was initially formulated as a deterministic
marks, that is sources of best practice. This approach is exem-
model, but there have been numerous developments that take
plified by the first application that the first author encountered
into account uncertainty including chance-constrained (CC)
working with Emmanuel Thanassoulis and John Foster who
DEA, imprecise DEA (IDEA), Monte Carlo simulation and
was then seconded to the (UK) Audit Commission. The appli-
bootstrapping. The authors have used DEA to analyse various
cation of concern was local government rates departments in
real-world situations involving, for example, local government
England (Thanassoulis et al, 1987; Dyson and Thanassoulis,
departments, bank branches, universities and public houses.
1988). Rates were at that time the local property tax. The
In these and other applications, DEA has been used determin-
efficiency of local government was the concern of the Audit
istically and any uncertainty in the situation has been handled
Commission and the then Deputy Director Ross Tristem, who
only implicitly or by sensitivity analysis. In this paper, we
return to several of the situations and examine the data and was in frequent contact with the Operational Research group
at Warwick, expressed his concern that in local government
the models in order to consider the nature of any uncertainty
there were views about which authorities were the most effi-
and how it might best have been brought explicitly into the
cient, but he was interested in exploring the possibility of a
analysis. Our interest lies in the interface between uncertainty
more objective approach and questioned whether DEA might
as met in practice and the theoretical tools proposed in the
literature. be an appropriate mechanism. The Audit Commission at that
time was charged with identifying and spreading best practice
The origins of DEA lie with Farrell (1957), who advo-
throughout the system of local government and, interestingly,
cated identifying an empirical efficient frontier, that is, one
explicitly agreed not to publish comparative tables of perfor-
formed by a set of real units and based on observed best prac-
mance (which have obsessed governments and DEA analysts
tice, rather than formulating a theoretical frontier based on a
in recent years). Their approach, therefore, was primarily
production function, fashionable at that time in economics.
developmental rather than the more summative stance preva-
Chames et al (1978) developed the approach to encompass
lent in the public sector today with the widespread publica-
tion of league tables and a name and shame mentality. (See
* Correspondence: RG Dyson, Operational Research and Management
Sciences, Warwick Business School, University of Warwick, Coventry, CV4 Amado and Dyson (2009) for a recent developmental study in
7 AU UK the public sector.) The Audit Commission at that time sought
E-mail: r.g.dyson@warwick.ac.uk best practice, studied the mechanisms delivering it and then

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
26 Journal of the Operational Research Society Vol. 61, No. 1

aimed to spread it throughout the system by the use of best dimensions. It would therefore be reasonable to assume that

practice guidelines. This is entirely consistent with the oper- there was a high chance (say 95%) of the true score being
ational research (OR) approach of studying real situations, within two points either side of the recorded score. From this
engaging with the key personnel and developing methods for information, we could construct a normal distribution on the
improvement as developed by Blackett (1950). Risk analysis, data subjectively. At first sight, this seems problematic at the
an area we will make appeals to later in this paper, also has top end where a recorded score of 24 might be given but we
a history of involving managers. This engagement is a vital could either use a different distribution at the extremity or
source of information for model building and engenders confi- simply assume that the underlying continuum went beyond 24.
dence in the results of the analysis In this way, a deterministic teaching rating could be replaced
This paper proceeds as follows. In the next section, by a subjective probability distribution as in the risk analysis
we examine four situations where DEA has been applied, approach (Hertz, 1964). With only a general understanding of
consider the nature of any sources of uncertainty in the data the nature of the uncertainty, the recorded score would have
and illustrate some of the ways that the nature of that uncer- to be taken as the mean value (or mode for a skewed distribu-
tainty would impact on the efficiency scores of the units. We tion). If, however, underlying information about the construc-
then review various approaches to dealing with uncertainty tion of the recorded score were available, it may be that the
in DEA in the literature and consider to what extent they are true mean would be estimated to be a little lower or a little

appropriate vehicles for handling the situations we outline. higher than the recorded score, in which case that value could
Finally, we discuss the challenges facing the use of DEA be used. Also the degree of uncertainty could conceivably be
under uncertainty from an OR perspective. different for different departments. This however would need
detailed information about the assessments of every depart-
ment which was not available at the time.
The nature of data in DEA applications
The approach to assessing the quality of research was done
In this section, we consider four situations where DEA has
through a national research assessment exercise (RAE) where
been applied but focus our attention on the data itself and the
the research quality of individuals in each department was
nature of any uncertainty in that data. The examples selected
assessed by consideration principally of their output in terms
come from cases where there was no shortage of data so that
of journal articles and books. In determining a departmental
the issues discussed are as far as possible not driven by lack score some account would also have been taken of the number
of information.
of doctoral students, the value of research grants and contracts
and the esteem of the department. The research quality of the
UK universities
department was given a rating on a scale 1-7 (actually, 1, 2,
The first example relates to the assessment of universities, 3(a), 3(b), 4, 5 and 5* were used) where the highest value was
see Athanassopoulos and Shale (1997), Sarrico et al (1997) intended to describe a department in which the majority of
and Sarrico and Dyson (2000). In the latter two papers, DEA the output was of international standard and at the bottom end
models were used to assess the relative efficiency of depart- the research was of sub-national standard. The most recent

ments in universities, for example physics, business studies, RAE has adopted a different scoring mechanism although
across the UK university sector. We will concentrate on the the methodology is broadly similar. Given the thoroughness
model in Sarrico et al (1997), which was from the perspective of the assessment process it would have been unlikely that
of the students prior to identifying their preferred universities, errors of more than one point in either direction would be
updated with the types of data used by Sarrico and Dyson made, so again a subjective symmetrical distribution can
(2000). The outputs were a measure of teaching quality, a be constructed around the recorded score. The seven-point
measure of research quality, the employment prospects of scale and the descriptors used suggest that the assumption of
students, the amount of accommodation on the campus, the an interval scale is not unrealistic.

cheapness of living at the university and the expenditure on The employability of the students was measured by the
libraries. A nominal input was used as the outputs were not percentage in employment or further training 6 months after
volume related. At that time the measurement of the quality graduation. The uncertainty in the data here relates to the
of teaching in the universities was by direct assessment with a fact that not all students could be contacted after 6 months

team of assessors spending 2-3 days with a university depart- and the information was self-reported. The self-reporting can
ment, sitting in on lectures, examining materials, looking lead to a bias as it is understood that graduates in a job
at assessment schemes and so on. The assessment had four are more likely to respond to an enquiry than those unem-
dimensions and up to six points could be awarded on each, ployed. Consequently, it was felt that an optimistic estimate
giving a potential range of scores from 0 to 24. There was of the true score would be 2% higher than the recorded score
a well-developed and consistent process for the assessment but a pessimistic one would be 5% lower. Here, we would
that the different teams complied with. The result of this was have an asymmetric distribution which could be represented
that it was rare for scores to be considered to be incorrect by say a lognormal or beta distribution. In this analysis, a
by more than a single point on each of the four different key perspective was that of the student seeking to apply to

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
RG Dyson and EA Shale - Data envelopment analysis, operational research and uncertainty 27

the university, and for accommodation the assumption was that the distress warrants measure is in fact a surrogate for
that a higher percentage of university-owned accommodation some measure of difficulty in collecting rates due to socio-
made the university more attractive. As the university would economic conditions such as a highly mobile population and
know the addresses of their students the measure would have this issue of surrogate measures will be considered in more
a relatively small error associated with it. A further measure detail in the next example.
was the cost of living at a particular university based on the
average debt of graduates of that university. This was felt to Public houses
be important, at least for many students, and London was
known to be rather expensive. The actual measure used was The second author led an investigation into the efficiency of
the difference between the maximum average debt at any a chain of branded public houses (Shale, 1996, 2009). The
university and the average debt of that particular university, inputs here were labour costs, population within a 5 mile
so that the one with the lowest debt would have the highest radius of the public house (pub) and weighted managerial
score. As the level of debt might well depend on the demo- talent. The outputs were the 'wet' and 'dry' take, that is the
graphics of the student population as well as the cost of living revenue from beverage and food sales, and a proxy measure
this measure would have a relatively high degree of uncer- of the attractiveness of the location. The labour cost data

tainty. Finally, the library expenditure per student was seen as were held in two locations collected by different mechanisms
a surrogate for academic support costs, but this was not avail- and for separate purposes, accountancy and human resources
able for the entire sector and so was not always used in the (HR). The latter function was responsible for the company's
analysis. No explicit measure of staff resources was used as profit related pay scheme and the control of labour costs was
it was assumed that the level of resource would be reflected a major contributor to that scheme. These data can therefore
in the teaching and research scores. The assumption here is be expected to be accurate as there is a clear validation mech-
that from a student perspective more means better, but from anism available. The population within a 5 mile radius was
an institutional perspective, this might hide inefficiencies in chosen as a measure of potential custom. The use of the entire
the use of resources - Sarrico and Dyson (2000) take an insti- population was not a compromise but actively selected due
tutional perspective. to the marketing of facilities for all ages from 'toddlers' to
It can be noted from the previous discussion that the data senior citizens. The radius was chosen as a result of statis-

points could have been replaced by subjective probability tical analysis but confirmed by exit polls that showed that
density functions, rather than, for example, simply ranges and 80% of customers travelled 5 miles or less. There remain

that some of the measures exhibited relatively high degrees of futher sources of uncertainty in these data however. Using
uncertainty whereas others such as the percentage of univer- these data as a measure of potential custom assumes a homo-
sity accommodation exhibited a relatively low degree. This geneous response in terms of leisure preference and dispos-
phenomenon can have profound implications for the efficiency able income across all the locations and ignores potential
scores and this issue is considered in the section on approaches access barriers such as major trunk routes which are known
to uncertainty. to modify consumption behaviour. Furthermore, as these data
are based on census returns demographic changes have to
be estimated in the years between the census returns by the
Rates departments
database provider. With respect to the original data, the UK
In the application to rates departments (see the Introduction census includes statements on coverage and completeness
section for context and references), there was a single input of returns (Office for National Statistics, 2005). This type
which was the cost or budget of the department and three of measure can be categorised as both environmental and a
outputs, one relating to the number of households from which proxy or surrogate for the latent variable of demand for these
the rates or taxes were to be collected, a second relating to services if delivered in the specified location.
the number of businesses and a third relating to the number The weighted managerial talent measure was constructed
of distress warrants issued. This latter measure was one of the in different ways in the two studies but there were signifi-
steps taken in seeking payment from recalcitrant rate payers cant similarities. In both cases, the measure was constructed
and was included as it was clear that those rate payers would as the product of the floor area available for the genera-
consume significantly more resource. Here again it would be tion of revenue and an index designed to capture the talent
reasonable to assume that errors in the data could be repre- of the manager for running this type of outlet (the use of
sented by subjective distributions but the distress warrants floor area was necessary to convert the talent index into a
measure presents an interesting complexity. The issue here is volume measure for compatibility with the other measures).
that the actual data itself would be known with a consider- The floor area was estimated from the development blueprint
able degree of accuracy; however, the uncertainty lies more and although there may have been some layout changes this
in the impact on the department of pursuing recalcitrants and would have a minor impact. The managerial talent measure
this might vary considerably from individual to individual and was much more open to debate. Initially, a qualitative judge-
indeed from department to department. It could be argued ment was made based on all available evidence by the HR

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
28 Journal of the Operational Research Society Vol. 61, No. 1

department and the area managers. This was accepted within known because of the regulations surrounding the booking of
the organisation to be subject to some bias, profitability of an flights. In the minority (for this type of business) of book-
outlet sometimes being seen as 'proof of talent. The company ings without flights or other travel, insurance issues should
invested in a well-respected psychometric profiling exercise confirm numbers.

run by an internationally recognised external provider. This The model contained 1 1 inputs which will be reviewed here
removed the potential for bias and within the exercise itself in order of increasing data uncertainty. The number of service
reduced the uncertainty of the numeric measure which was points in an agency is known and layout or other changes can
a count of desirable responses to a battery of 108 ques- easily be incorporated as non-integer values for the period of
tions. The greatest remaining source of uncertainty is whether analysis. Sales, support and managerial staff were recorded
talent can be measured in this way. To deny this type of in the model as full time equivalents. It is expected that these
exercise on principle is possibly to denigrate psychometric would be known very accurately as employment contracts are
methods as a whole. Such a debate is beyond the scope of this legal documents. There might be issues of absence, overtime
paper. or the number of employees varying across the period of anal-
The revenue data were held in the company's accounting ysis but these could be resolved by using numbers of hours
database and derived directly from electronic point of sale salaried. Branch costs (excluding payroll) were taken from the
data. Although some errors are inevitable given the number of accountancy system and thus were expected to be consistent
'keystrokes' the tills receive and the potential for till malfunc- within the limits of such a system. The average industry expe-
tion, this type of data is subject to intense scrutiny, routine rience of staff in years might be subject to some doubt about
cross-checking and cleaning and can therefore be expected candidates' exaggerating their experience before joining the
(in the absence of undetected fraud) to be accurate. company - otherwise it would be accurate - what is uncertain
The final measure chosen (population within 1 mile) was is the relationship between experience and generating sales.
intended to capture the attractiveness of the site; the company The floor space dedicated to customer facing tasks is known
thinking was in terms of 'aspirational' locations. Irrespec- but not every square metre will be equally easy to use just as
tive of whether the public houses were located in a rural or every member of staff is not equally effective. This is not a
semi-rural position, adjacent to park land or open water, or serious problem even though the DEA model works on the
in an upmarket low density housing area, in all these cases assumption of equality. If a unit is found to be inefficient then
the population in a 1 mile radius was low. This is a quanti- a review of the quality as well as the quantity of the resources
tative proxy measure for a qualitative aspect of the location. is a natural managerial reaction to the assessment.
Its appropriateness is context dependent. As a data item, it is The remaining four inputs are external, environmental
subject to some of the same issues as the population figures and proxy measures subject to far greater uncertainty.
discussed above. However, from the uncertainty perspective Market potential was included by using the population at the
it is its nature as a proxy that is the source of greatest doubt. 'town' level; the company's preferred measure as the outlets
The precise relationship between the measure used and the were located on high streets and in shopping centres that
concept of an 'aspirational' location cannot be fully known as drew customers from a wide radius. This is subject to the
any data depending upon perception, such as attractiveness, same provisos as the similar figures in the pubs case. This
must be subject to uncertainty. There is a significant inter- 'potential' was complemented firstly by a count of travel
disciplinary literature on the measurement of perceptions and agents in the town as a measure of competitive pressure,
expectations, for example Bitner (1990) looked at the effect although there is again a homogeneity of effect issue, and
of the physical surroundings on the perception of service secondly by a measure of average household spend taken
quality. Attractiveness might seem to be naturally an input from an external database, an indicator of wealth and hence
but the measure used meant that more - higher density of potential for spending disposable income on travel. Finally,
population - was less attractive so was designated an output in there was a variable intended to capture the quality of the
the model. Using an anti-isotonic input variable as an output retail location itself. In effect this recorded how far the

is a technical issue discussed in Shale (1996) and Dyson agent was from the nearest Marks and Spencer (M&S) store!
et al (2001). Although this measure did not appeal to the analysts involved,
the company management were absolutely convinced it
worked on the basis of their knowledge of M&S's location
Travel agencies
decisions and the similarity of the customer bases of the two
A larger scale DEA exercise involved a chain of travel agents. companies.
Their core business included the sale of flights and 'package' Our experience of DEA model building and knowledge of
holidays and it is this aspect of the operation that was anal- the literature suggests that there are generic data categories
ysed. Two outputs were used and in both cases the data was and sources of uncertainty that may have wide applicability.
expected to be accurate. The revenue generated was extracted We have summarised some of the most common of these and

from the accountancy system and subject to all the usual their characteristics in Table 1 , although the categories are not
checks of audit. The number of customers would be accurately exclusive or exhaustive.

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
RG Dyson and EA Shale - Data envelopment analysis, operational research and uncertainty 29

Table 1 Data uncertainty, sources and effects

Type of data/uncertainty Sources of uncertainty Effects and characteristics


Measurement error Human error or technical Symmetric if random error but skewed under bias. Large errors
malfunction or systematic bias likely to be detected by routine data cleaning.

Judgmental data Subjective nature of the quan- A highly variable category. Undertaken by a team of experts
tification of the measure using a carefully controlled process subject to wide-ranging
(perhaps public) scrutiny the measure is likely to be accurate,
with limited uncertainty. Individual or small group assessment
using no formal methodology and not subject to further review
may be highly uncertain.

Environmental or external data Complexity, quality control Environmental factors may be required to summarise a complex
and timeliness situation in a single or few numbers, for example a count of
the number of competing establishments does not necessarily
capture the strength of the competitive effect. By its nature a
summary omits the fine detail and is hence a potential source
of uncertainty. Data sourced from external providers potentially
has quality issues outside the control of the user. The review
and updating of the data may not coincide with the internally
generated data leading to timeframe mismatch.

Self-reported data Behavioural bias Non-response bias and behavioural side-effects when certain
responses are perceived to be more desirable.

Psychological data for example Effectiveness of measurement The analytical response to this type of data varies widely by
perceptions/expectations and instruments and behavioural discipline. It is the bedrock of the work of psychophysicists
psychological profiling bias but is subject to criticism by scientists trained in the classical
paradigm. Accepting that people are difficult to quantify but
that it is essential to do so for the realistic DEA of a social
system, the focus needs to be on the design of the data collection
and identification of sources of uncertainty and bias within it.

Proxy/surrogate data Indirect measurement Here the data used is either a substitute for a measurable
but unrecorded factor or is representing a latent or unmeasur-
able quantity. The former is regrettable and may potentially be
addressed in the longer term. The onus in both cases is on the
analyst to understand as far as possible the relationship between
the data possessed and that desired. There is a considerable
literature on latent variables (in many disciplines).

In summary, from our approach in the above cases and Approaches to uncertainty in DEA
in other real-world applications, we conclude the following
about the nature of data uncertainty in DEA. There may be In this section we introduce and discuss a number of

uncertainty in any of the inputs and outputs. We would not approaches to handling uncertainty in frontier analysis and
necessarily expect to be able to describe this from panel DEA. We focus only on the basic methods, as our aim is
data as suggested, for example, by Kao and Liu (2009), as to differentiate and compare the main approaches. We do
we could not always assume stationarity through time due not attempt to survey the entire set of approaches and their
to learning effects and technological change. Where data variations.

exhibit uncertainty we would envisage both symmetrical A well-known approach to determining an efficient frontier
and skewed distributions, often in the same application, and under uncertainty is stochastic frontier analysis (SFA) (Aigner
these would typically be constructed subjectively. For an et al, 1911 ' Meeusen and van Den Broeck, 1977). Here the
approach to constructing subjective distributions see, for frontier is not an empirical one as in DEA. It is assumed
example, Hertz and Thomas (1983). Our units are character- that the true/theoretical efficient frontier is generated by a
ized by a common set of inputs and outputs but they are not model linking a single output to multiple inputs. There are
necessarily conforming to a common production function. two sources of error, one, often represented by a half-normal
A clear example of multiple production functions would be or exponential distribution, attributed to controllable factors
that described in Chames et al (1981) where they analysed leading to inefficiencies and thus less than optimal output,
two different educational programmes. and a second, assumed to be normally distributed, due to

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
30 Journal of the Operational Research Society Vol. 61, No. 1

Ac Ac "

10 10 A A
w d

5 D* #F 1
,,-^m J. '
5 10 _ -
5 1 ExC _ ExCu -

Figure 1 The DEA efficient frontier. Figure 2 Imprecise DEA.

Ac ''
uncontrollable factors often external to the unit. To specify the
stochastic frontier it is necessary to specify a functional form 10 A A'
for the production frontier and assume distributional forms
for the two error terms. A misspecification of the functional
form can lead to errors in the efficiencies.

The original formulations assumed a single output; later


5 D# p ^W C
developments address multiple outputs in the model using
distance functions at the cost of some complexity (Fare and
Grosskopf, 1990). The approach does not consider errors in
the inputs and outputs per se but can include them only as
5 10 ExCu
a part of the uncontrollable error term. However from our
perspective, we consider that the distributional assumptions Figure 3 Imprecise DEA.
required, the lack of focus on the uncertainty in individual
inputs and outputs, the difficulty in handling multiple inputs
and outputs and the focus on a theoretical frontier make the outperform the target unit. An example of such a composite
approach unattractive. unit would be H' in Figure 1 .
IDEA was first introduced by Cooper et al (1999) and If the output levels exhibit uncertainty and as in IDEA the
covered bounded and ordinal data. We will concentrate here uncertainty is captured by ranges then we have the situation in
on illustrating and discussing the use of bounded data (ranges) Figure 2. In the example, it is assumed that the uncertainty in
to describe any uncertainty using the simplified formulation of the academic measure is relatively low compared to the uncer-
Zhu (2003). Figure 1 illustrates the basic DEA model for a set tainty in the extracurricular measure. If we focus on school G
of schools where the two outputs are academic achievement which was just underperforming in DEA, in IDEA not only
per pupil and achievement in extra curricular activities such do we allow the weights to favour G, but also the values of
as sport and the performing arts. School A is a high academic the outputs. To achieve that, the outputs of G are allowed to
achiever but is perhaps over focussed on exam performance move to their highest levels in the ranges (denoted by G') and
and neglects extra curricular activities as a result. School I the outputs of the other schools forced to their lowest levels
on the other hand focuses on performance at the expense of (eg denoted by a for school A). The frontier for G is now
academic subjects. DEA allows units to weight their outputs (a, G', i), and G is now efficient. Also with B and C taking
(and inputs) favourably so that the basic model would identify on their lowest values they become just inefficient against G's
A, B, C and I as all being efficient and defining the efficient frontier. In general, the additional flexibility of IDEA will
frontier. The model would also identify FT as defining target thus allow some inefficient units to become efficient.

output levels for the underperforming school H, and C and When we focus on each school in turn, IDEA will poten-
I are peer schools as they have the closest output mix to H tially produce a frontier peculiar to each unit. For example,
among the efficient schools. Best practice from C and I could the efficient frontier for A is (A', c, i), see Figure 3, and
potentially be transferred to H to improve its performance. C, which regains its efficiency has a frontier of (a, C) with
Importantly, for the following discussion DEA identifies a I now being inefficient. Schools A, B, C, G and I will
unique efficient frontier. This is achieved conceptually either each have their own frontier under IDEA and be efficient

by finding the weights on the inputs and outputs showing a whilst the other schools will all be enveloped by the frontier
target unit in its most favourable light or constructing a linear (a, c, i). They will still be inefficient but have a higher effi-
combination of the other units to see if a composite unit can ciency score than under DEA. In general, IDEA will yield

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
RG Dyson and EA Shale - Data envelopment analysis, operational research and uncertainty 31

Ac -
Cu I a
10
10 S^^5-^ a

F G ' ' CC (0.05)


# ' '
/ ' '
J ! '
5 10 ExC
5 10 ExC
Figure 5 Monte Carlo simulation.
Figure 4 Chance-constrained DEA.

a higher number of efficient units than the standard DEA are interested in using real-data to examine the impact of
model. uncertainty on the identification of best practice. An example
The use of ranges appears equivalent to imposing a uniform of perturbing observed data to examine the effect upon DEA
distribution on the uncertain parameters, that is from a efficiency estimation can be found in Kao and Liu (2009).
managerial perspective there is no further information on the This approach circumvents the intractability of many analyt-
likely value within the range. It should be noted also that the ical methods by employing simulation where the uncertainty
IDEA approach is effectively selecting only extreme values of in the data can be described by any discrete or continuous
these ranges although the extent of the ranges should reflect distribution. As it can involve solving large numbers of DEA
different degrees of uncertainty in the inputs and outputs. models to obtain approximations to the distributions of the
IDEA is thus perhaps more akin to sensitivity analysis than efficiencies it can be computationally greedy. However, there
to stochastic analysis. However, it could identify the more are no restrictions in principle on the nature of the uncer-
robust efficient units and would therefore be more helpful tainty affecting the variables. Kao and Liu (2009) advocate
for benchmarking and identifying best practice than DEA. the use of several years of data to construct distributions of
In CC DEA (Land et al, 1988; Olesen and Petersen, 1995; the uncertain variables and illustrate the advantage of this
Cooper et al, 1996) the typical assumption is that the inputs approach over using averages. They found that in their bank,
are deterministic and the outputs are stochastic described for example, 2000 replications were sufficient to produce a
by a joint distribution, usually the normal distribution for result close to the true distributions of the efficiencies. Apart
tractability. The uncertainty is handled by allowing a unit to from any computational difficulties there is the concern that
lie outside the efficient frontier with a pre-determined prob- the processes are unlikely to be stable over time due to
ability. The CC DEA approach converts the CC models to changes in technologies and learning effects, particularly if
deterministic equivalents which can be solved to determine the aim of any analysis is to discover and spread best practice
a new efficient frontier enveloping the original DEA frontier. across the system. This does raise starkly the contrast between
In Figure 4, the ellipse around unit B is an equi-probability summative and formative approaches. In a summative culture,
contour of unspecified probability level representing the joint the best performers are rewarded and there is no incentive
distribution of B. Note that, due to the convexity of the DEA (or even a disincentive) to share best practice. An alternative
frontier, there will be a greater than a 0.5 chance of the true to using panel data is to construct subjective distributions of
value of B lying outside the efficient frontier. In CC DEA, we the uncertain variables as advocated in risk analysis (Hertz,
wish to locate an efficient frontier such that an efficient unit 1964). In this approach, experts are interrogated about the
will have only a small chance 0.05 of falling outside the fron- uncertainty in the variables providing, for example, esti-
tier. This is illustrated in Figure 4 by the frontier labelled CC mates of most likely, optimistic and pessimistic values.
(0.05), for example. Olesen (2006) shows how to construct CC In this way, parameters can be constructed and often a
frontiers where there are both stochastic inputs and outputs. beta distribution is assumed where the estimates indicate a
The approach can soon become intractable if it is desired to skewed distribution. This simulation approach is illustrated in
incorporate uncertainty in the inputs as well as the outputs Figure 5.
and if a mix of distributions is required. The distribution of the outputs of each unit is represented
Monte Carlo simulation has been advocated by a number by equi-probability contours and just two are shown. We will
of researchers as a means of incorporating uncertainty in data assume for illustrative purposes that in each replication the
into DEA (Banker et al (1987) being an early example) but output values are either at the mean or on one of the contours.
almost all studies use the approach with artificially generated In the simulation, the frontier with each unit taking on its
data to compare techniques and/or look at the accuracy of the mean output values will have the highest likelihood of occur-
generated efficiency scores. That is not our arena here - we rence and this is represented by the thickest line A, B, C and I.

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
32 Journal of the Operational Research Society Vol. 61, No. 1

A replication joining points on the outer contours will be the Ac-

least likely and shown by the thinnest lines. One joining a


io :A
mix of mean value and contour points will have an interme-
diate thickness. Each replication will give an estimate for the
frontier and a distribution of the efficiency of each unit will
be generated. It can be seen from the figure that the DEA
efficient unit I will often be dominated by C and in general
units deemed efficient because of their high score under the ; / yy/^ i '
measure with low degree of uncertainty, for example school
A, are stable in terms of their efficiency score, whereas for /< ^. - - - " **

units on the efficient frontier due to high scores on the measure 5 10 ExC
exhibiting high variability, for example school I, the chances
of them remaining consistently on the efficient frontier are Figure 6 Bootstrapped confidence intervals.
much reduced.

Simar and Wilson (1998) first advocated bootstrapping as


a way of capturing the uncertainty in DEA via constructing about uncertainty. Additionally, the homogeneity assumption
confidence intervals about the efficiency scores. It is assumed on efficiency makes it impossible to deal with the very real
that there is a potentially infinite population of units corre- instance of some observations being more uncertain than
sponding to a data generating process and the current set others.

is simply one sample from the population. There has been


development in the area of consistent estimators and we will
Applicability of the approaches
concentrate on these. Simar and Wilson (2008) provide an
overview. One such approach is based on repeatedly sub- In this section, we return to the nature of uncertainty in the
sampling the set of observations, the size of the sub-samples types of applications we have encountered and consider the
being critical to adequate estimation of confidence intervals applicability of the various approaches. We reiterate that
on the efficiency scores. However, Kneip et al (2008, p. 1664) we are interested in applying DEA in a formative way to
state there is 'no reliable way of determining a reasonable identify and spread best practice. Additionally, the units do
value of the sub-sample size in applied settings.' They propose not necessarily conform to a single production function. The
instead adopting an alternative bootstrap approach smoothing uncertainty of interest relates to the inputs and outputs of the
both the distribution of the observations and the initial esti- units and may exist due to measurement errors, uncertainty
mate of the frontier. They argue that the two smoothing param- in external data, perceptions and proxy measures. There
eters required can be selected by logical means but the method is typically no data to produce frequency distributions and
is considerably more complex than the original sub-sampling the approach to capturing the uncertainty would be through
approach particularly in the multi-input, multi-output case. management engagement and the construction of subjec-
This is well illustrated in Simar and Wilson (2008). In order tive probability distributions which could be skewed. This
to render the method more tractable, they suggest a simpli- approach has support within the DEA literature, see, for
fying assumption, namely that the distribution of efficiency is example, the arguments of Stewart (1996). The uncertainty
homogeneous across input-output space - a potentially very may also be unit specific rather than system wide.
restrictive assumption. In essence, the method generates arti- First of all, we repeat our concerns about the efficacy of
ficial data from the observations in order to create the multiple SFA in our contexts. These relate to the distributional assump-
bootstrap samples required. Full details are given in Simar tions required, the lack of focus on the uncertainty in indi-
and Wilson (2008). Constructing the symmetric confidence vidual inputs and outputs, the difficulty in handling multiple
interval about the DEA frontier relies upon applying the one- inputs and outputs and the emphasis on a single theoretical
sided variability observed in the bootstrapped frontiers lying production function rather than one or more empirical fron-
within the DEA frontier outside that frontier also. These confi- tiers.

dence intervals are illustrated in Figure 6 where the thick lines IDEA would be a possible approach, as it focuses on
on the rays from the origin represent the confidence intervals. actual units and allows uncertainty in the inputs and outputs.
They show, for example, that the DEA inefficient unit G could Also it makes no assumptions about an underlying produc-
outperform C and C could outperform the efficient unit I. tion function and could accommodate the different potential
This method enables the analyst to give a unit (ray) specific sources of uncertainty. It would therefore be consistent with
estimate of the uncertainty in the DEA efficiency scores and our 'spreading best practice' approach. The approach would
can identify potentially robust efficient units. However, it have two disadvantages however. Firstly, capturing the uncer-
cannot include any specific information on data uncertainty tainty by ranges is equivalent to specifying a uniform distri-
in the observed values of inputs and outputs. It subsumes bution with equal probabilities within the ranges and zero
any particular source of variability into a general statement elsewhere. We believe that in many situations more realistic

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
RG Dyson and EA Shale - Data envelopment analysis, operational research and uncertainty 33

probability density functions can be constructed. Secondly, the ible and computationally less cumbersome. The drawbacks or
optimisation procedure will move the target unit to the most challenges relate to constructing the distributions, but there is
favourable data point and the others to the least favourable so a wealth of practice on that issue, and the potential computa-
that the approach has a second process in favour of the target tional greediness of the approach.
unit (the first being the choice of weights). This may be over-
compensation and will lead to a reduction in discrimination.
Conclusions and challenges
Note that there is now no longer a unique frontier but poten-
tially each unit has its own frontier. There will also of course Our tentative conclusions on the various approaches to
be no distribution of efficiency for each unit. However, IDEA handling uncertainty in DEA are as follows. Given our focus
can nevertheless be valuable in identifying robustly efficient on real units and best practice, with the sources of uncer-
units which will be superior candidates for spreading best tainty in both inputs and outputs of those units, we are not
practice. attracted to bootstrapping, principally due to its approach
CC DEA focuses on the uncertainty in the inputs and to uncertainty and to the simplifying assumptions necessary
outputs although typically at a system level and again is to make the approach tractable. IDEA does focus on real
concerned only with the real set of units. The approach units and incorporates uncertainty on individual inputs and
involves specifying new frontiers enveloping the original one outputs with the ranges being constructed subjectively. The
and as with IDEA may identify the more robustly efficient IDEA analysis can identify robust efficient units which may
units as the frontiers will be closest together where the vari- be examples of best practice, but cannot provide informa-
ability in efficiency is the lowest. The approach can become tion on the distribution of efficiencies. It is computationally
intractable where there are uncertainties in inputs and outputs feasible. CC DEA, like IDEA focuses on real units and
and the normality assumption is relaxed. For our purpose, uncertainty in individual inputs and outputs. It could handle
it is also not clear to what extent it adds value beyond that subjective distributions. However, it can become intractable
obtained by IDEA. where there are skewed distributions and uncertainties in

Bootstrapping produces confidence limits on the efficien- both inputs and outputs. It could identify robust efficient
cies of the units thus attempting to capture the true efficient units but it is not clear whether the added complexity over
frontier within the specified interval. The units with the confi- IDEA gives significant added value. Monte Carlo simulation
dence limits of highest efficiency could thus be examples of appears to have all the flexibility necessary for the situations
best practice. There are however two key drawbacks from with its only disadvantage being that it is computation-
our perspective. The sources of uncertainty are at the unit ally greedy. We would thus consider IDEA as a relatively
level and inferred implicitly from the distribution of the units simple and useful step in the direction of incorporating
in input output space, rather than addressed explicitly at the uncertainty into DEA and Monte Carlo simulation as the
individual inputs and outputs, and the assumptions necessary most flexible and comprehensive approach in terms of our
to make the approach tractable, in particular the assumption requirements.
of homogeneity of efficiency across the input output space, Finally, in this 60th anniversary issue of JORS, we would
could be unrealistic. Inefficiency across the space would not like to look forward by reflecting upon two more general
be homogeneous if operating practices within an organisation questions related to our foregoing discussion. What are some
are not equally well developed, that is mature across all the of the challenges facing those interested in applying DEA but
functions captured in the model. aware of the tension between a deterministic technique and an
Having indicated the drawbacks of the previous approaches uncertain world? Stepping back from the detail of our partic-
we can now assess the potential of the Monte Carlo simulation ular topic and field, what generic issues have been revealed
approach - the last, but often effective, resort of the stochastic by the investigation? Firstly, we suggest that there has been
modeller. The approach focuses on the real set of units and too much emphasis on the 'research' element of OR and too
makes no underlying assumptions about a production func- little on the 'operational'. Models and model extensions are
tion. In principle, it transforms the best information available being published because we can develop the mathematical
into subjective probability distributions with the uncertainty and statistical tools. At the extreme this becomes a branch of
on a particular input or output systematic across the set of applied mathematics or perhaps not even particularly applied.
units or peculiar to an individual unit. The replications of the Real-world, organisational issues remain unresolved or any
simulation will lead to a frequency distribution for the effi- resolution provided by OR practitioners goes unpublished.
cient frontier and there will be a distribution of efficiency We do not believe that this is unique to DEA but is a general
scores for each unit. malaise of OR, nor is it a new problem but that does not
Thus, there will be no strict division between efficient and render it less important (Ackoff, 1979). The debate over the
inefficient units but the units with the greatest concentration definition and role of OR continues in the 21st century. WW
at the highest efficiency score would be candidates for best Cooper argued for application-driven theory in his Blackett
practice. Desai et al (2005) compare CC DEA and Monte lecture to the Operational Research Society (Cooper, 1999).
Carlo simulation and conclude that the latter is more flex- The operational researcher applying DEA needs to engage

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms
34 Journal of the Operational Research Society Vol. 61, No. 1

both with the operational context and the nature and source of Dyson RG, Camanho AS, Podinovski VV, Sarrico CS and Shale
the uncertainties in order to realise the better in The Science EA (2001). Pitfalls and protocols in DEA. Eur J Opl Res 132:
245-259.
of Better'. There seem to be too many models whose sole
Fare R and Grosskopf S (1990). A distance function approach to price
published use is illustrated with unrealistic data and with little efficiency. J Public Econ 43(1): 123-126.
discussion of the practical significance of the development. Farrell MJ (1957). The measurement of production efficiency (with
It would be interesting if mainstream OR journals required discussion). J R Stat Soc Ser A 20: 253-281.

from papers, upon submission, not only an indication of appli- Hertz DB (1964). Risk analysis in capital investment. Harvard
Business Review 42: 95-106.
cability (currently often cursory in the extreme) but paired
Hertz DB and Thomas H (1983). Risk Analysis and its Applications.
submissions; one paper outlining the theoretical development Wiley: Chichester.
perhaps with a simple example and a second paper illus- Kao C and Liu S-T (2009). Stochastic data envelopment analysis in
trating its use in a realistic scenario. We understand that this measuring the efficiency of Taiwan commercial banks. Eur J Opl
Res 196: 312-322.
might be impractical but it would certainly be a challenge
Kneip A, Simar L and Wilson PW (2008). Asymptotics and consistent
for the OR community. It might require greater collaboration
bootstraps for DEA estimators in nonparametric frontier models.
between academics and practitioners or greater involvement Econ Theory 24: 1663-1697.
of academics with practice. We do not view either of these as Land KC, Lovell CAK and Thore S (1988). Chance-constrained data
undesirable developments. envelopment analysis. Managerial Decision Econ 14: 541-554.
Meeusen W and van Den Broeck J (1977). Efficiency estimation
from Cobb-Douglas production functions with composed error.
References International Economic Review 18: 435-444.
Office for National Statistics (2005). Census 2001 Quality Report
Ackoff RL (1979). The future of operational research is past. J Opl for England and Wales, http://www.statistics. gov.uk/downloads/
Res Soc 30: 93-104. census2001/census_200 l_quality_report.pdf, accessed 17 April
2009.
Aigner D, Lovell CAK and Schmidt P (1977). Formulation and
estimation of stochastic frontier production function models. Olesen OB (2006). Comparing and combining two approaches for
J Econometrics 6: 21-37. chance constrained DEA. J Prod Anal 26: 103-1 19.

Amado CAF and Dyson RG (2009). Exploring the Use of DEA for Olesen OB and Petersen NC (1995). Chance constrained efficiency
evaluation. Mnet Sci 41: 442-457.
Formative Evaluation in Primary Diabetes Care: An Application
to Compare English Practices. / Opl Res Soc 60: 1469-1482. Sarrico CS and Dyson RG (2000). Using DEA for planning in
Athanassopoulos AD and Shale EA (1997). Assessing the comparative UK universities - An institutional perspective. J Opl Res Soc 51:
789-800.
efficiency of higher education institutions in the UK by means of
data envelopment analysis. Educ Econ 5: 1 17-134. Sarrico CS, Hogan SM, Dyson RG and Athanassopoulos AD (1997).
Banker R, Chames A, Cooper WW and Maindiratta A (1987). A Data envelopment analysis and university selection. J Opl Res Soc
48: 1163-1177.
comparison of data envelopment analysis and translog estimates of
production frontiers using simulated observations from a known Shale EA (1996). Performance measurement of branded public
technology. In: Dogramaci A and Fare R (eds). Applications of houses: An application of data envelopment analysis to the outcome
Modern Production Theory: Efficiency and Productivity, Kluwer of investment decisions. MSc dissertation, University of Warwick.
Academic Publishers: Boston: MA, pp 33-55. Shale E A (2009). The integration of psychometric data into the
Bitner MJ (1990). Evaluating service encounters: the effects of performance assessment of a retail service chain. Working Paper.
physical surroundings and employee responses. J Marketing 54: University of Warwick.
69-82. Simar L and Wilson PW (1998). Sensitivity analysis in efficiency
Blackett PMS (1950). Operational research. Opl Res Quart 1(1): 3-6. scores: How to bootstrap in nonparametric frontier models. Mngt
Sci 44: 49-62.
Charnes A, Cooper WW and Rhodes E (1978). Measuring the
efficiency of decision making units. Eur J Opl Res 2: 429-444. Simar L and Wilson PW (2008). Statistical inference in non-parametric
Charnes A, Cooper WW and Rhodes E (1981). Evaluating program frontier models: Recent development and Perspectives. In: Fried
and managerial efficiency: An application of data envelopment HO, Lovell CAK and Schmidt SS (eds). The measurement of
analysis to program follow through. Mngt Sci 27: 668-697. productive efficiency and productivity growth, Oxford University
Cooper WW (1999). Operational research/management science: Press: New York, p 638.
Where it's been. Where it should be going? The Blackett Memorial Stewart TJ (1996). Relationships between data envelopment analysis
Lecture 18 November 1997. J Opl Res Soc 50: 3-11. and multicriteria decision analysis. J Opl Res Soc 47: 654-665.
Cooper WW, Huang ZM and Li S (1996). Satisficing DEA models Thanassoulis E, Dyson RG and Foster MJ (1987). Relative efficiency
under chance constraints. Ann Opns Res 66: 279-295. assessments using data envelopment analysis: An application to
Cooper WW, Park KS and Yu G (1999). IDEA and AR-IDEA: Models data on rates departments. J Opl Res Soc 38: 397-411.
for dealing with imprecise data in DEA. Mngt Sci 45: 597-607. Zhu J (2003). Imprecise data envelopment analysis (IDEA): A
Desai A, Ratick SJ and Schinnar AP (2005). Data envelopment review and improvement with an application. Eur J Opl Res 144:
513-529.
analysis with stochastic variations in the data. Soc Econ PI Sys 39:
147-164.
Dyson RG and Thanassoulis E (1988). Reducing weight flexibility in Received May 2009;
data envelopment analysis. J Opl Res Soc 39: 563-576. accepted September 2009 after one revision

This content downloaded from 197.255.68.201 on Wed, 06 Jul 2016 17:35:29 UTC
All use subject to http://about.jstor.org/terms

Você também pode gostar