Keywords

1 Introduction

Data envelopment analysis (DEA) is a mathematical programming-based technique to evaluate the relative performance of organisations. While the main applications have been in the evaluation of not-for-profit organisations, the technique can be successfully applied to other situations competing with other techniques as cost benefit analysis and multi criteria decision making as can be seen, for instance, in a recent study about the best choice for traffic planning, namely, the design and location of a highway in Memphis (Bougnol et al. 2005).

DEA is suited for this type of evaluation because it enables results to be compared making allowances for factors (Thanassoulis and Dunstan 1994). DEA makes it possible to identify efficient and inefficient units in a framework where results are considered in their particular context. In addition, DEA also provides information that enables the comparison of each inefficient unit with its “peer group”, that is, a group of efficient units that are identical with the units under analysis. These role-model units can then be studied in order to identify the success factors which other comparable units can attempt to follow. Thanassoulis (1993) argue that DEA is preferable to other methods, such as regression analysis, which also make it possible to contextualise results.

This chapter is structured as follows. The next section describes the development and fields of application of the technique, while Sect. 3.3 introduces the DEA models followed by a numerical and graphical example. Section 3.4 presents the mathematical formulation and the last one ends up with the main conclusions.

2 History and Applications of DEA

DEA is a mathematical programming technique presented in 1978 by Charnes et al. (1978), although its roots may be found as early as 1957 in Farrel’s seminal work (Farrell 1957) or even to Debreu’s, which introduced in the early fifties the “coefficient of resource utilisation” (Debreu 1951). It deserves special attention and also the work of the Dutch Nobel-prized Tjalling Koopmans and his “activity analysis concepts” (Koopmans 1951).

The DEA technique is usually introduced as a non-parametric one, but in fact it rests on the assumption of linearity (Chang and Guh 1991) and for the original constant returns to scale (CRS) models even in the more stringent assumption of proportionality.

Its application has been focused mainly on the efficiency assessment of not-for-profit organisations, since these cannot be evaluated on the basis of traditional economic and financial indicators used for commercial companies.

The first application of DEA was in the agriculture field; as a matter of fact, Farrell applied it to 1950 data of 48 states in the United States of America, considering 4 inputs and 2 outputs. At that time, the DEA term was not yet created, so in fact the first time the term DEA and that technique was applied was in the area of education, specifically in the analysis of Program Follow Through, conducted in the USA, in the late seventies (Rhodes 1978). Since then it has been used to assess efficiency in areas such as health (Wilson et al. 2012), county gaols (Seiford and Zhu 2002), courts (Schneider 2005), universities (Bougnol et al. 2010) and many other not-for-profit sectors. Nowadays DEA can be seen to have spread to other fields such as transit (Chiu et al. 2011), mining, (Chen et al. 2010), air transportation (Pestana e Dieke 2007) and even banking (Emrouznejad and Anouze 2010).

In data envelopment analysis, the organisational units to be assessed should be relatively homogeneous and were originally termed decision-making units (DMUs). As the whole technique is based on comparison of each DMU with all the remaining ones, a considerable large set of units is necessary for the assessment to be meaningful. We will assume that each DMU produces N outputs by means of M inputs.

3 The Meaning of DEA Efficiency

We will introduce some simple examples based on the following data set with 4 variables: 2 inputs – X 1 and X 2 – and 2 outputs – Y 1 and Y 2. Since this data will be used for pedagogical proposes, it is a small data set with just 12 decision-making units (DMU) (Table 3.1).

Table 3.1 Illustrative data set

First, to introduce some basic concepts, we will suppose that only X 1 and Y 1 would be important for our analysis.

In this case, we can graph the 12 observations on a scatter plot, and it would be obvious that the most efficient one will be the DMU 3 since a straight line originating at the (0;0) point towards DMU 3 has the higher slope than any of the remaining (Fig. 3.1).

Fig. 3.1
figure 1

Example with 12 hypothetical farms consuming X 1 and producing Y 1

The straight line originating at the (0;0) point towards DMU 3 is called the efficiency frontier and together with the X axis it defines a cone with its vertex at the origin. This cone is called the production possibility set, since it contains all real data, and according to DEA axioms, only points inside this cone correspond to possible working conditions based on best-achieved performance.

We will analyse in greater detail unit 1. There are two ways for DMU 1 to reach efficiency: increasing output, till it reaches M.O. (maximisation of output) or reducing the input till m.i. (minimisation of input).

The actual value for efficiency is defined as the ratio between the distances: \( {{{d\left( {\mathrm{m}.\mathrm{i}. - \mathrm{z}.\mathrm{i}.} \right)}} \left/ {{d\left( {1 - \mathrm{z}.\mathrm{i}.} \right)}} \right.} {=} {(2 - 0) \left/ {(4 - 0) } \right.} {=} 50\% \). On a similar way, the actual value for inefficiency is defined as the ratio between the distances: \( {{{d\left( {\mathrm{M}.\mathrm{O}. - \mathrm{Z}.\mathrm{O}.} \right)}} \left/ {{d\left( {1 - \mathrm{Z}.\mathrm{O}.} \right)}} \right.} {=} {(4 - 0) \left/ {(2 - 0) } \right.} {=} 100\% \) . In these calculations, z.i. means the zero input point and Z.O. the zero output point for the DMU 1.

We conclude that in DEA, there are two different means of optimization related with radial measures of score: input minimisation and output maximisation. The artificial points M.O. and m.i. are termed targets or composite points for DMU 1. Point 3 is the only efficient one, and it is the peer for all remaining points.

Under constant returns to scale, efficiency is the reciprocal of inefficiency; the peer set is also the same, regardless of the orientation, although their targets are different as we conclude from the exposition above.

Typical DEA software like efficiency measurement system (EMS), developed by Holger Scheel (Scheel 2000) at Dortmunt University (using Csaba Mészáros’ BPMPD interior-point solver) would give us results as depicted in Fig. 3.2.

Fig. 3.2
figure 2

Example of EMS results under constant returns to scale minimisation of inputs

EMS highlights the efficient units and provides us all the necessary values:

  • Score means efficiency.

  • X 1{I}{V} presents the virtual input of variable X 1, in a similar way Y 1{O}{V} presents the virtual output of variable Y 1.

  • Benchmarks means that all farms, except for the third, are benchmarked against farm 3 and so is used 11 times for comparison. The coefficients in this column mean that if we multiply farm 3 by those values, it will result in the composite-projected target point.

  • {S}X 1{I} presents the slacks for input X 1, in a similar way {S}Y 1{O} presents the slacks for output Y 1.

Thinking again around unit 1, it means that its target under input minimisation is 0,33 × (6;6) = (2;2). These are exactly the coordinates of point m.i. in Fig. 3.1. The concepts of virtual multipliers and slacks will be clarified after the presentation of the mathematical formulation.

So far it is very important to emphasise that there are two different radial orientations: input minimisation and output maximisation. There are target points and a set of efficient units that benchmark the non-efficient ones.

The results for output maximisation are depicted on Fig. 3.3.

Fig. 3.3
figure 3

Example of EMS results under constant returns to scale maximisation of outputs

We can confirm that the conclusions are very similar to those from Fig. 3.2; more specifically the score now is the inefficiency, the reciprocal of the scores under input minimisation. Slacks are identically null as in the minimisation of inputs.

We can proceed with our analysis including output Y 2 in our analysis. In this case we will have 3 variables, and we cannot represent them on the XY plane. Anyway, since we assume that constant returns to scale prevail, we can normalise the outputs by the single input X 1. That way, we get the Fig. 3.4, where we can easily spot three efficient farms: Farm 3 remains efficient (whenever we add new variables, efficiencies never decrease), and farms 4 and 10 turn out efficient, joining the new efficiency frontier.

Fig. 3.4
figure 4

Cut of the PPS at X 1 = 1, under CRS, two outputs: Y 1 and Y 2

We point out in Fig. 3.4 the zero output point (Z.O.) at the origin and the zero input area that lies outside the production possibility space (PPS). The PPS is now a polyhedral cone (an inverted pyramid) with 3 specific edges: the rays that depart from the origin to each of the 3 efficient units (3;4;10). There are other 3 edges that are not specific, since they exist always regardless of the specific data set; one belongs to the plane defined by X = 0, the other belongs to the plane with Y = 0 and finally the third is the Z axis with both X = 0 and Y = 0 (Fig. 3.4).

The graph in Fig. 3.4 is a cut of the PPS at X 1 = 1, since we normalised the outputs. The dashed line is the intersection of the horizontal plane X 1 = 1 with the efficiency frontier (the polyhedral cone). The dotted line is an isoefficiency contour.

It is clear that DEA clusters DMU according to their specificities; farms 1 and 8 are related to farms 4 and 10, their peer group. The same way farm 6 will be similar to farm 3, and so on. A different situation arises with farm 2, which cannot be expressed as a finite linear positive combination of the efficient ones; in that case, it is necessary to have a slack in Y 1. The same situation happens with farms 7 and 9, they “need” slacks in Y 2, and this means that after their projection on the vertical part of the dashed line, they still have to increase its production of Y 2. This can be confirmed by the EMS results for this case. Typically this kind of figure is associated with input minimisation, since we have only one input to reduce. It can be associated also to output maximisation, where an equiproportional maximisation of both outputs (keeping the output mix constant) is required.

Finally, we examine the case for two inputs and a single output Y 2. This time we cannot anticipate the results, since we did not study efficiencies for pairs of these three variables. Again, since we assume that constant returns to scale prevail, we can normalise the inputs by the single output Y 2. That way we get Fig. 3.5 where we plotted X 2/Y 2 versus X 1/Y 2 where we can easily spot three efficient farms:

Fig. 3.5
figure 5

Cut of the PPS at Y 2 = 1, under CRS, two inputs: X 1 and X 2

The PPS is again a polyhedral cone (an inverted unbounded pyramid) with 3 specific edges: the rays that depart from the origin to each of the 3 efficient units (3;7;10). There are no other edges since the inefficient units are located in an unbounded region that tends to infinity as outputs approach zero.

Farm 12 is not strongly efficient, since farm 7 has the same ratio on X 2/Y 2, but a smaller one on X 1/Y 2. Indeed there is a slack on X 1. We can see in the graph that the two points differ 0.5 on X 1/Y 2. This happens at the cutting plane Y 2 = 1, as the true scale of operation of farm 12 is Y 2. The slack is 0.5 × 4 = 2 as can be seen in Fig. 3.6 where we present the results for this case. We can notice that the only farms that EMS classifies as efficient are farms 3, 7 and 10, those that are on shadowed lines of the table. Farm 12 presents a score of 100%, although it is not efficient since it has 2-unit slack in X 1; its efficiency according to the Charnes Cooper and Rhodes model (CCR model) is 100% − 2ε where ε is a small non-Archimedean entity. These units that are on the “horizontal or vertical” parts of the efficiency frontier are termed “weakly efficient DMUs”; we will come back to this matter when we present the models in the remaining of this section.

Fig. 3.6
figure 6

EMS results for the two inputs, X 1 and X 2, and a single output, Y 2

So far we have assumed we are operating in a constant returns to scale situation. We will now present only one graphical example for the variable returns to scale (VRS) model, introduced by Banker, Charnes and Cooper in 1984 (Banker et al. 1984).

We will study again the one input (X 1) producing one single output (Y 1) case, but according to Banker, Charnes and Cooper the efficiency frontier is not anymore a polyhedral cone, but instead the intersection of all the polyhedral cones, one for each DMU; this result is based on the convex hull of the all DMUs. In the CCR model, we had flexibility on the choice of the weights; now in the BCC model, we are free to choose also the origin of our data (the vertex of the polyhedral cone).

In Fig. 3.7, we have a vertical facet that ends at DMU 1; this facet means that we assume a fixed amount of input (X 1 = 4) necessary just to start the business. DMU 2 under VRS will have a much higher efficiency than under CRS, especially in the minimisation of inputs case. DMU 2 is spending 5 units of input, and it was expected to spend just one in the CRS case; this leads to a CRS efficiency of 1/5 = 20%; in the more benevolent BCC model, it was expected to spend 4 units, so its efficiency rises to 4/5 = 80%. The ratio between CRS efficiency and VRS efficiency is called scale efficiency and increases from the origin (X 0; Y 0) till DMU 3 and then starts decreasing. The first region is the increasing returns to scale region, and the second is the decreasing returns to scale region. In higher dimensions it seems more difficult, but it is much easier than it seems at first look. The most productive scale size (MPSS) is simply the intersection between the VRS and the CRS efficiency frontiers. From the origin till the MPSS, we are in the increasing returns to scale region; from the MPSS till infinity, we are on the decreasing returns to scale region.

Fig. 3.7
figure 7

Comparison of efficiency frontiers of the CCR and BCC models

In CRS, we had a simple relation between the scores of input minimisation and output maximisation (they were reciprocals: efficiency 1/inefficiency). Under VRS it does not happen anymore. In fact there is no clear relation between those two scores since the peer set most of the times differs from one orientation to the other. For instance, DMU 10 under input minimisation has DMU 1 and 3 as peers, but if we intend to maximise its output, the reference set is 3 and 4.

In Fig. 3.8, we present the results for the one input, X 1, and a single output, Y 1, under VRS and minimisation of inputs. We cannot represent graphically any 3 variables case under variable returns to scale, so now we have to introduce the mathematical formulation that relies on linear programming and duality.

Fig. 3.8
figure 8

EMS results for the one input, X 1, and a single output, Y 1, under VRS

4 Mathematical Formulation for DEA

In DEA, efficiency (Ef a ) of a specific decision-making unit (DMU a ) under analysis is defined as the ratio between a weighted sum of its s outputs Y ra and a weighted sum of its m inputs X ia , a natural extension of the concept of efficiency used in the fields of physics and engineering (Charnes et al. 1978).

$$ \mathrm{E}{{\mathrm{f}}_a} = \frac{{\sum\nolimits_{r = 1}^s {{\mu_{ra }}{y_{ra }}} }}{{\sum\nolimits_{i = 1}^m {{v_{ia }}{x_{ia }}} }} $$
(3.1)

When assessing a set of J organisations, where X ik stands for the ith input of the kth DMU, with a similar meaning for Y rk , the weights μ rk and v ik , in Eq. (3.1), are chosen for each DMU j , under evaluation as those that maximise its efficiency as defined by Ef a . Several constraints have to be added to the maximisation problem:

  • The strict positivity (Charnes et al. 1978) of the weights μ rk and v ik , (also known as virtual multipliers).

  • For scaling purposes, all n DMUs under analysis must have efficiencies not exceeding an agreed value, typically one or 100%, as is usual in engineering definitions of efficiency.

  • A third kind of restriction has to be included since otherwise this linear fractional program would yield an infinite number of solutions. In fact, if a set of weights μ rk and v ik returns the optimal solution, so would rk and kv ik . Making the denominator in Eq. (3.1) equal to one or 100% circumvents this situation.

So, we have to solve the following linear programmingmaximisation problem for each one of the J DMUs under analysis:

$$ \max \;\mathrm{E}{{\mathrm{f}}_a} = \frac{{\sum\nolimits_{r = 1}^s {{\mu_{ra }}{y_{ra }}} }}{{\sum\nolimits_{i = 1}^m {{v_{ia }}{x_{ia }}} }} $$
(3.2)
$$ \mathrm{s}.\mathrm{t}.\;\;{\mu_{ra }} \geqslant \varepsilon > 0\quad r = 1\ldots s $$
(3.3)
$$ {v_{ia }} \geqslant \varepsilon > 0\quad i = 1\ldots m $$
(3.4)
$$ \mathrm{E}{{\mathrm{f}}_k} = \frac{{\sum\nolimits_{r = 1}^s {{\mu_{rk }}{y_{rk }}} }}{{\sum\nolimits_{i = 1}^m {{v_{ik }}{x_{ik }}} }} \leqslant 1\quad k = 1\ldots n $$
(3.5)
$$ \sum\limits_{i = 1}^m {{v_{ia }}{x_{ia }} = 1} $$
(3.6)

This fractional linear program can be solved by the Charnes and Cooper transformation (Charnes and Cooper 1962) which yield the following linear program:

$$ \mathrm{Max}\;\mathrm{E}{{\mathrm{f}}_a} = \sum\limits_{r = 1}^s {{\mu_{ra }}{y_{ra }}} $$
(3.7)

s.t.

$$ \sum\limits_{i = 1}^m {{v_{ia }}{x_{ia }} = 1} $$
(3.8)
$$ \sum\limits_{i = 1}^m {{v_{ia }}{x_{ik }} - \sum\limits_{r = 1}^s {{\mu_{ra }}{y_{rk }} \leqslant 0\quad j = 1{\ldots} J} } $$
(3.9)
$$ {\mu_{ra }} \geqslant \varepsilon > 0\quad r = 1{\ldots} s $$
(3.10)
$$ {v_{ia }} \geqslant \varepsilon > 0\quad i = 1{\ldots} m $$
(3.11)

The problem above is known as the multiplier problem, since its unknowns are the weights, which are usually lower bounded by a small quantity non-Archimedean entity, ε (Eqs. 3.10 and 3.11), so that all inputs and outputs are considered in the evaluation (Charnes et al. 1979) even if with a minor weight ε, set typically equal to 10−6.

The dual of this problem, which we shall call the envelopment problem, provides important information about economies that could be achieved in all the inputs; it also indicates which efficient units the inefficient unit being assessed should emulate. Those efficient units are usually referred to as the reference set or peer group of the unit under evaluation.

Any linear problem has a dual, which we will call the envelopment problem that can be written as follows:

$$ \min :{Z_{\mathrm{a}}} - \varepsilon \ \left\{ {\sum\limits_{r = 1}^s {S_{ra}^{+ } + \sum\limits_{i = 1}^m {S_{ia}^{- }} } } \right\} $$
(3.12)
$$ \mathrm{s}.\mathrm{t}.\sum\limits_{k = 1}^n {\lambda_k {x_{ik }} + S_{ia}^{- } = {Z_a}{x_{ia }}} \qquad i = 1\ldots m $$
(3.13)
$$ \sum\limits_{k = 1}^n {\lambda_k {y_{rk }} - S_{ra}^{+ } = {y_{ra }}\qquad r = 1\ldots s} $$
(3.14)
$$ {\lambda_k} \geqslant 0\qquad k = 1\ldots n $$
(3.15)
$$ S_{ra}^{+ } \geqslant 0\qquad r = 1\ldots s $$
(3.16)
$$ S_{ia}^{- } \geqslant 0\qquad i = 1\ldots m $$
(3.17)

This formulation can be interpreted as follows: Given DMU a, find the composite unit which has no smaller outputs than this one and whose inputs are smaller than those of DMU a scaled down by a factor \( {Z_a} \) as small as possible. This is why this formulation is known as input minimisation; since we are minimising \( {Z_a} \), we are seeking the minimal inputs that, based on best-achieved performance, could still produce the same amount of outputs as DMU a is currently producing. These composite units are finite linear combinations of efficient units more specifically finite positive combinations which lie on the efficiency frontier. These efficient DMUs are known as the peer group, the role-model units that inefficient DMU a should try to emulate.

In the envelopment problem, it is easy to understand the role of the small non- Archimedean entities ε; they simply multiply by the sum of slacks so that slacks are not ignored in the overall score.

As a whole, the interpretation of the DEA technique is straightforward and can be put in the following terms:

4.1 Multiplier Problem

Evaluate each DMU with the set of weights which maximises its efficiency, provided that all other DMUs, rated with that set of weights, have efficiency not greater than unity.

4.2 Envelopment Problem

Find the smallest proportion \( {Z_a} \) of inputs that would bring the current DMU a to the enveloping surface of all DMUs.

The model presented above, named CCR after the acronym of the authors Charnes et al. (1978), assumes constant returns to scale. This means that when the input of an efficient unit is multiplied by a given factor, its output level is also multiplied by the same factor. In this case, the production possibilities set will be a closed convex polyhedral cone in \( {{\mathcal{R}}^{n + m }} \) in the positive orthant.

In other situations this is not the case. The scale of operations may have an impact on the outputs, creating “economies” or “diseconomies” of scale. The BCC model, developed by Banker et al. (1984), can deal with variable returns to scale.

The BCC envelopment model can be obtained from the CCR envelopment model by adding the convexity constraint to the envelopment problem:

$$ \sum\limits_{k = 1}^n {\lambda_k } = 1 $$
(3.18)

The multiplier problem is the dual of the envelopment formulation, and the extra restriction originates an extra free variable. It is important to note that if we relax the convexity constraint to \( \sum\nolimits_{k = 1}^n {\lambda_k \geqslant 1} \) than its dual, variable will become non-negative and it becomes clear that this model (the non-decreasing returns to scale model) is nothing else but a CCR model with an extra output identically unitary.

The same analysis could be conducted based on the maximisation of outputs, leading to similar formulations.

For instance, DMU 10 will be rated under VRS as 5/8 efficient and 4/8 = 50% under CRS. It is straightforward to conclude that VRS efficiency is never smaller than the CRS efficiency. This became clear if we note that the envelopment problem has an extra constraint, and it is a minimising program. The ratio between those two efficiencies is called scale efficiency, and it is important in the efficiency analysis. All those definitions can be made also for the output maximisation orientation.

In VRS we can define several regions, the increasing returns to scale and the decreasing returns to scale that expands until the infinity. Those regions are separated by the most productive scale size, where the frontier of the pointed cone intersects the convex hull of the DMUs. In Fig. 3.7 it corresponds to point 3, the intersection between the dotted line (a linear hyperplane of dimension 1) and the dashed polygon.

5 Concluding Remarks

Using all the typical graphical representations of data envelopment analysis, the authors introduced based on a visual and spatial approach the basic ideas and principles of DEA. The two models for CRS and VRS, known as the CCR and BCC models, respectively, were presented, as well as both input minimisation and output maximisation orientations; this makes a total of 4 different models, and for each of those 4, there is a dual (weights = multipliers versus envelopment problems).

Many new developments of DEA could be presented here, most specifically weights restrictions, non-controllable inputs and outputs and what may be the most important one, the superefficiency model, that was introduced by Andersen and Petersen from Odense University by 1993 (Andersen and Petersen 1993). This technique is rather appealing because they allow efficiency to be greater than one, discriminating between efficient DMUs that otherwise are all ranked equal; it avoids also ambiguity on weights of those units (multiple solutions on the weights = multipliers problem, degeneracy in its dual).

The equations presented allow the implementation of DEA software in many programs as Excel, SAS, Mosel, Stata, OpenOffice and on any other program with linear programming solving capabilities.