Introduction

In October 2006, the European Commission published the Action Plan on Energy Efficiency. The plan aims to realize a 20% energy efficiency improvement by 2020 (EC 2006a). This objective was reinforced in January 2007 when the European Commission presented her energy package for the twenty-first century. The package included a 20% emissions reduction target for greenhouse gases by 2020, among others, to be achieved by improving energy efficiency by 20% (EC 2007). Realizing this potential requires the introduction of good new energy efficiency policies as well as strengthening and enforcing existing policies on the European and the national level. This raises the question: what characterizes good and effective energy efficiency policies and their implementation? Systematic ex post evaluation of energy efficiency policies can reveal factors that explain success and failure related to policy design as well as the implementation process. However, while an increasing number of energy efficiency policy instruments are being introduced in Europe and elsewhere, only few of the instruments are evaluated systematically. Evaluations are often hard to compare due to the diversity in methods and indicators used. Furthermore, methods currently applied in Europe in ex post policy evaluation are mostly focused on ‘final effects,’ i.e., energy savings and cost effectiveness. Limited research is focused on bringing policy evaluation methods on an equal footing and on systematically assessing successes and failures in the implementation of energy efficiency policies. Efforts so far include:

  • The SAVE project entitled “A European Ex Post Evaluation Guidebook for Demand Side Management (DSM) and Energy Efficiency (EE) Service Programmes” (SRC 2001), which provides general guidelines for ex post evaluation of DSM and EE services. These guidelines were tested for a number of DSM and EE service programs in the European Union.

  • The evaluation guidebook “Evaluating Energy Efficiency Policy Measures and DSM Programmes” (IEA 2005) published by IEA DSM IA (task 1). They provide guidance for systematically evaluating the implementation process of energy efficiency policy instruments. The developed method was applied for various types of instruments implemented in IEA countries.

Broader experience in systematic ex post evaluations that are not just focusing on “final effects” can be found in the USA. For example:

  • “California Energy Efficiency Evaluation Protocols: Technical, Methodological, and Reporting Requirements for Evaluation Professionals” (Hall et al. 2006). This guidebook is to be used in evaluating California’s energy efficiency programs. The guidebook states that an evaluation plan shall at least include a brief description of the program(s) being evaluated including a high-level presentation of the program theory. The Protocols are the primary guidance tools for policy makers and evaluation contractors and has been applied various times to assess the impact, and success and failure of policies.

  • Evaluations performed within the framework of the New York Energy $mart Programme (NYSERDA 2008). Examination of causality is an important component of the New York Energy $mart evaluation efforts. It is aimed at helping to validate program interventions, justify the expenditure of public benefits funds, and assist policy makers in decisions regarding future funding of public benefit programs.

  • Handbook prepared for the US Department of Energy “Impact Evaluation Framework for Technology Deployment Programs” (Reed et al. 2007). The handbook provides an approach for quantifying retrospective energy savings, clean energy advances, and market effects of technology deployment programs, which includes preparing a program theory.

Within the EU-funded project “Active Implementation of the European Directive on Energy Efficiency” (AID-EEFootnote 1), we tried to contribute to the further development of a generic framework for ex post evaluations within Europe (Joosen and Harmelink 2006). The method is based on the theory-based policy evaluation as described by Blumstein et al. (2000). The method is designed to systematically assess all steps of the policy implementation process with the aim to determine success and failure factors but also final effects such as target achievement, energy savings impact and cost effectiveness, as well as success and fail factors. By developing this method, we aim to contribute to the further development and harmonization of ex post policy evaluation methods and to creating comparable evaluation outcomes. Within the project, the method was applied to evaluate 20 energy efficiency policy instruments applied across Europe, Japan, and the USA (Ecofys et al. 2007).

This paper starts with an overview of methods applied in ex post policy evaluation, followed by a description of the theory-based policy evaluation method and the practical approach developed within the AID-EE Project. Next, overall findings are presented of the 20 case studies with respect to target achievement, energy savings impact, cost effectiveness, and typical success factors per instrument. This paper does not provide the detailed results from the various case studies. These can be found on the project website http://www.aid-ee.org and are also presented in (Harmsen et al. 2007; Bongardt and Kebeck 2007; Khan and Nordqvist 2007; Nordqvist 2007). Finally, conclusions and a discussion are presented on the practical application of the method for policy makers.

Methods applied in ex post policy evaluation

A variety of methods are applied in ex post evaluation of policy instruments. One is the assessment of aggregate ‘top-down’ indicators on energy consumption per sector or end user. Based on statistics, a hypothetical baseline is constructed assuming that energy efficiency stays unchanged from the base year (frozen energy efficiency) or is adjusted for autonomous efficiency improvements. The actual energy use is subtracted from this amount, and the difference is defined as the amount of energy saved. This method generally does not provide much insight in the impact of individual policies due to the aggregated level of analysis.

Another approach is econometric modeling of the impact of policy instruments. In econometric modeling, a list of factors (one of which is the analyzed policy instrument) is drawn up that potentially could affect (specific) energy use of a sector. Through statistical methods, the impact of the analyzed policy instruments can be estimated. The approach is suitable for evaluating general policy instruments, such as taxes. These methods, however, do not provide insight in ‘why’ an instrument performed or did not perform as expected and what could be done to improve it.

A third approach is detailed bottom-up policy evaluation that focuses on determining the ‘final effects’ of a single policy instrument or a package of instruments. A bottom-up calculation method means that energy savings obtained through the implementation of a specific type of energy efficiency improvement measure (e.g., a CFL) are determined in GJ or kWh that can be attributed to specific energy efficiency policy instruments. However, it must be noted that it is often difficult to determine the effect of single-policy instruments because of the fact that energy efficiency instruments often come in a package, that autonomous efficiency improvement cannot be clearly separated from policy-induced savings, and that the impact of rebound effect, free riders, etc. are insufficiently known. A combination of top-down and bottom-up evaluation methods will be the officially applied methods to evaluate the EU Energy End-use Efficiency and Energy Services Directive (ES Directive; EC 2006b).

A fourth approach complementary to all three but particularly to bottom-up impact evaluation is using policy theory. The general principle is that a likely theory is drawn up on how the policy instrument should achieve its targeted impact in terms of energy efficiency improvement. Several terms are used for this kind of approach, including logic model analysis (Megdal 2005), realistic evaluation theory (SRC 2001), program theory (IEA 2005), theory-based approach (Rufo et al. 1999; Goldstone et al. 2000), and intervention theory (Dunn 2003). The advantage of this approach is that insight can be gained on the full implementation process, including explanatory factors behind the impact.

Applied methodology: theory-based policy evaluation

Theory-based policy evaluation

The theory-based approach is not new and has been used numerous times to evaluate policies. The method of “theory-based” policy evaluation is extensively described and illustrated in Rossi et al. (2004). The application of the method for evaluating energy efficiency policies in Europe has been limited. The previous sections showed that there is a longer tradition with these types of evaluations in evaluating DSM programs in the USA. In Europe, the method was, e.g., applied to evaluate energy efficiency policies applied in the built environment (Joosen et al. 2004) and to perform a mid-term evaluation of the Reduction Plan on Non-CO2 Greenhouse Gases (Harmelink et al. 2006). However, this did not result in a tradition to systematically set up guidance for the future evaluations of energy and climate change policies.

The theory-based approach requires an iterative process of program design, evaluation, and redesign based on the lessons learned from the initial program (see Fig. 1). In practice, this means that theory-based policy evaluation establishes a plausible theory on how a policy instrument (or a package of instruments) is expected to result in energy efficiency improvements, and who is expected to take action at which point in time. The basic idea is to unravel the whole policy implementation process. Through this unraveling, insight is gained on ‘where something went wrong in the process of policy design and implementation’ and ‘where the keys are for improving the impact and cost effectiveness.’

Fig. 1
figure 1

Outline of the policy cycle and the role of the program theory in the policy cycle

Practical framework to evaluate policy instruments

Within the framework of the AID-EE Project, the theory-based policy evaluation method was translated into a practical six-step approach that was applied to evaluate the 20 case studies. An earlier version is described in Harmelink et al. (2005). The approach applies the following steps to evaluate individual instruments or packages of instruments:

  1. 1.

    In the first step, the policy instrument or policy package is characterized. This includes, among others, a description of targets, the period the policy instrument was active, the target groups, the policy-implementing agents, the available budget, the available information on the initially expected energy savings impact, and the cost effectiveness of the instrument.

  2. 2.

    In the second step, a policy theory is drawn up. A policy theory includes all the assumptions on the way the policy instruments should reach its targeted impact. A policy theory can be either explicit or implicit. In the ideal case, an explicit theory is available. This means that the policy makers have clearly described how they think the policy instrument is going to work before its implementation. That is, that they have clearly stated which actor needs to take action and that they have stated the expected outcome of each action. Often, the theory is largely implicit, and such a description is lacking. In this case, the evaluator has to try to reconstruct the policy theory. Drawing up a policy theory in practice includes documenting all implicit and explicit assumptions in the policy implementation process and mapping the cause–impact relationship, including the relationship with other policy instruments.

  3. 3.

    In the third step, the policy theory is translated to concrete and preferably quantitative indicators. This means that, for each assumed cause–impact relation, an indicator is drawn up to “measure” whether the cause–impact relation actually took place and to “measure” whether the change (or part of the change) that took place is due to the implementation of the policy instrument (i.e., the policy instrument was the causal force). This step also includes the development of the necessary formulas to calculate the impact and cost effectiveness.

  4. 4.

    In the fourth step, the cause–impact relations and the indicators are visually reflected in a flowchart. An example of such a flowchart is provided in Fig. 2.

  5. 5.

    In the fifth step, the policy theory is verified and, if necessary, adjusted. In step 2, the policy theory was drawn up with the help of available (official) documents or experiences with similar instruments. In the fifth step, the policy theory is verified through interviews with policy makers, implementing agents, and other actors involved in the implementation and monitoring of the policy instrument.

  6. 6.

    In the sixth and final step, (a) available information is gathered and analyzed to draw up the indicators, (b) conclusions are drawn on the energy savings impact and cost effectiveness of the policy instrument using the formulas and indicators, (c) analyses are made on the success and failure factors attributed to the analyzed instruments, and (d) recommendations are formulated to improve the energy savings impact and cost effectiveness.

Fig. 2
figure 2

Example of a flow chart for the Energy Investment Deduction Scheme (EIA) in The Netherlands showing cause–impact relations, indicators, success and failure factors, and interactions with other instruments

Characteristics of the selected instruments

A great variety of policy instruments are in place in EU member states, on the EU level, and in countries outside the EU to stimulate energy efficiency improvement in different sectors. Instruments range from direct regulations, financial incentives, and voluntary agreements. Within the AID-EE Project, we aimed for a representative selection of instruments to evaluate; i.e., the selected instruments should be a good representation of the great variety of different types of policy instruments applied in the different sectors to improve energy efficiency. The following criteria were applied for selecting the 20 case studies:

  1. 1.

    Selected instruments should be aimed at achieving substantial energy savings and/or be aimed at market implementation of energy-efficient technologies at the national level.

  2. 2.

    Selected instruments should be aimed at the implementation of energy end-use efficiency improvement measures (i.e., we excluded, e.g., combined heat and power policies).

  3. 3.

    Selected instruments should be applied in sectors that are covered under the ES Directive.

  4. 4.

    Balanced breakdown of selected instruments among different sectors.

  5. 5.

    Balanced breakdown of selected instruments among different types of instruments; i.e., the total package of selected instruments should be a good representation of the existing variety of implemented instruments.

  6. 6.

    Some monitoring data should be available.

Table 1 presents the instruments that were selected as case studies in the AID-EE Project. The instruments are grouped by type of instrument: regulation, financial, informative, voluntary agreements, and procurement. It must, however, be noted that most instruments come in a package (e.g., regulation that is linked to information campaigns and financial incentives) so that it is not always easy to clearly put the instruments in a specific category. Hence, the categorization is somewhat arbitrary. This also implies that we did not only assess individual instrument but in most cases packages of instruments.

Table 1 Overview of the analyzed instruments

Results

This section presents the overall findings of the 20 case studies with respect to data availability, targets and target achievement, impact on energy savings, cost effectiveness, and typical success factors per instrument. Individual case studies are available at http://www.aid-ee.org.

Data availability

Figure 3 shows the extent to which it was possible to assess the energy savings impact, the target achievement, critical indicators determined in the policy theories developed for the instruments, the side effects, and the costs. The main observation from the case studies is that most instruments lack a comprehensive monitoring system. The availability and quality of monitoring data turned out to be much lower than expected at the start of the project. Data to assess target achievement and energy savings impact were available for roughly half of the evaluated instruments. Quantitative information on indicators that could explain success or failure is hardly available. For 17 of the evaluated instruments, figures on government costs were available. Information on costs for end users and society were only available for a small number of the evaluated instruments.

Fig. 3
figure 3

Availability of monitoring information for the 20 different case studies

Targets and target achievement

A target was defined as a specific, either quantitative or qualitative objective that has been set at the time the policy instrument was introduced. Target achievement was defined as the extent to which a policy instrument achieved its stated target(s). Figure 4 shows that, from the selected case studies, only five instruments focus on energy savings only and do not explicitly seem to have other objectives. Energy savings policies and CO2 emission reduction policies can often not be separated. In seven case studies, CO2 emission reduction is the primary objective, and it is mentioned as an additional objective in three case studies. Market transformation is an important additional objective in seven case studies.

Fig. 4
figure 4

Type of target set for the 20 different case studies

Table 2 presents an overview of some of the characteristics of the evaluated policy instruments with respect to the targets set (quantitative or qualitative), the evaluated period, and the target achievement. In general, for regulatory instruments and voluntary agreements, quantified targets are set, whereas this quantification lacks for most of the informative instruments. Six of the evaluated instruments reached their target. We also observed that, for most instruments, quantitative targets are set for the targeted end-year and generally not for intermediate years.

Table 2 Evaluated period, target, and target achievement for the 20 case studies

Impact on energy savings

The impact (in the literature also referred to as effect) of a policy instrument was defined as the extent to which a policy instrument made a difference compared to the situation without a policy instrument (business as usual). Business as usual is defined as the development of energy consumption/demand in the absence of the evaluated (package of) policy instrument(s). In Fig. 5, the energy-saving impact is presented as the annual energy efficiency improvements in the evaluated period. The impact was corrected for free-rider effects. Other gross-to-net correction factors due to rebound and spill-over effects were not taken into account due to lack of data. Total energy use of the targeted sector was taken from the European Energy and Transport Trends 2030 (EC 2003).

$$\operatorname{Ir} = \frac{{\operatorname{Es} }}{{t \times \operatorname{Eu} }}$$
Ir:

impact of energy savings in percent per year

Es:

total cumulative energy savings in the end-year of evaluations corrected for free-rider effects in PJ

t :

evaluated period in years

Eu:

total energy use of the targeted sector end in end-year of evaluation in PJ

Fig. 5
figure 5

Annual percentage energy savings improvements per instrument in the evaluated period

Table 2 shows the evaluated period, target, and target achievement for the 20 case studies.

Figure 5 shows the calculated annual energy efficiency improvement with a breakdown by different instruments. The transport sector shows the lowest improvement rates. It can also be observed that there are almost no instruments in place that exclusively focus on the service sector. All instruments that target the service sector also address either industry or households. Instruments that exclusively focus on the service sector are rare. As the Flemish and many US commercial sector programs show, it is not more difficult to achieve savings in the service sector than elsewhere. Figure 5 does not show clear differences in savings for the different types of instrument. One should note that, for some instruments, such as the Danish VA scheme, the Dutch labeling, and the Finnish audit program, the results are given for the policy package and not for individual instruments. This might explain why the energy audit and the labeling scheme (both including financial incentives) have a higher improvement rate than the other mainly informative instruments. Furthermore, in some sectors and countries, there was relatively much “low-hanging fruit,” i.e., highly cost-effective measures that are not implemented due to various barriers. In such cases, high-energy efficiency improvement rates may be possible. Some instruments focus on just a small part of the sector, which makes the energy efficiency improvement rate at sector level rather small. An example is an energy performance standard for new buildings that only slowly penetrates because the demolition rate in most countries is rather low.

It must be noted that data on the energy-saving impact of instruments are surrounded by relatively high uncertainties because of limited availability of data. This is reflected with the error bars in Fig. 5. Due to lack of monitoring data, several assumptions had to be made on “real-life” performance of energy-saving measures (actual energy efficiency performance, operational hours, use of buildings and appliances, etc.).

Figure 5 shows the annual percentage energy savings per instrument and sector in the evaluated period (see Table 2 for the evaluated period). Error bars reflect uncertainty in the calculated impact.

Most evaluated instruments are part of a policy package (see Table 1). Often links between instruments are intended and meant to increase the impact of the whole package with respect to energy savings. In general, it is difficult to determine the isolated impact of a single instrument in a policy package. Informative instruments, which are generally implemented to support other instruments, constitute a good example. Their isolated impact is generally small or even zero. However, our analysis shows that both regulatory and financial instruments, as well as voluntary agreements, would not be so effective without informing target groups on their obligations, financial benefits, etc. Literature on the reinforcing or mitigating effect of policy instruments is scarce, and quantitative results are mostly lacking. Boonekamp (2005) developed a method to analyze the interaction between two instruments. This method was further elaborated by Michelsen (2005) for policy instruments providing incentives for the procurement of energy-efficient cold appliances and complemented with empirical evidence by interviewing experts.

Cost effectiveness

Cost effectiveness refers to the ratio between the energy-saving impact and the amount of money needed to achieve this impact and can, e.g., be expressed in euros per GJ or kWh saved. For a complete assessment of the cost effectiveness of a policy instrument, costs and benefits should be assessed from different perspectives:

  • End user. The costs for the end user provide an indication of the costs as experienced by the end user responsible for the implementation of the energy efficiency measure. These costs are defined as all additional costs that have to be made by the end user compared to the reference situation in case the evaluated energy efficiency policy instrument would not have been in place. The benefits include all savings on the energy bill.

  • Society. The costs for society provide an indication of the costs as experienced by the society as a whole for the implementation of the energy efficiency measure. The main difference between cost–benefit analysis from an end-user perspective and from a social perspective is the time preference. The social perspective is translated into a discount rate that is generally much lower than cutoff discount rates used by end users. In the case of cost–benefit analysis from the social perspective, the discount rate is called the social discount rate.

  • Governments. Costs for the government are defined as all expenditures that have been made by the government, which can be related to the implementation of the evaluated energy efficiency policies. Government expenditure includes budgets for subsidies, grants for research and development, costs for monitoring, and the administrative machinery. However, this also includes reduced government income due to fiscal measures and lowered energy tax income.

In the case studies, we have tried to determine the cost effectiveness for the three perspectives. However, due to the limited availability of data, it was impossible to make a useful comparison of the cost effectiveness from the perspective of the end user and the society. Government expenditures were typically between 0 and 5 EUR/GJ saved in the cases where this number could be determined. Although cost efficiency is an important evaluation criterion, it was beyond the means of the case studies to collect primary data for assessing this. In some cases, evaluations that included cost-efficiency assessments existed, but due to lack of transparency and documentation, and apparent differences between approaches, a comparison between estimates is not meaningful.

Success and fail factors

Using a policy theory approach facilitates a systematic and structured assessment of the cause–impact relationships in the implementation of policy instruments and, hence, the identification of success and fail factors. The 20 policy instruments evaluated in the AID-EE Project are of different types; they target different sectors, and they are implemented in different organizational and institutional contexts. Nevertheless, it is possible to make some general observations based on the recurrence of certain observations made independently in the different case studies. Table 3 in Appendix indicates in what circumstances that different types of instruments are suitable and, derived from the case studies, lists a number of specific characteristics that have been identified as potentially important determinants of success or failure.

There are also a few general factors in the process of design and implementation of policy instruments that appear as important. These interrelated factors include the existence of clear goals and a mandate for the implementing organization, the ability to balance and combine flexibility and continuity, the involvement of stakeholders, and the ability to adapt to and integrate adjacent policies or develop consistent policy packages.

It is not possible to conclude from the case studies whether it is important that the implementing agency is independent or not from the regulator. However, there is a strong indication that an organization or program with a clear mandate, responsibility, and adequate resources is a first and foremost prerequisite for success. The continuity of a program is important and can be ensured either by strong political backing within central bureaucracies or through creating more independent implementing agencies. Continuity in terms of stable and predictable conditions has been identified as an important success factor in several case studies. However, flexibility in the sense of being able to adapt to changing conditions or to remove potential fail factors in the implementation process is equally important. Technical and market developments, organizational changes, and changes in other policy areas may motivate adjustments. In several cases, the ability of the implementing agency to quickly remove specific barriers was noted as important. Such barriers include lack of information, tools, and skills that require the development of guidelines, procedures, analytical tools, education, etc.

The involvement of stakeholders in the design and implementation of policy instruments is noted as an important success factor in several cases. It serves the dual purpose of increasing the acceptance and, at the same time, improving efficiency by ensuring that stakeholder needs and preferences are considered throughout the process. Stakeholders include both the targeted actors and the various other actors that may be affected, e.g., energy consultants, equipment suppliers, and third-party auditing companies. A common trait is that the stakeholders appreciate simple and clear rules, as well as short lead times in cases where decisions or approvals have to be made by the implementing agency. Stakeholder involvement facilitates flexibility as noted above but also makes it easier for the implementing agency to benefit from other developments and conditions, for example, the existence of environmental management systems. An implementing agency with the ability to communicate well with all stakeholders but also with the mandate to adjust and evolve the policy instrument stands a better chance of being successful.

Conclusions and discussion

The case studies and the overall assessment of results presented in this paper lead to a number of observations concerning the usefulness of the policy theory approach, the experience with energy efficiency policy instruments, as well as the general success and fail factors. Based on these observations, policy implications in the context of the ES Directive are discussed.

The evaluations of 20 energy efficiency policy instruments applied across Europe, the USA, and Japan show that:

  • Energy efficiency policies often lack quantitative targets and clear timeframes.

  • Policy instruments often have multiple and/or unclear objectives.

  • The need for monitoring information often does not have priority in the design phase.

  • For most instruments, monitoring information is collected on a regular basis. However, this information is often insufficient to determine the impact on energy saving, cost effectiveness, and target achievement of an instrument.

  • Monitoring and verification of actual energy savings have a relatively low priority for most of the analyzed instruments.

  • There is no such thing as the ‘best’ policy instrument. However, typical circumstances in which to apply different types of instruments and generic characteristics that determine success or failure can be identified.

Despite the weaknesses noted above, as well as the diversity of instruments and contexts in which they were used, it is possible to make a few general observations concerning success factors in the process of design and implementation of policy instruments. Success factors include:

  • The existence of clear goals and a mandate for the implementing organization.

  • The ability to balance and combine flexibility and continuity.

  • The involvement of stakeholders.

  • The ability to adapt to and integrate adjacent evolving or new policies or develop consistent policy packages.

The instruments were evaluated by applying a practical framework based on theory-based policy evaluation. With the development of this framework, we aimed to contribute to a further harmonization of evaluation processes and create comparable evaluation outcomes among the case studies. We conclude that the theory-based policy evaluation method has several benefits over other ex post evaluation methods since:

  • The whole policy implementation process is evaluated, and the focus is not just on the final impacts (i.e., realized energy savings).

  • Through the development of indicators for each step in the implementation process, the “successes and failures” can be determined to the greatest extent possible.

  • By applying this approach, we not only learn whether policies are successful or not but also why they succeeded or failed and how they can be improved.

We experienced a number of practical problems that in real cases often make it difficult to exactly follow all steps of the developed methodology. These problems are mainly associated with the fact that evaluation and monitoring aspects did not have priority in the design and implementation phase of our case studies. Such problems include:

  • Lack of monitoring data and information. Baseline data or data on penetration rates, savings, awareness, etc., are often difficult and sometimes impossible to reconstruct.

  • Lack of time and resources mean that compromises have to be made regarding data collection and reconstruction. Ideally, decisions should be made early in the policy process regarding level of ambition and focus of the evaluation to identify important indicators and data.

  • The policy theory is not clear and not explicitly formulated during the design and implementation phase. When it has to be reconstructed, there is a risk of misinterpretation unless it is carefully checked with several informants.

  • In many cases, it is difficult to determine clear cause–effect relations. Integrating policy theory in the policy design phase would at least ameliorate some of these difficulties.

  • It may be difficult to identify success and fail factors, and determine their importance. Different respondents sometimes have different views on this, but through an iterative process including several respondents as well as factual information, a reasonably accurate representation can be constructed.

Based on the analysis and results, it is worth making a few observations in the context of the ES Directive and the 20% target for 2020. The ES Directive requires that savings are monitored and verified. Our analysis shows that experiences in this area from existing policy instruments in Europe are limited. Hence, the development of harmonized evaluation methods under the ES Directive starts from a relatively low level. It is also in the initial stage entirely focused on quantifying the savings. The analysis presented here shows that the policy theory approach can add considerable value in the process of policy design, implementation, and evaluation. Mainly, this results from systematizing the search for cause–impact relationships and explanatory factors behind the impact of instruments. In the design phase, it helps policy makers think about the whole implementation process, the relation and overlap with other policy instruments, and the relevant indicators to monitor so that evaluation becomes an integral part of the policy instrument and not a later add-on. During the implementation process, the approach can help identify problems and barriers at an early stage and allow the policy instrument to evolve and adjust to changing technical, organizational, economic, and other conditions. For verifying the savings, a policy theory approach can improve the quality of evaluations by ensuring that data needs and data gathering is an integral part of the policy. This is important not least for handling gross-to-net correction factors. Effective and efficient energy efficiency policies in terms of delivering large savings at low cost will be crucial for reaching energy and climate policy goals for 2020 and thereafter. For this purpose, evaluations must go beyond determining merely the savings as stipulated in the ES Directive and include approaches that facilitate continuous policy learning and policy transfer.