The improvement of Ex-Post Impact Assessment as a tool for Policy Making in the field of RTDI Peter Teirlinck – Hogeschool-Universiteit Brussel & Universiteit Antwerpen Acknowledgements This paper presents a summary of the literature review undertaken for FP7 OMC-net project 234501 ‘Optimising the Policy Mix by the Development of a Common Methodology for the Assessment of (Socio-) Economic Impacts of RTDI Public Funding’. Launched in March 2009, this project brought together a network of 15 partners - mainly public administrations and agencies - from around Europe (a mix of EU member states and Associated States) with the goal of exchanging information and mutual learning on national activities with respect to the assessment of the (socio-) economic impacts of public funding (policies and instruments) of Research, Technological Development and Innovation (RTDI). The work presented here takes on board the insights provided in a joint effort with Henri Delanghe (European Commission, DG Research). 1. Introduction Europe is facing major societal challenges related to global competition, the transformation to a knowledge and service economy, demographic sustainability, social justice, climate change, secure, sustainable and competitive energy, migratory pressure, and security and safety (European Commission, 2007). In response to these challenges the European Commission in close collaboration with the Member States presented - as a follow-up of the Lisbon strategy - ‘Europe 2020 - A strategy for smart, sustainable and inclusive growth’ (European Commission, 2010a). Instrumental to achieve such growth is the Flagship Initiative ‘Innovation Union’ (European Commission, 2010b) which states that close monitoring of the impact of various policy measures (mainly related to the creation of adequate framework conditions for innovation) is needed. Public European and, even on a larger scale, national (and regional) Research, Technology Development and Innovation (RTDI) funding plays an important role in making the EU 2020 ambitions real. The key role assigned to public funding relates to the existence of market failures, which prevent the private sector from investing in research at the socially desired optimum level (Arrow, 1962). Prior to public intervention governments must carefully design the way in which they will intervene (SEC, 2007), and RTDI support should be provided in the most appropriate areas (those with large spillover effects and where the private sector would not get involved in on its own) and through the most effective instruments. Increasing budget constraints for public funding in Europe (enforced by the economic and financial crisis starting in 2008-2009) enhances this selection process. Moreover, from a policy mix perspective (Nauwelaers, 2009) there is a growing demand to justify each euro spent for RTDI in comparison with other policy intervention areas (like e.g. health, environment, ...), as well as to consider the interactions and synergies between policy measures in different policy areas. In contrast with several other policy areas, assessing the socio-economic impacts of public funding for RTDI is not a common practice in Europe (Delanghe and 2 Teirlinck, 2009). In order to convince policy makers not to cut down research and innovation budgets, there is a necessity to pay more attention to the long term socioeconomic outcomes of research. This paper provides insights into a framework to do so by addressing methodological aspects and a focus on the missing link between impact assessment and policy design (Delanghe and Teirlinck, 2009). Section 2 defines ‘ex-post impact assessment’, presents the ‘logic intervention model’ as a framework for analyzing the impacts of public funding for RTDI, and highlights the fact that the policy cycle is not closed. Section 3 presents empirical evidence on shortcomings in existing impact assessment/evaluation exercises. The focus is on impact assessment studies ordered by public administrations. Section 4 concludes and provides recommendations for improving the usefulness of impact assessment exercises as a tool for policy making. 2. Logic intervention model and ex-post impact assessment in the policy cycle Assessing the socio-economic impacts of public policy is becoming increasingly important as the changing role and position of government has resulted in a growing demand for evidence-based policies (OECD, 2007: 4). Doing so allows policy makers to address fundamental questions: which intervention and support instruments work, which do not, and how can they be improved? In this section the concept of impact is defined and its interrelations with policy design are brought together in so-called logic intervention models. These models play an important role in ‘evidence-based policymaking’. 3 2.1 The logic intervention model and ex-post impact assessment Regardless of its nature (policy, programme, measure, project), a public intervention can be analyzed as a set of financial, organizational and human resources mobilized to achieve, in a given period of time, an objective or set of objectives, with the aim of solving or overcoming a problem or difficulty affecting targeted groups. Logic intervention models (Figure 1) can be defined as ‘logically consistent descriptions of the design of a programme as well as of the expected impact’ (Lengrand et al., 2006: 76). They help identify and set out the relationship between the socioeconomic needs to be addressed by the intervention and its objectives (initial statements of the outcomes intended to be achieved by an intervention), inputs, outputs, and outcomes. Outcomes include results (immediate changes that arise for direct addressees at the end of their participation in a public intervention) and impacts (longer term effects of the intervention). In contrast with results (which are directly influenced by the policy intervention), impacts can be defined as what a policy measure can influence only indirectly. They refer to longer term socio-economic consequences which can be observed only after a certain period after the completion of an intervention (hence expost impact assessment), and which may affect either direct addressees of the intervention or indirect addressees falling outside the boundary of the intervention, who may be winners or losers. Important to note is the broad scope (socio-economic, direct and indirect, positive and negative) and time dimension (longer term, observed after a certain period after the completion of an intervention) of impacts. 4 Figure 1: The Logic Intervention Model Target populations Global impacts Intermediate impacts Needs, Problems and issues Results Outcomes Design and implementation Objectives Inputs Outputs Efficiency Relevance Effectiveness Evaluation Utility and sustainability Source: Adapted from European Commission (2004: 72). See also European Court of Auditors (sd: 17). 5 The different building blocks of the logic model can be linked through different interrelations capturing ‘relevance’, ‘effectiveness’, ‘efficiency’, and ‘utility and sustainability’ (European Commission, 2004: 39). ‘Relevance’ relates to the extent to which an intervention’s objectives are pertinent to the needs, problems and issues to be addressed. Closely related to relevance are the extent to which the intervention logic does not contradict other interventions with similar objectives (‘coherence’) and the extent to which resources are available in due time, in appropriate quantity and quality, and at the best price (‘economy’). ‘Effectiveness’ refers to the extent to which the objectives set are achieved in the intended results and impacts. ‘Efficiency’ refers to the extent to which the desired outcomes are achieved at a reasonable cost. The principle of efficiency is concerned with the best relationship between resources employed and outcomes. A last interrelation links needs with results and impacts. ‘Utility’ refers to the extent to which outcomes correspond with the needs, problems and issues to be addressed. ‘Sustainability’ means the extent to which positive outcomes are likely to last after an intervention has stopped. ‘Ex-post impact assessment’ focuses in particular on the broad, longer term impacts of public interventions. As such it looks mainly at effectiveness, utility and sustainability. It should be seen in relation to the broader activity of evaluation which can be defined as ‘a matter of seeing how well a policy or programme has achieved the objectives set for it’ (Lengrand et al., 2006: 31). The focus on the broad, longer term impacts of public interventions raises the important challenge to look beyond the single policy, programme, measure or project to the broader policy mix. The policy mix refers to the combination of policy instruments, which interact to influence the quantity and quality of R&D investments in public and 6 private sectors. Policy instruments are all programmes, organizations, rules and regulations with an active involvement of the public sector, which intentionally or unintentionally affect R&D investments. Interactions point to the fact that the influence of one policy instrument is modified by the co-existence of other policy instruments in the policy mix (OECD, 2010). 2.2 Ex-post impact assessment and closing the policy cycle The policy cycle is generally conceived to consist of four basic stages. In stage one, the ‘agenda-setting, problem identification’ stage, policy problems are defined and policy issues are raised, introduced to the political stage by different governmental institutions, individuals, interest groups, or specific events. In stage two, the ‘policy formulation’ stage, analysis and politics determine how the agenda item is translated into legislation. This encompasses several aspects including the development and consideration of alternative policy options, the selection of a preferred option, and its adoption. Next, in stage three, the ‘implementation’ stage, the adopted policy is implemented (administered and enforced) by the bureaucracy, or by a government agency. The bureaucracy or agency translates the policy into a concrete set of actions, and makes judgments as to the intent, goals, timetables, program design, and reporting methods. Finally, in the ‘evaluation’ stage, the implementation of policy is evaluated in order to find out what is working and what is not. The impacts of the policy are assessed. If goals exist, the effectiveness of the policy and its components can be determined and side-effects have to be discovered and evaluated. 7 Figure 2 presents the place of ex-post impact assessment in the policy cycle taking into account that (i) the policy formulation process is supported by ex-ante evaluation/impact assessment; (ii) the policy implementation process is supported by monitoring and interim evaluation; and (iii) that the evaluation process consists of expost evaluation and ex-post impact assessment. Figure 2: Ex-post impact assessment in the policy cycle Agendasetting Policyformulation Implementation Evaluation Ex-ante evaluation Monitoring Interim evaluation Ex-post evaluation Ex-post Impact assessment Source: Delanghe and Teirlinck (2009). 8 The different activities underpinning evidence-based policy-making assume the closure of the policy cycle through full feedback effects and learning from monitoring, interim evaluation, ex-post evaluation and ex-post impact assessment. The closure of the policy cycle depends first and foremost on the usefulness for policy making of the evaluation product. This is determined by a variety of factors including (Harty and Newcomer, 2004 and Bayhan et al., 2011): the degree of clarity of the initial intervention logic (and the feasibility of the assessment); the degree of stakeholder involvement and consensus on evaluation criteria and expectations; the timing of the assessment; the extent to which failures are assessed as well as successes; the degree to which reasoned, pre-tested methodologies and indicators are used to generate information which is easily understood and interpreted; the extent to which the information produced is relevant for broader programme and policy design rather than highly context-specific; the balance between data collection and analysis; and the extent of dissemination of information produced. In the next section the obstacles hampering feedback effects from ex-post impact assessment in policy design will be looked after in an empirical way. 3. Use and usefuleness of ex-post impact assessment exercises This section looks in an empirical way at the use (section 3.1) and usefulness (section 3.2) of impact assessment practices in Europe. With regard to the use of impact assessment, it is investigated to what extent impact assessment is implemented at 9 country level in Europe. With regard to usefulness, attention is paid to factors that hamper the usefulness of ex-post impact assessment for policy making. As starting point of the analysis we take 21 ex-post impact assessment exercises ordered by government institutions. The focus of the analysis is on ex-post impact assessment/evaluation exercises in three policy clusters: public funding of private RTDI, public funding of industry-science relations, and public funding of public research organizations. The clusters are related to challenges for National Innovation Systems in most European countries related to: (i) low R&D investments and the need to increase the overall R&D spending (especially in the private sector); (ii) a need for improvement of the interactions between industry and science actors; and (iii) a better governance of research and innovation in public research organizations. 3.1 Use of ex-post impact assessment at country level in Europe In this section a global analysis of the R&D and innovation policies of 27 EU countries will be offered. The used information is based on the database of the European Inventory of Research and Innovation Policy Measures created - on behalf of the European Commission - by the ERAWATCH organization, and combined with a questionnaire among the eleven countries participating in the OMC-net project. Based on an analysis of the qualitative data of the ERAWATCH-website, the evaluation of R&D and innovation policies turns out to have developed strongly in European Union Member States. However, it seems to have become a standard practice in some countries only, and the countries can be classified in four groups based on the 10 existing evaluation culture (Eparvier, 2009). A first group brings together leading countries in evaluation. At present, these countries (Austria, Denmark, Finland, Germany, Ireland, the Netherlands, Sweden and the United Kingdom) have a wellestablished culture of evaluation and carry out systematic evaluations of programs and institutions. A second group is made of countries that have strongly reinforced evaluation practices, structures and culture (Belgium, Czech Republic, Estonia, France, Portugal and Slovenia). These countries recently engaged in a process of systematizing evaluation of programs and of institutions. A third group gathers countries that have recently established or that are about to set up evaluation practices, structures and culture (Bulgaria, Italy, Luxembourg, Malta, Romania and Slovakia). A last group is composed of countries that do carry out evaluations but not on a systematic basis (Cyprus, Greece, Hungary, Latvia, Lithuania, Poland and Spain). Based on an in-depth analysis of 814 instruments (for which detailed information was available) used in the 27 EU Member States (ERAWATCH Research Inventory Report: overview across EU countries - http://cordis.europa.eu/erawatch) the classification of the use of evaluation in Europe could be made as presented in Table 1. 11 Table 1: The ex-post evaluation/assessment culture: a qualitative versus quantitative assessment Qualitative assessment of the evaluation culture included in the ERAWATCH country profiles published in the web site Percentage of instruments culture evaluated ex post Systematic or Low evaluation Frequent but not well established systematic evaluation evaluation (Turkey*) culture Czech Republic, Less than 10% Lithuania, Portugal, (Cyprus*) Hungary, Poland, Belgium Malta, Greece Between 10 and 30% Luxembourg Spain Italy Romania, France, Netherlands, Ireland, Sweden, Austria, United Slovenia Kingdom Between Slovak Republic, 30 and 40% Latvia, Estonia Over 45% Bulgaria Finland, Denmark, Germany Source: Delanghe et al., 2011. *No qualitative assessment was obtained for Cyprus and no statistical data were available for Turkey. 12 A more detailed analysis of funding measures in the three target policy clusters revealed that each cluster accounts for about 10% of the total RTDI government budgets. Important with respect to policy mix and causal linkages between policy measures and outcomes is that more than half of the policy instruments target several policy clusters at the same time. In terms of evaluation, more than half of the instruments have not been evaluated so far and there are marked differences in propensity to evaluate according to the policy area and according to the country. Specifically with regard to ex-post evaluation, the propensity to do an assessment differs according to the type of policy instrument and according to the country of origin. Instruments aimed at the direct or indirect promotion of private R&D investments are evaluated in one out of four cases when it concerns indirect support for R&D; for direct support this amounted to one out of six. About half of the instruments related to public research organizations and universities were evaluated ex-post. Finally, close to onefourth of the instruments focused on public-private R&D cooperation and of the measures considered in support of infrastructure for public-private linkages were evaluated ex-post. In the case of the instruments to foster knowledge transfer, this percentage was almost 14% (for more detailed information, see Delanghe et al., 2011). An additional survey among the OMC-net project partners identified that the planning of impact assessment in the policy cycle remains predominantly ad hoc. Nevertheless, there is an important trend towards more interest in the use of impact assessment in RTDI policy making (Table 2). 13 Table 2: Planning of impact assessment in the policy cycle for policy measures funding RTDI Extent of planning of impact assessment in the policy cycle for policy measures funding RTDI Occasional Not/seldom – ad hoc Systematic Austria (Legal Bulgaria Recent and Estonia planned changes in the Increased Malta use of impact use Spain assessment in Turkey RTDI policy Iceland making Austria (for Basis for funding not universities)* oriented at Czech Republic universities) (for institutional Belgium support for research France - focused at Sweden (Not scientific impact!) VINNOVA - Sweden Focused at (VINNOVA – scientific focused at impact in impact) terms of sustainable development) No change Note: in each of the countries having experience in impact assessment the results of these exercises are used mainly for the design of new policy measures or for the improvement of existing ones. *The systemic planning of impact assessment of Austrian universities is indicated in Austrian law, but de facto not reality until now. 14 3.2 Usefulness of ex-post impact assessment exercises In the analysis of the impact assessment exercises attention is paid to ‘methods, data collection and indicators’ and to ‘user involvement and recommendations’. The focus of the analysis is on the 21 ex-post impact assessment exercises (presented in Table 3) that have been selected from 7 countries and that have been considered important exercises at country level since the year 2000. A common characteristic of these exercises is that they are ordered by government institutions. 15 Table 3: Overview of impact assessment exercises in the field of public funding for RTDI Country Impact assessment study Intellectual Property Rights at Austrian Universities: Evaluation of the Austria uni:invent programme Evaluation of the Austrian Industrial Research Promotion Fund (FFF) - Austria Impact Analysis Impact of the Federal Scientific Institutions on the economic, scientific, social Belgium and cultural development in Belgium VRWB research: impact analysis of fiscal stimuli on R&D expenditures in Belgium Flanders Belgium The impact of public R&D funding in Flanders A look into the Black Box: What difference do IWT R&D grants make for Belgium their clients Making the difference: The evaluation of 'Behavioural Additionality" of R&D Belgium subsidies. Bulgaria Peer-review of the national R&D system under the OMC forth cycle Bulgaria Evaluation of the National Science Foundation (NSF) PREDIT 3: National research and innovation programme in terrestrial France transports 16 Evaluation of PREBAT (Research and exploitation of energy in buildings France 2005-2009) Impact study SRC 2006 / impact of French research and technology France organizations (RTO's) France Report on the valorization of research Evaluation of the first phase (2006-2008) of the policy on competitiveness France clusters (pôles de compétitivité) France Competititveness clusters (pôles de compétitivité), what can we expect? Which articulation between PRES, RTRA and competitiveness clusters (pôles France de compétitivité) Information report on the economic incidences of an augmentation of research France expenditures in Europe Impact of the research tax credit on the funding of private R&D: an France econometric evaluation Spain Impact Assessment of the FP6 in the Spanish R&D public system Spain Impact of R&D on the Industry Turkey An Assessment of the Industrial Technology Project To collect and harmonize the information for the 21 studies, a survey has been set up in order to get the maximum synergy possible with other works already done in the area, mainly the Inno-Policy Trendchart repository of European policy measures and 17 Inno-Appraisal repository of European evaluation studies (http://www.proinnoeurope.eu/index.cfm?fuseaction=page.display&topicID=262&parent ID=52). Four types of impacts (scientific, technological, economic, and societal) were considered. Economic impacts were most prominent (86% of the cases), followed by scientific impacts (48%). In less than one third of the cases broader societal impacts were investigated. The latter can be closely related to the fact that only 2 out of the 21 studies had a broader societal impact as initial objective. Technological impacts were the focus in less than one fourth of the studies. 3.2.1 Methods and indicators Insights into the use (good practices) and possibilities for harmonization in terms of methodologies and indicators for impact assessment are crucial elements to promote impact assessment at the level of policy makers and to stimulate them to make more use of impact assessment in the policy making process (OECD, 2009). Methods The harmonized analysis of the 21 case studies reveals that 40% of the exercises are not based on previously designed methodologies. Moreover, in close to eighty percent of the cases these methodologies have not been used for other impact assessment exercises of the same policy measure (this case sensitivity can be related either to the fact that the methodology was not considered appropriate or to the fact that no other impact assessment exercises have been performed). 18 The focus of the studies is mainly on the consequences or results of the measure in terms of ‘outcomes’, and to verify if the goals of the measure were achieved. Moreover, considerable attention is paid to input, output, and behavioural additionality. Programme or project implementation efficiency as well as policy/strategy development are covered in less than one third of the exercises. This seems to indicate that impact assessment exercises not directly aim at policy (re-)design or fail to do so. 19 Table 4: Topics covered in impact assessment exercises of public funding for RTDI Programme Implementation Outcomes 86% Efficiency 24% Goal Attainment/Effectiveness 62% Quality of outputs 24% Output additionality 38% Internal consistency 14% Behavioural additionality 33% Coherence/Complementarity 14% Input additionality 33% External consistency 5% Project implementation Policy/Strategy Development 29% Efficiency 5% 24% Gender issues 5% Value for Money/Return on Investment/Cost-Benefit Analysis Source: Questionnaire on impact assessment for public funding of RTDI, analysis of the 21 case studies presented in Table 3. Note: responses may be biased due to difficulties for institutions to criticize exercises they or other ministries/agencies have ordered. 20 The appropriateness of the design of the study, the chosen methodology related to the objectives of the impact assessment and the nature of the policy measure - on average can be considered fairly appropriate (measured on a five-point Likert Scale). Lack of experience and data and too much hypotheses underlying the relationship between RTDI funding and impacts are the hampering factors for the appropriateness of the design of impact assessment exercises. Complexity to measure net impacts is a main barrier for the usefulness for mutual learning of these exercises. Indicators Preferred ways of data collection include participant surveys, existing databases and interviews. Descriptive, case-study analysis and econometric analysis are the main data analysis methods for impact assessment exercises. About three-fourths of the studies combine qualitative and quantitative approaches for data analysis. Barely in one third of the cases impact indicators are considered at the set-up of the policy measure (or in other words included in the logic intervention model). Most exercises make use of existing indicators. Only in one case completely new indicators in the field of impact assessment were created. 21 Table 5: Appropriateness of indicators for impact measurement Impact Appropriateness of indicators for the Efficiency (in measurement of impacts terms of Not Sometimes appropriate appropriate Appropriate information provided/ resources devoted)* Scientific 11% 33% 56% 2.57 Technological 0% 50% 50% 2.67 Economic 0% 27% 73% 2.86 83% 0% 1.00 Societal (including 17% environmental) Source: Questionnaire on impact assessment for public funding of RTDI, analysis of the 21 case studies presented in Table 3. * Average efficiency: average based on a 5-point Likert scale (not efficient = 0; moderately efficient = 1; average efficient =2; fairly efficient = 3; very efficient = 4). 22 The high reliance on existing indicators could explain the limited perceived appropriateness of the indicators, especially with regard to societal, technological and scientific impacts. Important in relation to the broad EU 2020 challenges is the finding that in none of the exercises indicators were found fully appropriate to measure societal impacts. This could be related to the context specificity and to the need to consider a broader policy mix. For example: the impact on society due to both research and norms & legislation when considering the decrease of mortal accidents, or the decrease of the emission of greenhouse gases. In a considerable number of cases indicators to measure scientific and societal impacts were found completely inappropriate. Not surprisingly, these findings resulted in a lower perceived efficiency of the indicators to measure societal impacts. Important shortcomings in terms of indicators are data availability (including the high number of potentially useful indicators and the possibility to make comparisons) and lack of appropriateness to adapt to specificities in terms of the actors benefiting from the public funding. 3.2.3 User involvement, recommendations and consequences Based on the analysis of the 21 case studies, several hampering factors for the usefulness of impact assessment as a tool for policy making have been identified. A main finding is that too little attention is paid to policy/strategy development and to project implementation efficiency in impact assessment. This means that the policy cycle is not closed. This can be related to the fact that evening less than half of the impact assessment studies ordered by government administrations/agencies, politicians 23 and the program management are involved. This confirms the existence of a gap between policy formulation and implementation. Moreover, relatively low involvement can be witnessed of those directly or potentially supported by the policy measure or programme. Therefore, it comes as no surprise that impact assessment exercises mainly provide recommendations for redesign, expansion or management of the programme/measure under consideration. Related to the policy mix, broader policy design is not found to be an objective in more than half of the impact assessment exercises and the influence and relation with other policy measures (e.g. in terms of re-design or merger) are limited. 24 Table 6: Composition of the user group of the impact assessment exercise Policy Makers (Government Potential Users of the Policy officials) 95% Measure 33% Programme Management 48% Auditors/Financial Authorities 14% Policy Analysts 48% General Public 14% Those directly supported by the External/ (co)sponsor of the measure 43% Policy Makers (politicians) 43% measure 10% Source: Questionnaire on impact assessment for public funding of RTDI 25 Table 7: Recommendations and consequences of impact assessment exercises % of % of Recommendations? cases case Consequences? s Termination/merger of the No 14 measure 0 52 Re-design of the measure 52 management 48 Re-design of another measure 14 Yes, targeted at broader policy design 43 Expansion of the measure 29 No consequences 38 Yes, targeted at the design of the measure Yes, targeted at the programme Source: Questionnaire on impact assessment for public funding of RTDI 26 4. Reflections on how to improve impact assessment as a tool for policy making Over the last decade increasing resources have been allocated to RTDI policies and expost impact assessment has been stepped up with some delay. Progress has been in methods for impact assessment and more data are now available. Thus, there is both more demand for the assessment of policies that absorb increasing resources and more supply in terms of sophisticated methods, data, and experience. However, there are marked differences in the practice of impact assessment according to the policy area under consideration and also between countries. On average, in Europe, impact assessment of public funding of RTDI still rather occurs in an ad hoc manner and little attention is paid to the broad societal impacts of public funding for RTDI. Assessing the economic, scientific, technologic and societal impacts of RTDI policies is not a straightforward task. Methodologically there are several challenges, as there are challenges concerning the utilisation of impact assessment findings and results in the design of policy measures. Four major methodological lessons can be derived from the impact assessment exercises. First, a need for awareness of the difficulty to measure (quantify and qualify) indirect economic and technological effects (not least by the receivers of public funding for RTDI). Second, the importance to consider different policy measures at the same time (policy mix). Also, there is a need for recognition of problems in terms of attribution (both to specific measures and to specific territories) and timing. Third, awareness is necessary of the lack of data to analyze and to benchmark the impacts resulting in a lack of evidence-based conclusions. A final lesson is the need of experienced people in evaluation/impact assessment itself, especially to bridge the gap between the science community and policy stakeholders. 27 Also with regard to the full utilization of impact assessment and its results for the optimization of RTDI policies, several hampering factors can be identified. A main barrier consists of the lack of (early) involvement of government administrations/agencies, politicians and the programme management agencies/organizations in impact assessment studies. This results in a disrupted feedback loop between policy formulation, implementation and policy optimization. A second important hampering factor relates to the relative isolation in which impact assessment is carried out. This means that influence on, and relation with, other policy measures (other instruments in the policy mix) is often (too) limited. A final main lesson of the mutual learning enhanced during the production of this review is the recognition that different terminologies are used with regard to evaluation and impact assessment. The logic intervention model has been presented as a common methodological approach to impact assessment. It offers the advantage to better link policy objectives with long term impacts and to consider methods and indicators appropriate for doing so from the moment the policy measure is introduced (i.e. in an ex-ante manner). References Arrow, K.J. (1962): Economic Welfare and the Allocation of Resources for Inventions, in Nelson R.R. (ed.), The Rate and Direction of Inventive Activity: Economic and Social Factors. Princeton University Press, Princeton. Bayhan, D., Bukulmez, E., Ozdemir, A.H., Gok, A. and Edler, J. (2011): Governance and Usefulness of Impact Assessments. Chapter 3 in Teirlinck, P. (ed.) Lessons for ex- 28 post impact assessment of public funding for RTDI in the policy cycle - Findings from FP7 OMC-net project 234501: Optimising the Policy Mix by the Development of a Common Methodology for the Assessment of (Socio-) Economic Impacts of RTDI Public Funding (CIA4OPM), European Commission, Brussels. Delanghe, H. and Teirlinck, P. (2009): Optimising the Policy Mix by the Development of a Common Methodology for the Assessment of (Socio-)Economic Impacts of RTDI Public Funding (CIA4OPM), DSTI/STP/TIP(2009)18, pp. 67, OECD, Paris. Delanghe, H., Teirlinck, P., Heijs, J. and Sachwald, F. (2011): Practice of Impact Assessment in the Policy Cycle, Chapter 2 in Teirlinck, P. (ed.) Lessons for ex-post impact assessment of public funding for RTDI in the policy cycle - Findings from FP7 OMC-net project 234501: 'Optimising the Policy Mix by the Development of a Common Methodology for the Assessment of (Socio-) Economic Impacts of RTDI Public Funding (CIA4OPM), European Commission, Brussels. Eparvier, P. (2009): ERAWATCH Research Inventory Report for: Overview Across EU Countries. European Commission. European Commission (2004): Evaluating EU Activities – A Practical Guide for the Commission Services, Office for Official Publications of the European Communities, Luxembourg. European Commission (2007): Europe in the global research landscape. European Commission, DG Research, Brussels. European Commission (2010a): EUROPE 2020 - A strategy for smart, sustainable and inclusive growth, 3.3.2010, COM (2010), European Commission, Brussels. 29 European Commission (2010b): Europe 2020 Flagship Initiative Innovation Union – COM (2010) 546, European Commission, Brussels. European Court of Auditors (sd), Performance Audit Manual, Court of Auditors, Luxembourg. Hatry, H.P. and K.E. Newcomer (2004), ‘Pitfalls of Evaluation’, in: J.S. Wholey, H.P. Hatry and K.E. Newcomer, Handbook of Practical Program Evaluation. Lengrand, L. & Associés, PREST, ANRT, Reidev Ltd. (2006), SMART INNOVATION: A Practical Guide to Evaluating Innovation Programmes, ECSC-ECEAEC, Brussels-Luxembourg. Nauwelaers, C. (2009), Policy Mixes for R&D in Europe. UNU-MERIT, April 2009. A study commissioned by the European Commission - Directorate-General for Research. OECD (2007), Assessing the Socio-Economic Impacts of Public Investment in R&D, OECD, Paris. OECD (2009), Enhancing Public Research Performance through Evaluation, Impact Assessment and Priority Setting, OECD, Paris. OECD (2010), STI Outlook, OECD, Paris. SEC (2007), SEC(2007)213 of 21/02/2007 Responding to strategic needs: Reinforcing the use of evaluation. http://ec.europa.eu/budget/library/documents/evaluation/ communication/eval_comm_sec_2007_213_en.pdf. 30
© Copyright 2025 Paperzz