Handbook
Handbook
Evaluation of tax incentives for R&D: an overview of issues and considerations A paper produced by the Crest OMC Working Group Evaluation and design of R&D tax incentives
Introduction................................................................................................................. 3 Considering evaluations when designing the scheme................................................. 5 Organising the evaluation ........................................................................................... 6 3.1 When should the evaluation be carried out? ....................................................... 6 3.2 Planning the evaluation....................................................................................... 6 3.3 Who should perform the evaluation?.................................................................. 7 3.3.1 Internal evaluations..................................................................................... 7 3.3.2 External evaluations.................................................................................... 8 3.3.3 A combination of internal and external evaluation..................................... 9 3.4 Independence and publishing evaluations .......................................................... 9 4 Issues to be evaluated................................................................................................ 10 4.1 General evaluation issues.................................................................................. 10 4.1.1 Input additionality: Does the scheme generate more R&D? .................... 10 4.1.2 Result or output additionality: What are the effects of the investments? . 10 4.1.3 Behavioral additionality: Do firms change their R&D strategy?.............. 11 4.1.4 Administrative costs and efficiency.......................................................... 12 4.2 Examples of more specific evaluation issues.................................................... 12 4.2.1 How the scheme works together with other R&D measures .................... 12 4.2.2 The role of research institutes ................................................................... 12 4.3 Policy recommendations................................................................................... 12 5 Methods of evaluation............................................................................................... 14 5.1 Different methods of evaluation ....................................................................... 14 5.1.1 Event studies ............................................................................................. 14 5.1.2 Case studies and surveys........................................................................... 15 5.1.3 Econometric studies .................................................................................. 15 5.2 The absence of random variation: A problem for all methods ......................... 18 5.3 Summing up: Different questions require different methods ........................... 19 6 Data issues ................................................................................................................ 21 6.1 What is the ideal situation? ............................................................................... 21 6.1.1 Creating an evaluation database................................................................ 21 6.1.2 Data describing firms................................................................................ 21 6.1.3 Data describing the subsidized research projects: .................................... 22 6.2 What are the possible data sources?.................................................................. 23 6.2.1 Administrative registers ............................................................................ 23 6.2.2 Regular censuses and surveys carried out for other purposes................... 23 6.3 How can shortcomings be remedied at what cost? ........................................ 24
1 2 3
1 Introduction
This paper is part of the work of the OMC Working Group on the evaluation and design of tax incentives for R&D. The purpose is to give a general overview and considerations concerning the evaluation of tax incentives for R&D. More in particular, the intention is to highlight important stages and choices in the evaluation process that policy makers and operating agencies might find useful to take into consideration. The evaluators themselves will have more detailed and operational knowledge than what is covered here. There is no single recipe for doing evaluations evaluations of tax incentives for R&D are no exceptions. Although the primary objective most often is to increase R&D spending within firms, there may be a range of other objectives that evaluations should capture. In addition the national circumstances may vary and evaluations of tax incentives should be seen in the broader context of the national innovation systems and mix of government tools for innovation and R&D. Last but maybe most important access to various data sources varies across countries and can have a large impact on how evaluations may be carried out. Evaluations give a stronger knowledge base for making policies, but evaluations in themselves do not create policies. There has to be a translation of findings from evaluations to policy making, and policy makers have to understand and be able to use findings from the evaluations in their work. Evaluations should therefore be seen as a tool for learning that should be used in a proper way to improve policy making. Since the political systems differ across EU/ CREST member states this makes for a wide range of approaches in the follow up of evaluations. Evaluations do not give answers that are final. This goes for evaluations in general and for evaluations of both direct and indirect R&D support measures. Evaluations can, however, be used both to develop a stronger knowledge base for targeting certain policy aims and objectives and in order to create a stronger empirical basis for research concerning innovation and R&D. These are all themes that are addressed in the report of the OMC Working Group on the evaluations and design of R&D tax incentives. This paper is not meant as a detailed handbook, but shows important items to be aware of when planning and doing evaluations of tax incentives for R&D. From the outset the Working Group had as its mandate to produce a report. During its work, the group was requested to expand its work on evaluation of tax incentives. Because of the need for policy makers to get a better understanding of how evaluations might be carried out, it was decided to do this through a paper with an overview of issues and general planning-advice on evaluation methods. In this process, the Working Group had a need to get on board more expertise, and it is greatly appreciated that Torbjorn Haegeland, Head of Research at Statistics Norway, and Bruno van Pottelsberghe, Director at the European Patent Office, agreed to take part in the final rounds of discussion on the report and the handbook. Torbjorn Haegeland is also the main contributor to the writing of the handbook.
This paper is not intended as a detailed handbook and should not be taken as such. Rather, it aims to highlight and explore important issues that might usefully be considered when planning and conducting an evaluation of tax incentives for R&D.
One of the most important determinants of when an evaluation should be performed is when it will be possible to estimate the different effects of the policy. This will reflect both the period over which effects emerge (at least sufficiently to be quantified) and the time needed for sufficient data to become available to measure the effect. The additionality effects of R&D tax incentives, for example, can take a number of years to emerge and the methods often used to measure such effects can require a number of years of data. The risk of attempting such an evaluation too early is that misleading results might be produced, for example showing no policy effect when it is simply too soon for effects to show up. On the other hand, however, it is important for an evaluation plan to recognise that results are needed to inform improvements to the policy and so, as far as possible, evaluations should also aim to produce some results as early as possible. This might be in areas where effects can be seen and assessed more quickly, such as on the administration and delivery (for example, testing the awareness and take up of the policy and the ease with which firms can understand and claim support)
potential evaluation methods to tackle these questions; consideration of the counterfactual and how it will be measured; data that will be needed and where it will come from; how the different methods will fit together; the expected outputs from the evaluation (such as reports); what results will be published, how and when; the timescale for the evaluation; and the cost of the evaluation
It may be helpful to draw up an evaluation plan or strategy setting these issues out. This will form a basis for reference when it comes to actually performing the evaluation and will provide a benchmark against which progress can be measured. An evaluation plan can also help to manage expectations amongst policy makers and politicians of when they can expect different results to become available. It can also be useful to open the evaluation plan to peer review, such as from policy makers and other evaluation experts who can provide advice. This should help to test, for example, whether the evaluation questions will inform policy improvements, whether the methods are appropriate, and whether the timetable is practical and realistic.
their association with the policy. Internal evaluation may also be of poorer quality methodologically or less efficient, if internal evaluators are less aware of the latest developments in evaluation methods in the area or where the absence of competition has allowed inefficiencies to survive. Some of the disadvantages of internal evaluation can be resolved by exposing the evaluation to some level of public scrutiny. This could be through some form of peer review, either of the end product of the evaluation or at various interim stages whilst the evaluation is carried out. Making the results of the evaluation publicly available may help to raise the perceived quality of the work, although it may still leave some doubts over the quality of the methodology, particularly amongst those not experienced in evaluations.
4 Issues to be evaluated
The main task for evaluations of government measures is to find out to what extent the objectives of the measure is fulfilled, and whether the benefits to society generated by the measure are larger than the costs involved. Evaluations of tax incentives for R&D should of course have the same focus. The main objective for tax incentives for R&D is generally the same across countries: to generate benefits for society as a whole by inducing an increased level of R&D investments, with returns exceeding costs involved. Though they share the same main objective, different schemes have different specific objectives. These specific objectives may be regarded as different strategies to achieve the main objective of the scheme. Examples of specific objectives may be to increase the level of R&D in SMEs, to stimulate collaboration between firms and R&D institutes or to stimulate the creation of knowledge-based, research-intensive firms. Since evaluations should focus on the objectives of the measure in question, the issues to be evaluated will vary between countries, although some of the main components will be identical. In the following, an extended checklist of issues that should be included in most evaluations is presented. Following this issues that could be included, depending on the objective of the scheme is shown.
10
Evidence suggests that the returns to such investments on average are high. However, the literature also points out that there is vast heterogeneity in the returns to R&D investments. This is partly due to general idiosyncrasies, but there may also be systematic differences between types of investments. Hence, there may be differences between types of government support with respect to the results of the investments they generate. Input additionality may be a necessary condition for the success of a scheme, but not a sufficient condition. It is therefore necessary to evaluate output or result additionality: whether and how the R&D investments cause improved economic performance of the firms undertaking them and for society as a whole. One may think of the following "chain of results" for a tax incentive:
The first effect is input additionality, while the subsequent effects relate to output/result additionality. In addition to increased pace of innovations this might be increased quality and quantity of innovations. There may also be considerable external effects, i.e. effects accruing to others. It should be noted that result additionality is more demanding to identify and quantify than input additionality. The main reason for this is that effects are in general more delayed. In addition the effects of R&D may be difficult to distinguish from other factors that are important for profitability and enterprise external effects are particularly difficult to ascertain. There may also be a range of other objectives for tax incentives that give reason for evaluations that cover other results than those described here.
11
Does learning about benefits of R&D lead to sustained higher levels of investment?
Policy recommendations must take into account that there are variations as to what extent different effects can be observed and identified. When effects occur may also vary and not all effects can be expected to be observed within the evaluation period.
13
5 Methods of evaluation
This section gives a brief introduction to different methods that are widely used in evaluations. It should be noted that what follows is far from a detailed "how-to-evaluatetax incentives" guide. It is rather an attempt to give a short and fairly non-technical description of the different approaches, pointing out strengths and weaknesses of the different methods in addressing different aspects of the scheme. What should be learned from this section is that there is no "magic bullet." There is no single method that can be used to throw light on all the aspects of the scheme that is to be evaluated. Different methods have their strengths and weaknesses with respect to the questions they are suited to analyse, and their costs, data requirements, etc. The design of the tax incentives themselves may also complicate evaluations when it comes to the identification of causal effects.
14
between planned and actual R&D than the tax incentive. Event studies are most suitable to study sudden events where effects materialize quickly. Neither of these criteria is likely to be fulfilled in the case of tax incentives for R&D.
15
Input additionality Hall and van Reenen focus their discussion on estimating the input additionality effect, through econometric estimation of demand equations for R&D. Within this framework, there are basically two approaches. The point of departure for both is a regression equation that predicts R&D investments at the firm level. In the first approach, the demand equation includes a variable that indicates whether the firm had access to the tax incentive, in addition to other variables that affect R&D investments: ln( R & D) it = + C it + X it + u it This equation expresses the logarithm of the R&D investments of firm i in year t as a function of the presence of a tax incentive (Cit equals one if firm I had access to the scheme in year t and zero otherwise) and other variables, which are contained in the vector Xit. Such variables may be previous R&D investments, previous output and sales, expected future output, cash flow, product prices, etc. Whether the firms have been granted R&D subsidies through other channels will also be an important factor. The -parameter measures the expected growth of R&D investment following a firm getting access to the scheme. The basic framework assumes that this effect is identical across firms. This assumption is hardly innocuous. It may be relaxed but this often increases the data requirements. Such models should be estimated on micro data, to utilize cross-section variations in access to the scheme across firms. Using macro data only, it is impossible to distinguish the effects of the scheme from unobserved macroeconomic shocks. The second demand equation approach has very much in common with the one described above. The major difference is that instead of just including a variable indicating existence of or access to a tax incentive for R&D, one calculates the so-called user cost of R&D investments, i.e. a variable that reflects the price of R&D investments for the firm, on the margin, taking into account R&D tax incentives, other tax rules, interest rates and depreciation. The introduction of say a tax deduction scheme for R&D will reduce the user cost of R&D. ln( R & D) it = + it + X it + u it The key advantage of the user cost approach compared to the first approach is that one may utilize variations in the generosity of the scheme between different firms, and also changes over time. Such variation may be very useful in identifying the effect of the scheme. In addition, variations in other components of the user cost (tax rules, interest rates, depreciation rates) may in theory help in identifying the effects of the tax incentive. The equations above implicitly assume that the effects of the tax incentive are homogeneous across different firms. A relevant evaluation question is whether the effects differ with respect to observed firm characteristics, e.g. previous R&D experience, the education level of the staff, industry and location. If one should find that the scheme has larger effects for some types of firms than for others, it may influence the future design of
16
such schemes. In principle, it is straightforward to extend the framework described above to allow for such different effects. However, the number of observations may limit how many different additionality parameters that may be estimated with a reasonable degree of precision. Some schemes have differential treatment for different types of R&D projects, e.g. a higher maximum tax deduction for joint projects with R&D institutes. The rationale behind this is a supposed higher effect of such projects. It is straightforward to extend the framework above to test this assumption. The short-term input additionality effects may differ from the long-term effects. The long-term effects may be larger because there may be substantial adjustment costs in R&D investments. Such costs are due to the fact that it takes time to adjust the use of inputs to new external conditions. In addition, the long-term effect may be larger because an increase in R&D investments adds to the firms knowledge base, thereby increasing the marginal payoff of future R&D investments. Evaluations should take these possible differences into account. Result additionality Econometric methods may also be useful in trying to evaluate the results (result additionality) of the R&D investments generated by the scheme. In some sense, the econometric framework for analyzing the results is similar to the framework for analyzing input additionality, but with some modifications. As discussed in the previous section, one may think of the effects of the tax incentive that accrues to the firm as coming in the following steps:
The first step belongs to input additionality, while the next steps belong to result additionality. Hence there are many candidates for variables that may be indicators of result additionality. The number of patents is one obvious indicator. Productivity (labour productivity, total factor productivity), total sales, profitability, number of employees etc. are all variables where effects of increased R&D investments in principle should show up. Hence, they are candidates (among others) as dependent variables in the analysis of result additionality. It is challenging to identify the effects of the tax incentive on R&D investments. Identifying the effects of the R&D investments they are in many respects even more demanding. The further along the result chain above that the evaluation goes, the more external factors may affect the result and it may be difficult, or even impossible, to control for these factors. Results will probably come with delays of variable and unknown length (more on the timing issue below). This also makes it harder to identify the result additionality. In
17
other words, the hope for clear and definitive answers is smaller the further away from the tax incentive the analysis is done. The analysis of result additionality should take into account that one cannot distinguish clearly the effects of R&D investments generated by a tax incentive in a given year from other R&D investments, within the scheme or not, made by the firm. To account for this, the concept of R&D capital may be useful. The idea is to define R&D capital as the sum of past R&D investments net of depreciation, and include this variable in the econometric analysis. Then all the firms R&D investments are included in the analysis, with a greater emphasis on recent investments. A number of studies have found that the distribution of returns to R&D investments is skewed. Behind the average returns reported in many analyses, there is often a majority of projects with zero or low gross returns, thus failing from a commercial point of view, and a small number of projects with very high returns. This possible skewness should be taken into account, using methods that allow for heterogeneity in returns. One important rationale for government support to R&D is the presumption that R&D investments generate positive external effects, i.e. effects that accrue to someone else other than those undertaking the investment. The standard econometric approach is to extend the models of result additionality with variables reflecting R&D investments outside of the firm, and to estimate to what extent such variables have an effect on different measures of firm performance. Identifying and quantifying external effects of R&D in general is a daunting task. To isolate external effects of R&D investments triggered by a specific tax incentive is even more challenging. Therefore, if analysis of external effects is a part of an evaluation, one should not expect any definitive findings.
18
problematic to look at firms who apply for support through the scheme, using those firms that do not apply as a comparison group. Firms that get a good research idea are more likely to apply for support through the scheme, but they would also be more likely to carry out the project in the absence of the support. Hence, firms "within" the scheme and outside the scheme are likely to differ with respect "research ideas", a highly unobservable and time-varying variable. Ideally, the question of whether a specific measure works or not should be answered by carrying out a controlled experiment, randomly dividing the population of firms into two groups, giving one group access to the scheme. This would provide the exogenous variation needed, and give a basis for comparing a treatment group with a control group using the above framework. This ideal situation is almost never feasible, cf. Jaffe (2002). Schemes and measures are often general in nature. This creates great challenges for evaluators. The more general the scheme that is, the more equal similar firms are treated in the scheme, the more complicated is the evaluation. The reason is that a higher degree of generality or equal treatment brings evaluations further away from the ideal setting. When all comparable firms either have access to the scheme or not, it is impossible to construct a control group providing information about the counterfactual situation. The challenge in a non-experimental setting, without a formal control group, is to deduce from historical data, what the situation would have been if the scheme had not been launched. In the absence of a controlled experiment, one needs to look for so-called quasi-experiments built into the scheme. A quasi-experiment might involve, for example, using variations in the scheme that may be regarded as random at least at the margin. In our setting, randomness implies that the variations are not systematically related to (unobserved) variables that affect firms R&D decisions. A potential quasi-experiment could use variations in the generosity of the scheme with respect to firm characteristics that are relatively fixed in the short term, e.g. number of employees. Assuming that firms around the border of the size restriction are comparable, and that it is in a sense random whether they were eligible for support through the scheme or not, this discontinuity creates a quasi-experiment, and one may study the effect by comparing firms just above and just below the threshold. The absence or rareness of random variation thus makes it very difficult to identify the true, causal effects of tax incentives for R&D. The problems due to non-randomness are most visible in the econometric evaluation approaches, but are also present in other approaches.
19
to be estimated can be correctly specified. Data collection through existing sources is also relatively cheap. In situations where data on many relevant variables are lacking, and the preferred model cannot be estimated, an interview-based approach may partly substitute for econometric analyses. As described above, in an ideal situation the interviewee implicitly controls also for factors that cannot be included in an econometric model when they give their answers on e.g. input additionality. However, there is no guarantee that the implicit model is the same across interviewees. Data collection through interviews is relatively costly. While econometric techniques are well suited to capture effects that may be quantified in a sensible way, they are not equally suited to identify behavioural additionality, i.e. changes in the way firms understand R&D and how R&D decisions are made. Data on these issues are seldom readily available from independent sources, and must be collected separately. Coding information on such issues in a way that they can be included in econometric model, often removes important parts of its content. Here, case studies and interviews are more relevant methods.
20
6 Data issues
Good data are crucial for evaluations. The lack of appropriate data severely limits the potential of an evaluation. Before starting the design of the evaluation project and even before implementing the scheme, governments or operating agencies should go carefully through the data requirements for the evaluation and compare them to the data actually available. Some of the key questions to ask are: What is the ideal situation? What are the possible data sources? What are the shortcomings of the actual data situation? How can they be remedied - at what cost?
Below, these questions are discussed in turn. Our general principle in the discussion of data issues is that one should, as far as possible, use already existing/available data to avoid extra red tape for companies and to be able to have long time-series of data at a modest cost. In some countries, there are few relevant pre-existing data sources, and the time-series dimension may be short. The advice to such countries is to rely more on data gathered specifically for evaluation purposes, and to start establishing a data infrastructure for future evaluations.
One set of variables relates to the effects of the scheme, both first and higher order effects. The relevant variables will of course depend on what the evaluation specifically focuses on (see below), but key variables are likely to include: R&D activity: Own R&D, purchased R&D, R&D personnel, financial sources, partners etc. Result variables: Revenue, value added, profits, gross investments, innovations, patents, change in decision processes for R&D etc.
A second set of variables relates to other more basic characteristics of the firm, such as size (number of employees), education level of the staff, localization etc. Note that the classification of variables into one of the two sets is not always obvious.
Firms applying for support through project-based measures, have to submit a substantial amount of information about themselves and their research project. Through the processing of the application at the government agency, a lot of information relevant for the evaluation is produced. Firms may also have to send in their final reports with information of how the project actually was carried out, and with a self-assessment of the results. In addition, the application and reporting procedures represent an opportunity to collect data that are necessary for the evaluation, but not available from other sources. Thus, the administrative routines surrounding the scheme should be designed also with evaluation in mind. This source may provide valuable data at a modest cost. However, one disadvantage is that the data only cover those who apply for support through the scheme. In other countries tax incentive works on a self-assessment basis and so there is often no information on the R&D other than the amount of qualifying spending. This might not be a major problem for evaluations which have a primary focus on the increase in R&D spending and not on the type of R&D supported. To study the effects of a tax incentive scheme for R&D, one needs a fairly long time horizon because the total economic effects of R&D, or even the effect on R&D itself, cannot be expected to be visible in the short run. The data should cover a sufficiently
22
long period both prior to and after the launching of the scheme. What should be regarded as sufficiently long can in principle not be established ex ante. In practice, one has to make a compromise based on the availability of historical data, the cost of gathering new data, and the time horizon for the evaluation project.
6.2.2 Regular censuses and surveys carried out for other purposes
As for register data, this type of data is often used for the production of statistics, but is suitable for research. Relevant data sources of this kind include manufacturing censuses, R&D surveys and innovation surveys. The advantages and disadvantages are mainly the same as above, with the exception that the surveys do not cover the whole population. The representativeness of the sample in surveys may create a challenge. Often, surveys limit the sample to firms of a certain size, e.g. 10 employees or more. If the scheme is targeted towards or turns out to be more popular among small firms, then this data source is not ideal in the sense that is does not cover a main user group of the scheme of the scheme.
23
24