Transcript Document
Indicators to characterize public funding systems Benedetto Lepori, 28th November 2008 The issue Is it possible to produce indicators to characterize/compare public funding systems including The role of different funding agencies The allocation methods The streams of money The beneficiaries Beyond R&D statistics which is essentially focused on performers With limited ability to map funders – performers flows With a too simple description of funders (by “sector”) Not taking into account the existence of intermediaries Obvious relevance for public policies of these indicators See for example the OECD work on steering and funding of public research where no data on allocation could be collected Two PRIME project on the field Exploratory project on public project funding Analysis of CEEC funding systems and of their change An overall view of public funding Four relevant system layers State Funding agencies Performers’ organisations Research groups/individuals) Two main allocation streams core funding project funding Key issues Interactions accross layers and funding modes Increasing role of performers in driving funding Policy layer National state Funding agencies (national) Agencies layer Core funding Organisations layer Research groups layer Project funding Higher Education Institutions and PROs Internal allocation International agencies (EU, ESA, etc.) Research units Project funding analysis Try to compare countries concerning the role and composition on project funding Identify in each country the instruments which we can identify as project funding As well as the main funding agencies A basic definition + a lot of comparative work to treat in the same way the same instruments Collect the data from different sources Ministries reports, agencies reports, etc. Per year and main beneficiary group Data collection and cleaning procedures Produce aggregations and international comparisons As well as analyses of the evolution in last 30 years Criteria We consider project funding if Attributed mostly for research purposes Limited in time and scope (not recurrent) Attributed by an external agency to the research organization No reference to competition/modes of allocation Issues / problems Use for research activities (since we are not looking to performers) – difficult to assess for example for contracts Long-term competitive schemes (centres of excellence) Internal competitive schemes to research organizations (or vertically integrated organizations like academy of sciences) Problem cases European Space Agency contracts Mostly to industry for development of rockets, satellites, etc. Some funding for scientific programs It is nearly impossible to ascertain how much should be considered as R&D We adopted a broad delimitation Including probably more that what would be considered as R&D funding French CNRS A very large organization with a double role A partially competitive allocation mechanism Funding its own laboratories (10%) Funding joint laboratories with universities (thus partially outside the CNRS perimeter) Labelling by CNRS gives access to its resources and there is a rather high turnover Mostly in form of personnel, but no mandatory allocation It is something intermediary between general and project funds Accounting for nearly 10% of public research funding in France, thus impossible to put in a footnote Our picture of the French system largely depends on the choice we do concerning these funds Considering it as project gives a view of France much nearer to other Western European countries See Theves et al. paper Discussion Overall the distinction works and is usable but One needs much care in the comparisons especially for time series We should try to develop a finer typology of funding instruments in the future Intensive discussion was needed to solve dubious cases Ensuring some comparability In-depth knowledge of national systems was essential for this exercise …this is not the end of the story… Categories Comparative analysis clearly needs common categories Beyond individual list of instruments/agencies at national level We devise three classifications By type of agency (international, ministry, intermediaries) By instrument type By beneficiaries Instrument types Intuitively it is clear that there are different instruments concerning their orientation / type of research they fund But designing a clear classification is very difficult since these concepts are largely multidimensional We end up with a simple distinction between “academic” / thematic / innovation-oriented instruments However, this is just a rough approximation of policy intentions, but not necessarily of the research done Different features need to be combined to better understand the significance of instruments for the research system Granularity is a problem since the level of aggregation is rather high Data sources and data collection issues There is no unique source of these data, but they had to be complied from: State accounts (France) Research ministry reports Agency reports Some existing databases Surveys (for example the Swiss R&D survey) Direct inquiry to ministries Most data are available but with limitations/problems in many cases In most cases it was possible to come back to the ‘70 Data problems Funding to companies through loans Estimate of cash value Project decisions instead of allocations One needs some kind of averaging European Space Agency Use national contribution as proxy (national return rule) Contracts from ministries Coverage is problematic outside the formal programs …even if the methdology is simple there is a lot of dirty work to get usable data… Some selected results PF as the second stream of research funding in the six considered countries ¼ to 1/3 of funding volume Strong increase in the role in the last years A composite model of funding instruments Some academic, some thematic, some innovation Try to accomodate the different goals instead of chosing a single rationale Differences between countries remain quite large In the organisation of funding agencies and the type of funding Strong dependence on history of organisational structures One needs to consider carefully this context when designing European funding policies Type of PF instruments 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Sw itzerland Academic EU FP Austria Italy Thematic national Norw ay Netherlands France Space (ESA + national programs) Innovation France w ith CNRS undivided PF as % of BIP 0.6 1970 1980 1990 2002 0.5 0.4 0.3 0.2 0.1 0 Switzerland Austria Italy Norway Netherlands France From academic to thematic and back… 90.0 France 1990 80.0 France 1982 70.0 France 2002 Netherlands 1980 60.0 % Thematic instruments Italy 2002 Netherlands 2002 Switzerland 1990 50.0 Italy 1990 Netherlands 1990 Switzerland 1980 Switzerland 2002 40.0 Netherlands 1975 Austria 2002 Switzerland 1970 30.0 Italy 1980 Austria 1990 Austria 1980 20.0 10.0 Italy 1971 Austria 1970 0.0 0.0 10.0 20.0 30.0 40.0 50.0 % Academic (bottom-up) instruments 60.0 70.0 From steering to markets Today’s highly differentiated project funding system resembles increasing to a market Where loosely coordinated agencies buy research services from performers With strong strategic behavior of applicants Depending on their strenghts, needs, etc. Building stable market structures Needed to support long-term research, but also predictable allocation of funding Crucially based on shared representations on others actors behaviour Of funding agencies Of potential competitors Strong segmentation by domain/type of research/topic Few key players with high success rates, many marginal players Some evidence that core markets are relatively small-scale CEEC project To analyse the changes in the organization of public funding in CEEC Three countries as a test (Czech Republic, Estonia, Poland) Look also for availability of data (including time series) Approach Identify the main funding streams and to characterize them Draw structural diagrams Understand the role of funding agencies and allocation criteria See Karel’s presentation for more details on specific results Poland MSHE Institutional funding 24,19% 70,66% 30,81% 17,37% Higher education 23,81% 22,48% AS Government units 7% HEI Institutes 13,20% 0,55% 0,49% Institutes 3,35% 0,22% 0,17% Institutes 5,2% 0,26% 0,68% 2,45% project funding Pri vat e 1,46% EU FP 1,57% EU SF MSHE fund FPS 1,52% 0,60% 0.23% Structural models A step towards identifying the main structural features of national funding models Taking into account the interactions between modes and layers (institutional complementarities) Different models related to different national contexts Which can be stable in an evolutionary perspective Ability of the State of steering the system Levels of delegation Performance vs. innovation Project-funding based Centralised Mixed models Characterize the models in terms of some key features Three basic models seems to emerge: The project-funding based model Most of research funding coming from highly differentiated set of non coordnated agencies High share of project funding and low of core funding State « buys » research services from groups With limited attempt of central coordination Mission agencies funding also basic research (DoD) Reliance on competition rather than on coordination to get the required results stratified university system with concentration in a few players International excellence as the main criterion Limitations of this model when the system is too small Cumulative effects suppress diversity of the system (the Estonian case) The centralised model A single large research organisation funded by the state (outside the university sector) The old CNRS model before joint laboratories The old system of Academy of Sciences in communist countries The today’s Polish system The PRO has more functions than just funding Decision on priority areas Creation and restructuring of laboratories Mostly coopting scientists in the PRO board A strongly centralised system Define priorities and concentrate effort Central planning, but with all its rigidities in face of new scientific development Less adapted to new science dynamics more based on complementarities than on heavvy investments (new « search regimes »; Bonaccorsi) The mixed model Some balance between core and project funding Say 60% to 40% Large university sector alongside a number of non-university research institutes A differentiated performers sector Trying to achieve some balance between Targeting the best groups and spreading out resources (also because of the link with education and regional development) Differentiating the performers sectors to answer to different missions vs. creating competition between performers Typical of most Continental European countries including Switzerland Best performing in small rich countries rather that on large ones Is the efficiency loss and the lack of central coordination acceptable? The Czech republic as a typical case Conclusions (1) It is basically feasible to characterize funding systems using existing data at national level Mostly detailed budgetary data Complemented with data from agencies and R&D statistics A very good knowledge of national systems is required To identify and classify funding streams To identify the really relevant features National experts are required for these tasks Conclusions (2) This works provides very interesting insights on the structure of funding systems Quantitative indicators are essential to characterize them and to distinguish variants The next avenue is to go towards data on micro-structures of funding systems References (see www.enid-europe.org) Lepori B., van den Besselaar P., Dinges M., van der Meulen B., Potì B., Reale E., Slipersaeter S., Theves J., (2007), Comparing the Evolution of National Research Policies: what Patterns of Change?, Science and Public Policy 34 (6), 372-388. Lepori B., Dinges M., Potì B., Reale E., Slipersaeter S., Theves J., van d. (2007). Indicators for Comparative Analysis of Public Project Funding. Research Evaluation, 16 (4), 243-255. Lepori B., Masso J., Jablecka J., Sima K., Ukrainski K. (2008). Research funding system in Central and Eastern European countries: a comparative analysis. Paper presented at the ENID-PRIME Indicators Conference, Olso, May 2008.