1 / 21

Evaluating Scientific Research Initiatives

Evaluating Scientific Research Initiatives. The Case of the International Agricultural Research Centers Leslie J. Cooksy, Ph.D. University of California – Davis ljcooksy@ucdavis.edu. Development of an impact assessment culture. Mid-1990s – Impact Assessment & Evaluation Group 2000 –

elwyn
Download Presentation

Evaluating Scientific Research Initiatives

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Evaluating Scientific Research Initiatives The Case of the International Agricultural Research Centers Leslie J. Cooksy, Ph.D. University of California – Davis ljcooksy@ucdavis.edu

  2. Development of an impact assessment culture • Mid-1990s – Impact Assessment & Evaluation Group • 2000 – Systemwide workshop on impact assessment • 2004 – Impact assessment culture and impact assessment studies monitored in new performance monitoring system

  3. Shifting the impact assessment culture – the cynical view Before • Impact assessment = Ex ante return on investment After • Impact assessment = Ex post return on investment

  4. Concerns about IA in the IARCs • Focus on demonstrating impact instead of questioning effectiveness • Over-selection of successful cases • Lack of attention to negative consequences of IARC research • Tendency to attribute all benefits to center activities (limited focus on context and interactions) • Disciplinary barriers to non-economic approaches to evaluation

  5. General challenges in assessing impact of research initiatives • Dependence on intermediaries to achieve long-term outcomes • The years that can pass between a research product and its impact • Risk inherent in research

  6. Challenges to impact assessment in the IARCs • Multiple levels of evaluation • Project • Program • Center • System • Lack of coordination across levels • Multiple external demands for evaluation • Limited evaluation expertise in research centers

  7. How do the IARCs evaluate impact? • Research efficiency (cost/benefit) • Evidence of use/adoption of technologies • Case studies • External review panels • Performance monitoring • New methods (e.g., narratives)

  8. External reviews of the IARCs • Panels of five or fewer internationally-respected scientists • Asked to assess: • Consistency of Center mission, strategies and priorities with CGIAR’s mission and priorities • Relevance and quality of science • Governance and management • Accomplishments and impact of Center research

  9. External review issues • Length of review process ( 2 years) • Length of time between reviews ( 5-7 years) • Limited pool of truly external researchers • Lack of coordination with Center-commissioned studies • Lack of explicit criteria  inconsistent application of criteria by different panels

  10. Goals for performance monitoring by the IARC system (CGIAR) • Links planning and evaluation • Provides annual data • Focuses on outputs and outcomes • Feeds into external reviews

  11. PM system indicators • Potential to Perform indicators: • Quality and Relevance of Research • Institutional Health • Financial Health • Results indicators: • Outputs • Outcomes • Impacts

  12. PM indicator: Institutionalization of impact assessment • Nature of the portfolio of IA studies • Innovation and advancement of IA methods and processes • Communication/dissemination and capacity enhancement • Impact culture (internal feedback and learning)

  13. PM indicator: Impact studies • 2 studies are submitted and ranked for: • Clear presentation • Reasonable and transparent assumptions • Reliable and representative data • Realistic counterfactual • Sound attribution of benefits to research • Distance down the impact pathway • External input

  14. Performance monitoring issues • Across centers: • Inconsistent definitions in the plans that are the basis for the assessment of outputs and outcomes • Differing opportunities and standards across the multiple disciplines represented in different IARCs • In general • Addressing outcome & impact of research activities • Distrust of process

  15. Addressing the challenges • Role of program theory (the “impact pathway”) • Establish the program, not individual projects, as the unit of analysis • Explicate role of intermediaries in uptake, adoption, and use • Explicate link between use of technology and change in the condition of end users • Focus data collection on the causal links

  16. Addressing the challenges • Case studies -- “Modus operandi method” (Scriven) • Detailed analysis of the configuration of the chain of events • Use of “tracers” • Role of meta-analysis/evaluation synthesis • Include studies that show negative as well as positive outcomes • Plan portfolio of studies that can be synthesized

  17. Factors Pushing Against Change • Decreased funding for research centers • Continued lack of coordination of expectations • Insufficient involvement of research centers in planning change • Tradition

  18. Factors Pushing for Change • Donors’ demand for accountability • Coordination at the system level • Reinvigorated oversight/review body (Science Council) • Cultural shift toward acceptance of M&E • Institutional Learning & Change (ILAC) movement

  19. Conclusion • Complex systems need multiple approaches to impact assessment • Impact assessment demands from multiple sources need to be coordinated • Strengths of traditional approaches need to be recognized • “Disputatious community of scholars” needs to be nurtured so that negative results are seen as opportunities to learn

More Related