1 / 20

Readings

Readings. Text: Riddick & Russell Ch1 stakeholders – p10 Ch 2 an evaluation system Proposal p25-36 Ch 4 – Lit Review Coursepack GAO report Ch 1 & 2, pp 177-185 MSU Sports pp 256-260. Evaluation is a process. Evaluability assessment Evaluation research proposal

magnar
Download Presentation

Readings

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Readings • Text: Riddick & Russell • Ch1 stakeholders – p10 • Ch 2 an evaluation system • Proposal p25-36 • Ch 4 – Lit Review • Coursepack • GAO report Ch 1 & 2, pp 177-185 • MSU Sports pp 256-260

  2. Evaluation is a process • Evaluability assessment • Evaluation research proposal • Review of the proposal • Conduct evaluation • Report results • Implementation of findings

  3. Evaluability Assessment • Is a program ready to be evaluated? • Description of program • Goals and objectives • Organization ready? Identify decision-makers • Political and social factors • Model of how program works • Are resources to conduct an evaluation available? • What part of program will be evaluated?

  4. Steps for Evaluability Study • Program description – review documents • Identify targets, objectives, inputs, outputs • Interview key personnel • Scout program • Develop program model • Get agreement to proceed/cooperation

  5. A Program Model Labor, Time Capital, $$$ Land/Facilities Use measures “Benefits to users” $$$ Process Outputs INPUTS Effects Impacts

  6. Evaluation criteria • Effort – qnty & qlty of inputs • Performance- qnty & qlty of outputs • Adequacy – meet needs? • Efficiency – costs/benefits • Process – how & why pgm works? • Equity – who benefits, who pays?

  7. Purposes of Proposal • Communicate with Client • Demonstrate your grasp of problem • Plan the study in advance, so others can evaluate the study approach • will it work? • have you overlooked something? • will results be useful to client? • Can we afford it?

  8. Proposal Format 1. Problem Statement - define program to be evaluated/problem to be studied, users & uses of results. Justify importance of the problem/study. 2. Objectives : Concise listing . In evaluation studies, the objectives usually focus on the key elements of program to be evaluated & the evaluation criteria. These are the study objectives NOT the program objectives. 3. Background/Literature Review - place for more extensive history/structure of program. Focus on aspects most relevant to proposed evaluation. Discuss previous studies or the relevant methods. 4. Methods - details on procedures for achieving objectives - data gathering and analysis, population, sampling, measures, etc. Who will do what to whom, when, where, how and why? 5. Attachments - budget, timeline, measurement instruments, etc. NOTE: Most “programs” must be narrowed to specific components to be evaluated. Think of a “Program of studies” rather than a single evaluation study. The proposal should define this specific study & how it fits into a broader program of studies.

  9. Sample Evaluation Objectives 1. Estimate benefits and costs of program 2. Estimate economic impacts of program on local community (social, environmental, fiscal). 3. Determine effects of program on target population. 4. Describe users and non-users of program 5. Assess community recreation needs, preferences 6. Determine market/financial feasibility of program 7. Evaluate adequacy or performance of program

  10. Typical Research Objectives Describe a sample or population Identify/test relationships between variables in a population: statistical cause-effect Quantify the relationship Average income of MI Snowmobilers in 1998 is $45K SB with higher incomes spend more money After safety program, SB’s have fewer accidents SB spend per day = $25 + .4 * Income

  11. Variable Terminology Variables : any characteristic that varies across individuals in a population (i.e. takes on different values for different individuals). Dependent variable is the one you are trying to predict or explain, usually the focus of your study Independent variables are the ones that help explain the dependent variable. In Program evaluation, the outcomes are generally the dependent variables and characteristics of program or target populations are independent. In a cause effect relationship, cause is independent & effect dependent.

  12. MSU Sports Programs What is program(s)? Inputs, outputs, process Stakeholders Which piece to evaluate Evaluation criteria Methods to use

  13. Methods Choices • Overall Approach/Design • Qualitative or Quantitative • Primary or secondary data • Survey, experiment, case study, etc. • Who to study - population, sample • individuals, market segments, populations • What to study - concepts, measures • behavior, knowledge, attitudes • Cost vs Benefit of Study

  14. Qualitative vs Quantitative Approaches Qualitative Focus Group In-Depth Interview Case Study Participant observation Secondary data analysis Quantitative Surveys Experiments Structured observation Secondary data analysis

  15. Qualitative vs Quantitative Quantitative Gen’l Laws Test Hypotheses Predict behavior Outsider-Objective Structured formal measures probability samples statistical analysis Qualitative Unique/Individual case Understanding Meanings/Intentions Insider-Subjective Unstructured open ended measures judgement samples interpretation of data Purpose Perspective Procedures

  16. Primary or Secondary Data • Secondary data are data that were collected for some purpose other than your study,e.g. government records, internal documents, previous surveys • Choice between Primary /Secondary Data • Costs (time, money, personnel) • Relevance, accuracy, adequacy of data

  17. Research Designs/Data Collection Approaches

  18. Survey vs Experiment Survey - measure things as they are, snapshot of population at one point in time, generally refers to questionnaires (telephone, self-administered, personal interview) Experiment - manipulate at least one variable (treatment) to evaluate response, to study cause-effect relationships (field and lab experiments)

  19. General Guidelines on when to use different approaches 1. Describing a population - surveys 2. Describing users/visitors - on-site survey 3. Describing non-users, potential users or general population - household survey 4. Describing observable characteristics of visitors - on-site observation 5. Measuring impacts, cause-effect relationships - experiments

  20. Guidelines (cont) 6. Anytime suitable secondary data exists - secondary data 7. Short, simple household studies - phone 8. Captive audience or very interested population - self-administered survey 9. Testing new ideas - experimentation or focus groups 10. In-depth study - in-depth personal interviews, focus groups, case studies

More Related