1 / 16

Archived File

Archived File. The file below has been archived for historical reference purposes only. The content and links are no longer maintained and may be outdated. See the OER Public Archive Home Page for more details about archived files. Priority Scores and Percentiles. James Onken, Ph.D.

Download Presentation

Archived File

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Archived File The file below has been archived for historical reference purposes only. The content and links are no longer maintained and may be outdated. See the OER Public Archive Home Page for more details about archived files.

  2. Priority Scores and Percentiles James Onken, Ph.D. Chief, Office of Program Analysis and Evaluation NIGMS

  3. Priority and Percentile Scores Priority Score = The average of individual reviewers’ scores on 1.0 - 5.0 scale (two significant digits), rounded to three significant digits, and multiplied by 100 Percentile Score = The rank of an application’s priority score, relative to all priority scores assigned in the three most recent review cycles Example: Average reviewer score = 1.53 Priority score = 153 Percentile score = 20.7 (MGA)

  4. Scoring Issues How reliable are they? Are they being expressed at an appropriate level of precision? Do the average priority score and percentile provide enough information? - To applicants - To NIH staff In one way or another, all three of these issues involve score variation and what this variation means.

  5. Sources of Variation in Scores • Individual Reviewer Scores • True differences among applications • True differences of opinion among reviewers • Error • Priority Scores • True differences among applications • Error • Percentile Scores • Variation in Priority Scores • Variation in score “base” used to calculate percentiles

  6. Reliability of Group Averages Reliability of individual reviewer score (rxx): True variance True variance + Error variance Expected reliability of group average (r*xx): r*xx = N rxx 1 + (N-1) rxx

  7. Expected Reliability of Priority ScoresAs Function of Group Size

  8. Misapprehension of the Advantages of Averaging “Averaging estimates is an effective way to improve accuracy when combining expert judgments, integrating group members’ judgments, or using advice to modify personal judgments... people often hold incorrect beliefs about averaging, falsely concluding that the average of two judges’ estimates would be no more accurate than the average judge. The experiments confirmed that this misconception was common across a range of tasks...” - Larrick and Soll, 2005 “Even subjects who were statistically sophisticated showed limited appreciation of the aggregation principle...” - Kunda and Nisbett, 1986

  9. Describing Variation in Reviewers’ Scores • Standard Deviation • Range

  10. Variance Indicators

  11. Variance of Proportions (Percentiles) Sampling errors in proportions are a function of: • proportion (largest around 0.5) • denominator (smaller with larger N)

  12. Standard Errors of Proportions

  13. Round-to-Round Variation in Percentile ScoresAssociated with Priority Scores within the Same Study Section

  14. Priority Scores Among Applications of Equivalent Relative Rank

  15. Priority Score VariationAmong Applications with Equivalent Relative Rank Coefficient of Variation =Standard Deviation Average Priority Score ≈0.10

More Related