1 / 48

Measuring Outcomes

Measuring Outcomes. Comparing a measurement to a ‘true’ value or ‘Gold standard’. Scott Harris October 2009. Setting up your “Zapper”. Press “GO” “4” “1” “GO” on your handset to match the unique frequency for this session.

igor-valdez
Download Presentation

Measuring Outcomes

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Measuring Outcomes Comparing a measurement to a ‘true’ value or ‘Gold standard’ Scott HarrisOctober 2009

  2. Setting up your “Zapper” • Press “GO” “4” “1” “GO” on your handset to match the unique frequency for this session. • You have set it up correctly if an orange-green light is on (for a short period of time), otherwise a red light will show then you will have to try again. • You will be using the handset to answer some questions during this session. • All handsets need to be collected at the end of the session or when you leave the lecture room.

  3. Learning outcomes By the end of this session you should be aware of techniques for assessing the accuracy/precision of measuring the following ‘true’ outcome types with the following measures: • Continuous vs. continuous • Binary vs. binary • Binary vs. continuous

  4. Contents • Introduction • Continuous vs. continuous • The most common mistake • Features of interest (Precision and accuracy) • Setting up a Bland-Altman plot (with examples) • Binary vs. binary • Sensitivity / Specificty • Positive predictive value / Negative predictive value • Binary vs. continuous • Diagnostic tests • Receiver operating characteristic (ROC) curves

  5. Statistics

  6. Statistical analysis • The full set of statistical analysis techniques are like a cookbook and depending on the ingredients available (the data) only certain analyses will be appropriate. How you record your measurements directly affects the analysis options open to you and this should be considered before data collection. • The statistics section of the RSB course will go through the different tests that are used in different situations with different ‘types’ of data. These sessions will look at associations, relationships and group differences.

  7. Statistical Support Statistical support for SoM PhD students is available from: • Dr Ruth Pickering (rmp@southampton.ac.uk ) • Brian Yuen (hmy@southampton.ac.uk ) • Scott Harris (sharris@southampton.ac.uk ) Up to 3 (1hr) consultations per student.

  8. Types of data • Quantitative – a measured quantity. • Continuous – Measurements from a continuous scale: Height, weight, age. • Discrete – Count data: Children in a family, number of days in hospital. • Qualitative – Assessing a quality. • Ordinal – An order to the data: Likert scale (much worse, worse, the same, better, much better), age group (18-25, 26-30…). • Categorical / Nominal – Simple categories: Blood group (O, A, B, AB). A special case is binary data (two levels): Status (Alive, dead), Infection (yes, no).

  9. Measurement comparison studies • Also known as method comparison studies or measurement error studies. • Can be either a measurement measured a number of times or a number of alternative measures for the same outcome. • Techniques discussed here will focus on only 2 measurements at a time. More advanced techniques do exist looking at repeatability when more than 2 measurements are taken (see a statistician for these).

  10. Measures of a ‘true’ or Gold standard value Aim: To determine the ability of a new test to replicate the score of a ‘true’ or gold standard value accurately and precisely or to distinguish those who have the feature of interest from those who do not. Procedure: • Recruit sample of patients or take a sample of measures in your population of interest. Need to include patients both with and without the outcome of interest if we are dealing with a binary outcome. • All recipients undergo the new test (index test). • Test result is compared against knowledge of true status (reference test or ‘gold standard’ test).

  11. Reference test Reference test = ‘Gold standard’ Often imperfect: • Often invasive or unpleasant (e.g. liver or other biopsy). • May require long term follow-up (may take too long to arrive). • May be resource intensive (time, money etc.) • As close to the truth as possible. (assumed to be the actual truth).

  12. Continuous vs. continuous Accuracy and precision. Bland-Altman plot.

  13. Most common mistake Comparison of European scale peak flow meters with digital spirometer. The most common mistake is that researchers will draw a scatter plot of the two measures against each other and either calculate a correlation or a linear regression. If these results are significant they then conclude agreement between the measures. This would be INCORRECT. The solid red line on the plot shows an approximate line of ‘best fit’. The true line of equivalence is shown with the dotted red line. M. Goyal et al : Comparison of Wright scale and European scale peak flow meters with digital spirometer . The Internet Journal of Pulmonary Medicine. 2008 Volume 9 Number 2

  14. Accuracy Precision • These images illustrate two features of random and systematic errors: • A Random error (measurement error) affects the precision, whereas a systematic error (bias) affects accuracy. • A test score can be highly precise (reliable), without being accurate. • An ideal test score is both accurate and precise.

  15. Producing a Bland-Altman plot • Calculate the mean of your two measures for each observation. • Calculate the difference between your two measures for each observation. • Calculate the mean difference across all observations. • Calculate the Standard deviation of the difference and multiply it by 1.96. • Draw a scatter plot as shown on the next slide:

  16. The mean value is plotted on the x-axis, with the difference on the y-axis. Bland-Altman plot Add in reference lines at the mean difference and the mean difference +/- 1.96xSD. These are then your limits of agreement. If this interval is narrow enough you can conclude agreement between the 2 measures.

  17. The Bland-Altman plot for CUM15, the index that was used to express cough sensitivity in hypertonic saline challenge. Koskela et al.Cough 2008 4:8

  18. Baseline After weight-loss Difference “Bias between the techniques was not observed, as indicated by a non-significant p value (p = 0.648, p = 0.408 and p = 0.665, respectively).” Minderico et al.Nutrition & Metabolism 2006 3:32  

  19. Binary vs. binary Sensitivity, specificity, positive and negative predictive values.

  20. Classification of results

  21. Theory: Sensitivity and specificity Sensitivity: Of those with the disease, the proportion with a positive test result. Specificity: Of those without the disease, the proportion with a negative test result. Sensitivity and specificity are inversely related to one another - If the level used to distinguish between a positive and a negative test is shifted, sensitivity and specificity will move in opposite directions.

  22. Theory: Sensitivity Sensitivity = a / (a + c)

  23. Theory: Specificity Specificity = d / (b + d)

  24. Sensitivity & Specificity: Example Sensitivity = a / (a + c) = 62/69 = 89.9% Specificity = d / (b + d) = 34/39 = 87.2%

  25. What does this mean for an individual? • The problem with sensitivity and specificity is that they are not useful for looking at results for individual patients. • If a person has a positive test for some outcome then what we want to say is how likely is it for that person to actually have the outcome. • This is where an alternative, positive and negative predictive values come in: Positive predictive value (PPV): Of those with a positive test result, the proportion who truly have the disease. Negative predictive value (NPV): Of those with a negative test result, the proportion who are truly without the disease.

  26. Theory: Positive predictive value (PPV) PPV = a / (a + b)

  27. Theory: Negative predictive value (NPV) NPV = d / (c + d)

  28. PPV & NPV: Example PPV = a / (a + b) = 62/67 = 92.5% NPV = d / (c + d) = 34/41 = 82.9%

  29. Binary vs. continuous Receiver operating characteristic (ROC) curve, sensitivity, specificity, positive and negative predictive values.

  30. Test Score Perfect diagnostic test performance Cut-off value Normal Diseased Predict as normal Predict as Diseased

  31. Perfect diagnostic test performance

  32. Imperfect separation? – Real world! Test value Imperfect test. Is it good enough? Normal Diseased

  33. Imperfect separation? Test value True positives True negatives Normal Diseased

  34. Imperfect separation? Test value False positives False negatives Normal Diseased

  35. Classification of results

  36. Cut point: Maximising Sensitivity Test value Normal Diseased

  37. Cut point: Maximising Specificity Test value Normal Diseased

  38. More ‘Normal’ than ‘Diseased’ Test value Prevalence = All with disease Population False positives False negatives Normal Diseased

  39. Theory: PPV & NPV Whereas sensitivity and specificity were looking at the group truly with the outcome (sensitivity) or the group truly without the outcome (specificity) separately, PPV and NPV look at the ratio of parts of both of these groups. This means that PPV and NPV are linked with prevalence, whereas sensitivity and specificity are not. Low prevalence - The number of negative test results will be much larger than false negatives: NPV very high but not very informative. High prevalence - The number of positive test results will be much larger than false positives: PPV very high but not very informative.

  40. Receiver Operating Characteristic (ROC) curves Quite often we have a continuous value for our new test and we could choose various ‘cuts’ to create a binary prediction of status. An ROC curve demonstrates the effect of varying the cut point on sensitivity and specificity: • Plots a curve of Sensitivity vs. (1 - Specificity) for all cut points that would alter at least one classification. • The top left corner of the plot signifies perfect performance. • Often includes a diagonal line, indicating an uninformative test. • The larger the area under the curve the better the test across the range of cut points. • The nearest one point gets to perfect performance the better the test performance using the single best cut point.

  41. Receiver Operating Characteristic (ROC) curves Perfect performance Better performance

  42. ROC curves: SPSS - Customised output

  43. Practical Questions

  44. Practical Questions A new test for a disease was evaluated on 50 affected individuals and 50 who were known to be unaffected. The test correctly identified 40 of the affected subjects, but incorrectly classified 5 of the unaffected ones. • What is the sensitivity of the test? • What is the specificity of the test? • What is the positive predictive value of the test? • What is the negative predictive value of the test? • If the real disease prevalence is 2%, how likely is it that someone with a positive test has the disease?

  45. Practical Questions (I to IV) Sensitivity = Specificity = PPV= NPV=

  46. Practical Questions (V) Sensitivity = Specificity = PPV= NPV=

  47. Summary You should now be aware of techniques for assessing the accuracy/precision of measuring the following ‘true’ outcome types with the following measures: • Continuous vs. continuous • Bland-Altman plot • Binary vs. binary • Sensitivity/specificity/PPV/NPV • Binary vs. continuous • ROC curves and sensitivity/specificity/PPV/NPV

  48. References • Bland JM, Altman DG. (1986). Statistical methods for assessing agreement between two methods of clinical measurement. Lancet, i, 307-310. • http://statpages.org/ctab2x2.html • Swets JA, Pickett RM. Evaluation of diagnostic systems. New York: Academic Press,1982. • Langlotz CP. Fundamental measures of diagnostic examination performance: usefulness for clinical decision making and research. Radiology 2003; 228:3-9. • Hanley JA, McNeil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982; 143:29-36. • Hanley JA, McNeil BJ. A Method of Comparing the Areas under Receiver Operating Characteristic Curves Derived from the Same Cases. Radiology 1983; 148:839-843.

More Related