180 likes | 330 Views
Content Analysis: Reliability. Kimberly A. Neuendorf, Ph.D. Cleveland State University Fall 2011. Reliability. Generally—the extent to which a measuring procedure yields the same results on repeated trials (Carmines & Zeller, 1979) Types: Test-retest: Same people, different times.
E N D
Content Analysis:Reliability Kimberly A. Neuendorf, Ph.D. Cleveland State University Fall 2011
Reliability Generally—the extent to which a measuring procedure yields the same results on repeated trials (Carmines & Zeller, 1979) Types: Test-retest: Same people, different times. Intracoder reliability. . . Alternative-forms: Different people, same time, different measures. Internal consistency: Multiple measures, same construct. Inter-rater/Intercoder: Different people, same measures.
Index/Scale Construction Similar to survey or experimental work e.g., Bond analysis—Harm to female, sexual activity Need to check internal consistency reliability (e.g., Cronbach’s alpha)
Intercoder Reliability Defined: The level of agreement or correspondence on a measured variable among two or more coders What contributes to good reliability? careful unitizing, codebook construction, coder training (training, training!)
Reliability Subsamples Pilot and Final reliability subsamples Because of drift, fatigue, experience Selection of subsamples Random, representative subsample “Rich Range” subsample Useful for “rare event” measures Reliability/variance relationship
Intercoder Reliability Statistics - 1 Types Agreement Percent agreement Holsti’s Agreement beyond chance Scott’s pi Cohen’s kappa Fleiss’ multi-coder extension of kappa Krippendorff’s alpha(s) Covariation Spearman rho Pearson r Lin’s concordance correlation coefficient (rc)
Reliability Statistics – 2 See handouts on (a) Bivariate Correlation and (b) Pearson’s and Lin’s Compared
Reliability Statistics - 3 Core assumptions of coefficients “More scholarship is needed”—these coefficients have not been assessed!
Reliability Statistics - 4 My recommendations Do NOT use percent agreement ALONE Nominal/Ordinal: Kappa (Cohen’s, Fleiss’) Interval/Ratio: Lin’s concordance Calculate via PRAM Reliability analyses as diagnostics, e.g., Problematic variables, coders (“rogues”?), variable/coder interactions Confusion matrixes (categories that tend to be confused)
Reliability Statistics - 5 “Standards” for Minimums for Rel. Stats. Percent Agreement: 90%?? Kappa (Cohen’s, Fleiss’): .40 minimally, .60 OK, .80 good Pearson correlation; Lin’s concordance: .70 (~50% shared variance) --???
Reliability Statistics - 6 • The problem of the “extreme” or “skewed” distribution • Can have a % agreement of .95 and a Cohen’s kappa of -.10!!! • Why? • What to do?
PRAM: Program for Reliability Analysis with Multiple Coders Written by rocket scientists! Trial version available from Dr. N!