270 likes | 515 Views
Administration. Administering Scoring Statistical computation. Administering. Preparing the testing environment - for paper-and-pencil tests - for listening tests - for speaking tests Communicating the instructions Detailed instructions for examiners and invigilators
E N D
Administration Administering Scoring Statistical computation
Administering • Preparing the testing environment - for paper-and-pencil tests - for listening tests - for speaking tests • Communicating the instructions Detailed instructions for examiners and invigilators • Maintaining a supportive testing environment The role of examiners and invigilators • Collecting the tests
Administration: Scoring Objective items Writing Total Score Total Score Adjusted score University A 65.00 30.01 20.00 13.80 10.63 Qinghua Uni. 65.00 45.41 20.00 12.88 14.89 Beijing Uni. 65.00 43.78 20.00 18.13 14.22 (Li, 1997: 114)
Administration: scoring • Scoring objective test items • Scoring subjective test items Measures for quality control • Marking scheme / Rubrics • Rater training • Place • Examinees’ names not exposed • Random allocation of test papers to raters • A third rater when necessary • Computer adjustment
Administration: Statistical computation • Analysis and evaluation of the test • Score distribution • skewness (positive vs. negative) • kurtosis (leptokurtic vs. platykurtic) • score interpretation • Score interpretation • Item analysis • Conclusion and feedback
Normal Distribution http://stat-www.berkeley.edu/~stark/Java/NormHiLite.htm
Sk<0 Sk>0
Leptokurtic Ku>0) Mesokurticku=0 Platykurtic Ku<0)
Reliability • Regression matrix (correlation) • Factor analysis MET1985至1987各年客观题部分的信度数据
Regression matrix (correlation) (Li, 1997: 129)
( Li, 1997: 130) • factor analysis
Administration: Statistical computation • Score interpretation TOEFL scores NMET scores
Administration: Statistical computation • item analysis (in pretest study and after the test) • subjective items (Rasch model) • objective items • facility value (pass rate): the ratio between those who pass the item and the total number of testees • Difficulty (Δ,delta) • Discriminability (rbis)
Statistical computation • Simple: • Rank students according to their total scores • Divide them into three groups (top and bottom have equal numbers of students) • Count how many students in the top group get an item right, and how many in the bottom group • Formula: • Example
D.I.= = = .047619 Statistical computation Facility value=44/62=71%
Statistical computation • Simple: • Rank students according to their total scores • Divide them into three groups (top and bottom have equal numbers of students) • Count how many students in the top group get an item right, and how many in the bottom group • Formula: • Example • Point biserial correlations (Rbis)