1. Rating Scales Derived from Student Samples: Effects of the Scale Maker and the Student Sample on Scale Content and Student Scores
- Author
-
Carolyn E. Turner and John A. Upshur
- Subjects
Linguistics and Language ,Scale (ratio) ,media_common.quotation_subject ,Applied psychology ,Sample (statistics) ,Language and Linguistics ,Education ,Test (assessment) ,Rating scale ,Quality (business) ,Set (psychology) ,Construct (philosophy) ,Psychology ,Social psychology ,Spoken language ,media_common - Abstract
Performance tests typically require raters to judge the quality of examinees' written or spoken language relative to a rating scale; therefore, scores may be affected by variables inherent in the specific scale development process. In this study we consider two variables in empirically derived rating scales that have not been investigated to date: scale developers and the sample of performances used by the scale developers. These variables may affect scale content and structure and (ultimately) final test scores. This study examined the development and use of scales using two samples of ESL student writing and three teams of rating scale developers to construct three empirically derived scales. A comparison of the scale content showed considerable variation even though all development teams used similar constructs of writing ability. Each team used its own scale to rate a different set of compositions. Comparison of the ratings showed that scale development team had a minor effect on ratings and that scale development sample had a major effect. We present implications of these findings on the nature of empirically derived rating scales, focusing particularly on how such scales are developed.
- Published
- 2002
- Full Text
- View/download PDF