NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 7 results Save | Export
Reardon, Sean F.; Kalogrides, Demetra; Ho, Andrew D. – Journal of Educational and Behavioral Statistics, 2021
Linking score scales across different tests is considered speculative and fraught, even at the aggregate level. We introduce and illustrate validation methods for aggregate linkages, using the challenge of linking U.S. school district average test scores across states as a motivating example. We show that aggregate linkages can be validated both…
Descriptors: Equated Scores, Validity, Methods, School Districts
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Jinming – Journal of Educational and Behavioral Statistics, 2012
The impact of uncertainty about item parameters on test information functions is investigated. The information function of a test is one of the most important tools in item response theory (IRT). Inaccuracy in the estimation of test information can have substantial consequences on data analyses based on IRT. In this article, the major part (called…
Descriptors: Item Response Theory, Tests, Accuracy, Data Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Kachchaf, Rachel; Solano-Flores, Guillermo – Applied Measurement in Education, 2012
We examined how rater language background affects the scoring of short-answer, open-ended test items in the assessment of English language learners (ELLs). Four native English and four native Spanish-speaking certified bilingual teachers scored 107 responses of fourth- and fifth-grade Spanish-speaking ELLs to mathematics items administered in…
Descriptors: Error of Measurement, English Language Learners, Scoring, Bilingual Teachers
Xu, Xueli; von Davier, Matthias – Educational Testing Service, 2010
One of the major objectives of large-scale educational surveys is reporting trends in academic achievement. For this purpose, a substantial number of items are carried from one assessment cycle to the next. The linking process that places academic abilities measured in different assessments on a common scale is usually based on a concurrent…
Descriptors: Case Studies, Trend Analysis, Computation, Educational Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Braun, Henry; Qian, Jiahe – ETS Research Report Series, 2008
This report describes the derivation and evaluation of a method for comparing the performance standards for public school students set by different states. It is based on an approach proposed by McLaughlin and associates, which constituted an innovative attempt to resolve the confusion and concern that occurs when very different proportions of…
Descriptors: State Standards, Comparative Analysis, Public Schools, National Competency Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Oranje, Andreas; Freund, David; Lin, Mei-jang; Tang, Yuxin – ETS Research Report Series, 2007
In this paper, a data perturbation method for minimizing the possibility of disclosure of participants' identities on a survey is described in the context of the National Assessment of Educational Progress (NAEP). The method distinguishes itself from most approaches because of the presence of cognitive tasks. Hence, a data edit should have minimal…
Descriptors: Student Surveys, Risk, National Competency Tests, Data Analysis
Davey, Tim; And Others – 1996
Some standard-setting methods require judges to estimate the probability that an examinee who just meets an achievement standard will answer each of a set of items correctly. These probability estimates are then used to infer the values on some latent scale that, in theory, determines an examinee's responses. The paper focuses on the procedures…
Descriptors: Academic Achievement, Achievement Tests, Elementary Secondary Education, Error of Measurement