NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)4
Since 2006 (last 20 years)7
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 14 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Jones, Andrew T.; Kopp, Jason P.; Ong, Thai Q. – Educational Measurement: Issues and Practice, 2020
Studies investigating invariance have often been limited to measurement or prediction invariance. Selection invariance, wherein the use of test scores for classification results in equivalent classification accuracy between groups, has received comparatively little attention in the psychometric literature. Previous research suggests that some form…
Descriptors: Test Construction, Test Bias, Classification, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Grabovsky, Irina; Wainer, Howard – Journal of Educational and Behavioral Statistics, 2017
In this article, we extend the methodology of the Cut-Score Operating Function that we introduced previously and apply it to a testing scenario with multiple independent components and different testing policies. We derive analytically the overall classification error rate for a test battery under the policy when several retakes are allowed for…
Descriptors: Cutting Scores, Weighted Scores, Classification, Testing
Moore, Joann L.; Li, Tianli; Lu, Yang – ACT, Inc., 2020
The Every Student Succeeds Act requires that English Learners (ELs) are included in annual state testing (grades 3-8 and once in high school) and included in each state's accountability system disaggregated by subgroup to ensure that they receive the support they need to learn English, participate fully in their education experience, and graduate…
Descriptors: College Entrance Examinations, Scores, English Language Learners, Accountability
Powers, Sonya; Li, Dongmei; Suh, Hongwook; Harris, Deborah J. – ACT, Inc., 2016
ACT reporting categories and ACT Readiness Ranges are new features added to the ACT score reports starting in fall 2016. For each reporting category, the number correct score, the maximum points possible, the percent correct, and the ACT Readiness Range, along with an indicator of whether the reporting category score falls within the Readiness…
Descriptors: Scores, Classification, College Entrance Examinations, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Bradshaw, Jenny; Wheater, Rebecca – Research Papers in Education, 2013
This review examined a range of approaches internationally to the reporting of assessment results for individual students, with a particular focus on how results are represented, the level of detail reported and the steps taken to quantify, report and explain error and uncertainty in the results' reports or certificates given to students in a…
Descriptors: Test Reliability, Error of Measurement, High Stakes Tests, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Verschuren, Olaf; Bosma, Liesbeth; Takken, Tim – Developmental Medicine & Child Neurology, 2011
For children and adolescents with cerebral palsy (CP) classified as Gross Motor Function Classification System (GMFCS) level III there is no running-based field test available to assess their cardiorespiratory fitness. The current study investigated whether a shuttle run test can be reliably (test-retest) performed in a group of children with…
Descriptors: Cerebral Palsy, Field Tests, Classification, Error of Measurement
National Centre for Vocational Education Research (NCVER), 2012
Developed for users of the Longitudinal Surveys of Australian Youth (LSAY), this user guide consolidates information about the LSAY 2009 cohort into one document. The guide aims to address all aspects of the LSAY data including: how to access the data; data restrictions; variable naming conventions; the structure of the data; documentation;…
Descriptors: Foreign Countries, Employment, Classification, Longitudinal Studies
Peer reviewed Peer reviewed
Brennan, Robert L.; Prediger, Dale J. – Educational and Psychological Measurement, 1981
This paper considers some appropriate and inappropriate uses of coefficient kappa and alternative kappa-like statistics. Discussion is restricted to the descriptive characteristics of these statistics for measuring agreement with categorical data in studies of reliability and validity. (Author)
Descriptors: Classification, Error of Measurement, Mathematical Models, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Kupermintz, Haggai – Journal of Educational Measurement, 2004
A decision-theoretic approach to the question of reliability in categorically scored examinations is explored. The concepts of true scores and errors are discussed as they deviate from conventional psychometric definitions and measurement error in categorical scores is cast in terms of misclassifications. A reliability measure based on…
Descriptors: Test Reliability, Error of Measurement, Psychometrics, Test Theory
Livingston, Samuel A. – 1976
A distinction is made between reliability of measurement and reliability of classification; the "criterion-referenced reliability coefficient" describes the former. Application of this coefficient to the probability distribution of possible scores for a single student yields a meaningful way to describe the reliability of a single score. (Author)
Descriptors: Classification, Criterion Referenced Tests, Error of Measurement, Measurement
Jaeger, Richard M. – 1975
Three new indicators of psychometric quality for objectives-based statewide assessments are proposed. These measures provide indication of the stability of reported data on item and objectives mastery, the validity of assessment items for members of various cultural groups, and the convergent validity of prescribed objectives mastery scores. The…
Descriptors: Classification, Cultural Influences, Educational Assessment, Error of Measurement
Peer reviewed Peer reviewed
Mellenbergh, Gideon J.; van der Linden, Wim J. – Applied Psychological Measurement, 1979
For six tests, coefficient delta as an index for internal optimality is computed. Internal optimality is defined as the magnitude of risk of the decision procedure with respect to the true score. Results are compared with an alternative index (coefficient kappa) for assessing the consistency of decisions. (Author/JKS)
Descriptors: Classification, Comparative Analysis, Decision Making, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Lei, Pui-Wa; Koehly, Laura M. – Journal of Experimental Education, 2003
Classification studies are important for practitioners who need to identify individuals for specialized treatment or intervention. When interventions are irreversible or misclassifications are costly, information about the proficiency of different classification procedures becomes invaluable. This study furnishes information about the relative…
Descriptors: Monte Carlo Methods, Classification, Discriminant Analysis, Regression (Statistics)
Forbes, Dean W. – 1976
Rasch calibration permitted the development of short achievement tests that were economical in testing time, and could be developed in a series of difficulty levels to suit student individual differences. Furthermore, these tests were of adequate reliability for practical educational measurement when individual students were assigned to tests of…
Descriptors: Academic Ability, Achievement Tests, Classification, Elementary Education