NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers3
Location
Australia1
Oregon1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 47 results Save | Export
Strauss, Christian L. L. – ProQuest LLC, 2022
In many psychological and educational applications, it is imperative to obtain valid and reliable score estimates of multilevel processes. For example, in order to assess the quality and characteristics of high impact learning processes, one must compute accurate scores representative of student- and classroom-level constructs. Currently, there…
Descriptors: Scores, Factor Analysis, Models, True Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Kylie Gorney; Sandip Sinharay – Journal of Educational Measurement, 2025
Although there exists an extensive amount of research on subscores and their properties, limited research has been conducted on categorical subscores and their interpretations. In this paper, we focus on the claim of Feinberg and von Davier that categorical subscores are useful for remediation and instructional purposes. We investigate this claim…
Descriptors: Tests, Scores, Test Interpretation, Alternative Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2019
We derive formulas for the differential item functioning (DIF) measures that two routinely used DIF statistics are designed to estimate. The DIF measures that match on observed scores are compared to DIF measures based on an unobserved ability (theta or true score) for items that are described by either the one-parameter logistic (1PL) or…
Descriptors: Scores, Test Bias, Statistical Analysis, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Yi-Hsuan; Zhang, Jinming – International Journal of Testing, 2017
Simulations were conducted to examine the effect of differential item functioning (DIF) on measurement consequences such as total scores, item response theory (IRT) ability estimates, and test reliability in terms of the ratio of true-score variance to observed-score variance and the standard error of estimation for the IRT ability parameter. The…
Descriptors: Test Bias, Test Reliability, Performance, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Jurich, Daniel P.; DeMars, Christine E.; Goodman, Joshua T. – Applied Psychological Measurement, 2012
The prevalence of high-stakes test scores as a basis for significant decisions necessitates the dissemination of accurate and fair scores. However, the magnitude of these decisions has created an environment in which examinees may be prone to resort to cheating. To reduce the risk of cheating, multiple test forms are commonly administered. When…
Descriptors: High Stakes Tests, Scores, Prevention, Cheating
Andrews, Benjamin James – ProQuest LLC, 2011
The equity properties can be used to assess the quality of an equating. The degree to which expected scores conditional on ability are similar between test forms is referred to as first-order equity. Second-order equity is the degree to which conditional standard errors of measurement are similar between test forms after equating. The purpose of…
Descriptors: Test Format, Advanced Placement, Simulation, True Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Leue, Anja; Lange, Sebastian – Assessment, 2011
The assessment of positive affect (PA) and negative affect (NA) by means of the Positive Affect and Negative Affect Schedule has received a remarkable popularity in the social sciences. Using a meta-analytic tool--namely, reliability generalization (RG)--population reliability scores of both scales have been investigated on the basis of a random…
Descriptors: Social Sciences, True Scores, Generalization, Affective Behavior
Peer reviewed Peer reviewed
Direct linkDirect link
Stoolmiller, Michael; Biancarosa, Gina; Fien, Hank – Assessment for Effective Intervention, 2013
Lack of psychometric equivalence of oral reading fluency (ORF) passages used within a grade for screening and progress monitoring has recently become an issue with calls for the use of equating methods to ensure equivalence. To investigate the nature of the nonequivalence and to guide the choice of equating method to correct for nonequivalence,…
Descriptors: School Personnel, Reading Fluency, Emergent Literacy, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
MacCann, Robert G. – Educational and Psychological Measurement, 2008
It is shown that the Angoff and bookmarking cut scores are examples of true score equating that in the real world must be applied to observed scores. In the context of defining minimal competency, the percentage "failed" by such methods is a function of the length of the measuring instrument. It is argued that this length is largely…
Descriptors: True Scores, Cutting Scores, Minimum Competencies, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Haberman, Shelby J. – ETS Research Report Series, 2008
In educational testing, subscores may be provided based on a portion of the items from a larger test. One consideration in evaluation of such subscores is their ability to predict a criterion score. Two limitations on prediction exist. The first, which is well known, is that the coefficient of determination for linear prediction of the criterion…
Descriptors: Scores, Validity, Educational Testing, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Biswas, Ajoy Kumar – Applied Psychological Measurement, 2006
This article studies the ordinal reliability of (total) test scores. This study is based on a classical-type linear model of observed score (X), true score (T), and random error (E). Based on the idea of Kendall's tau-a coefficient, a measure of ordinal reliability for small-examinee populations is developed. This measure is extended to large…
Descriptors: True Scores, Test Theory, Test Reliability, Scores
Peer reviewed Peer reviewed
Tucker, Ledyard R. – Psychometrika, 1979
A correction by Bond (TM 504 914) of an error by Tucker, Damarin, and Messick is acknowledged. A formula for the correlation between initial true test scores and true difference scores is presented. Observed score considerations should be replaced by emphasis on true score considerations. (Author/CTM)
Descriptors: Differences, Mathematical Formulas, Pretests Posttests, Scores
Peer reviewed Peer reviewed
Marks, Edmond; Lindsay, Carl A. – Journal of Educational Measurement, 1972
Examines the effects of four parameters on the accuracy of test equating under a relaxed definition of test form equivalence. The four parameters studied were sample size, test form length, test form reliability, and the correlation between true scores of the test forms to be equated. (CK)
Descriptors: Scores, Test Interpretation, Test Reliability, Test Results
Peer reviewed Peer reviewed
Bond, Lloyd – Psychometrika, 1979
Tucker, Damarin, and Messick proposed a "base-free" measure of change which involves the computation of residual scores that are uncorrelated with true scores on the pretest. The present note discusses this change measure and demonstrates that properties they attribute to a are, in fact, properties of b. (Author/CTM)
Descriptors: Differences, Pretests Posttests, Research Reviews (Publications), Scores
Peer reviewed Peer reviewed
Williams, Richard H.; Zimmerman, Donald W. – Educational and Psychological Measurement, 1977
The usual formulas for the reliability of differences between two test scores are based on the assumption that the error scores are uncorrelated. Formulas are presented for the general case where this assumption is unnecessary. (Author/JKS)
Descriptors: Correlation, Error of Measurement, Error Patterns, Scores
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4