NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
Elementary and Secondary…1
What Works Clearinghouse Rating
Showing 106 to 120 of 416 results Save | Export
Peer reviewed Peer reviewed
Bowers, John – Educational and Psychological Measurement, 1971
Descriptors: Error of Measurement, Mathematical Models, Test Reliability, True Scores
Peer reviewed Peer reviewed
Bond, Lloyd – Psychometrika, 1979
Tucker, Damarin, and Messick proposed a "base-free" measure of change which involves the computation of residual scores that are uncorrelated with true scores on the pretest. The present note discusses this change measure and demonstrates that properties they attribute to a are, in fact, properties of b. (Author/CTM)
Descriptors: Differences, Pretests Posttests, Research Reviews (Publications), Scores
Peer reviewed Peer reviewed
Conger, Anthony J. – Educational and Psychological Measurement, 1980
Reliability maximizing weights are related to theoretically specified true score scaling weights to show a constant relationship that is invariant under separate linear tranformations on each variable in the system. Test theoretic relations should be derived for the most general model available and not for unnecessarily constrained models.…
Descriptors: Mathematical Formulas, Scaling, Test Reliability, Test Theory
Peer reviewed Peer reviewed
Wilcox, Rand R. – Applied Psychological Measurement, 1979
Using a new coefficient, a rescaling of the Bayes risk is examined and a modification of this coefficient is described which yields an index that always has a value between zero and one. (Author/MH)
Descriptors: Bayesian Statistics, Measurement Techniques, Scoring, Technical Reports
Peer reviewed Peer reviewed
Dimitrov, Dimiter M. – Journal of Applied Measurement, 2003
Proposes formulas for expected true-score measures and reliability of binary items as a function of their Rasch difficulty when the trait (ability) distribution is normal or logistic. Provides an illustrative example for using the proposed formulas. (SLD)
Descriptors: Ability, Difficulty Level, Item Response Theory, Reliability
Peer reviewed Peer reviewed
Tisak, John; Tisak, Marie S. – Applied Psychological Measurement, 1996
Dynamic generalizations of reliability and validity that will incorporate longitudinal or developmental models, using latent curve analysis, are discussed. A latent curve model formulated to depict change is incorporated into the classical definitions of reliability and validity. The approach is illustrated with sociological and psychological…
Descriptors: Definitions, Development, Longitudinal Studies, Models
Peer reviewed Peer reviewed
Cliff, Norman – Psychometrika, 1989
This paper argues that: test data are ordinal; latent trait scores are only determined ordinally; and test data are used largely for ordinal purposes. A set of ordinal assumptions is presented, including an ordinal version of local independence. It is concluded that a purely ordinal test theory is possible. (TJH)
Descriptors: Equations (Mathematics), Latent Trait Theory, Regression (Statistics), True Scores
Peer reviewed Peer reviewed
Krus, David J.; Helmstadter, Gerald C. – Educational and Psychological Measurement, 1993
Negative coefficients of reliability, sometimes returned by the standard formula for estimation of the internal-consistency reliability, are neither theoretically nor numerically correct. Alternative strategies for test development in this special case are suggested. (Author)
Descriptors: Estimation (Mathematics), Reliability, Test Construction, Test Use
Peer reviewed Peer reviewed
Jiang, Hai; Stout, William – Journal of Educational and Behavioral Statistics, 1998
Proposes a new regression correction for the SIBTEST statistical tests (R. Shealy and W. Stout, 1993) that essentially uses a two-segment piecewise linear regression of the true on observed matching subtest scores. A simulation study illustrates the approach. (SLD)
Descriptors: Estimation (Mathematics), Item Bias, Regression (Statistics), Simulation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Haberman, Shelby J. – ETS Research Report Series, 2008
In educational testing, subscores may be provided based on a portion of the items from a larger test. One consideration in evaluation of such subscores is their ability to predict a criterion score. Two limitations on prediction exist. The first, which is well known, is that the coefficient of determination for linear prediction of the criterion…
Descriptors: Scores, Validity, Educational Testing, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Gaudron, Jean-Philippe; Vautier, Stephane – Journal of Vocational Behavior, 2007
This study aimed at estimating the correlation between true scores (true consistency) of vocational interest over a short time span in a sample of 1089 adults. Participants were administered 54 items assessing vocational, family, and leisure interests twice over a 1-month period. Responses were analyzed with a multitrait (MT) model, which supposes…
Descriptors: Vocational Interests, Correlation, True Scores, Longitudinal Studies
Stocking, Martha L.; And Others – 1988
A sequence of simulations was carried out to aid in the diagnosis and interpretation of equating differences found between random and matched (nonrandom) samples for four commonly used equating procedures: (1) Tucker linear observed-score equating; (2) Levine equally reliable linear observed-score equating; (3) equipercentile curvilinear…
Descriptors: Equated Scores, Item Response Theory, Sample Size, Simulation
Peer reviewed Peer reviewed
Schulman, Robert S.; Haden, Richard L. – Psychometrika, 1975
A model is proposed for the description of ordinal test scores based on the definition of true score as expected rank; its deviations are compared with results from classical test theory. An unbiased estimator of population true score from sample data is calculated. Score variance and population reliability are examined. (Author/BJG)
Descriptors: Career Development, Mathematical Models, Test Reliability, Test Theory
Peer reviewed Peer reviewed
Kearns, Jack; Meredith, William – Psychometrika, 1975
Examines the question of how large a sample must be in order to produce empirical Bayes estimates which are preferable to other commonly used estimates, such as proportion correct observed score. (Author/RC)
Descriptors: Bayesian Statistics, Item Analysis, Probability, Sampling
Peer reviewed Peer reviewed
Ng, K. T. – Educational and Psychological Measurement, 1974
This paper is aimed at demonstrating that Charles Spearman postulated neither a platonic true-error distinction nor a requirement for constant true scores under repeated measurement. (Author/RC)
Descriptors: Career Development, Correlation, Models, Test Reliability
Pages: 1  |  ...  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  12  |  ...  |  28