NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)2
Since 2006 (last 20 years)4
Publication Type
Journal Articles27
Speeches/Meeting Papers27
Reports - Evaluative12
Reports - Research11
Reports - Descriptive2
Guides - Non-Classroom1
Opinion Papers1
Education Level
Adult Education1
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 27 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wang, Lin; Qian, Jiahe; Lee, Yi-Hsuan – ETS Research Report Series, 2018
Educational assessment data are often collected from a set of test centers across various geographic regions, and therefore the data samples contain clusters. Such cluster-based data may result in clustering effects in variance estimation. However, in many grouped jackknife variance estimation applications, jackknife groups are often formed by a…
Descriptors: Item Response Theory, Scaling, Equated Scores, Cluster Grouping
Peer reviewed Peer reviewed
Direct linkDirect link
Gorges, Julia; Maehler, Débora B.; Koch, Tobias; Offerhaus, Judith – Large-scale Assessments in Education, 2016
Background: Despite the importance of lifelong learning as a key to individual and societal prosperity, we know little about adult motivation to engage in learning across the lifespan. Building on educational psychological approaches, this article presents a measure of Motivation-to-Learn using four items from the background questionnaire of the…
Descriptors: Adult Learning, Learning Motivation, Factor Analysis, Questionnaires
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wang, Lin; Qian, Jiahe; Lee, Yi-Hsuan – ETS Research Report Series, 2013
The purpose of this study was to evaluate the combined effects of reduced equating sample size and shortened anchor test length on item response theory (IRT)-based linking and equating results. Data from two independent operational forms of a large-scale testing program were used to establish the baseline results for evaluating the results from…
Descriptors: Test Construction, Item Response Theory, Testing Programs, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Dimitrov, Dimiter M. – Mid-Western Educational Researcher, 2010
The focus of this presidential address is on the contemporary treatment of reliability and validity in educational assessment. Highlights on reliability are provided under the classical true-score model using tools from latent trait modeling to clarify important assumptions and procedures for reliability estimation. In addition to reliability,…
Descriptors: Educational Assessment, Validity, Item Response Theory, Reliability
Peer reviewed Peer reviewed
Barchard, Kimberly A.; Hakstian, A. Ralph – Multivariate Behavioral Research, 1997
Two studies, both using Type 12 sampling, are presented in which the effects of violating the assumption of essential parallelism in setting confidence intervals are studied. Results indicate that as long as data manifest properties of essential parallelism, the two methods studied maintain precise Type I error control. (SLD)
Descriptors: Error of Measurement, Robustness (Statistics), Sampling, Statistical Analysis
Peer reviewed Peer reviewed
Brennan, Robert L.; Lee, Won-Chan – Educational and Psychological Measurement, 1999
Develops two procedures for estimating individual-level conditional standard errors of measurement for scale scores, assuming tests of dichotomously scored items. Compares the two procedures to a polynomial procedure and a procedure developed by L. Feldt and A. Qualls (1998) using data from the Iowa Tests of Basic Skills. Contains 22 references.…
Descriptors: Error of Measurement, Estimation (Mathematics), Scaling, Scores
Peer reviewed Peer reviewed
Brennan, Robert L. – Applied Psychological Measurement, 1998
Provides a comprehensive and integrated treatment of both conditional absolute standard errors of measurement (SEM) and conditional relative SEMs from the perspective of generalizability theory. Illustrates the approach with examples from commercial standardized tests. Examples support the conclusion that both types of conditional SEMs tend to be…
Descriptors: Error of Measurement, Generalizability Theory, Raw Scores, Standardized Tests
Peer reviewed Peer reviewed
Henson, Robin K.; Kogan, Lori R.; Vacha-Haase, Tammi – Educational and Psychological Measurement, 2001
Studied sources of measurement error variance in the Teacher Efficacy Scale (TES) (Gibson and Dembo, 1984). Used reliability generalization to characterize the typical score reliability for the TES and potential sources of measurement error variance across 43 studies. Also examined related instruments for measurement integrity. (SLD)
Descriptors: Error of Measurement, Generalization, Meta Analysis, Psychometrics
Peer reviewed Peer reviewed
Rowley, Glenn L. – Journal of Educational Measurement, 1989
The focus on the individual that is possible in analyzing behavioral data provides the possibility of investigating sequencing effects. Autocorrelation--as illustrated with classroom data from a previous study--can cause standard procedures to underestimate the magnitude of measurement error. Recommendations are made to reduce the effects of…
Descriptors: Behavioral Science Research, Data Analysis, Error of Measurement, Estimation (Mathematics)
Peer reviewed Peer reviewed
Viswesvaran, Chockalingam; Ones, Deniz S. – Educational and Psychological Measurement, 2000
Used meta-analysis to cumulate reliabilities of personality scale scores, using 848 coefficients of stability and 1,359 internal consistency reliabilities across the Big Five factors of personality. The dimension of personality being measured does not appear to moderate strongly either internal consistency or the test-retest reliabilities.…
Descriptors: Error of Measurement, Meta Analysis, Personality Assessment, Personality Traits
Peer reviewed Peer reviewed
Lee, Won-Chan; Brennan, Robert L.; Kolen, Michael J. – Journal of Educational Measurement, 2000
Describes four procedures previously developed for estimating conditional standard errors of measurement for scale scores and compares them in a simulation study. All four procedures appear viable. Recommends that test users select a procedure based on various factors such as the type of scale score of concern, test characteristics, assumptions…
Descriptors: Error of Measurement, Estimation (Mathematics), Item Response Theory, Scaling
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy T. – Journal of Educational and Behavioral Statistics, 1996
Two possible standard error formulas for the polytomous differential item functioning index proposed by N. J. Dorans and A. P. Schmitt (1991) were derived. These standard errors, and associated hypothesis-testing procedures, were evaluated through simulated data. The standard error that performed better is based on N. Mantel's (1963)…
Descriptors: Error of Measurement, Evaluation Methods, Hypothesis Testing, Item Bias
Peer reviewed Peer reviewed
Milanowski, Anthony T. – Journal of Personnel Evaluation in Education, 1999
Describes the temporal consistency of school classification observed in the Kentucky, and secondarily in the Charlotte-Mecklinberg (North Carolina), school-based performance award programs. Data from the Kentucky Department of Education show the extent to which temporal inconsistency could be due to measurement error. (SLD)
Descriptors: Academic Achievement, Achievement Gains, Classification, Error of Measurement
Peer reviewed Peer reviewed
Smith, Richard M. – Journal of Outcome Measurement, 1998
Restrictions due to loss of freedom in estimation, the targeting of the instrument, and the presence of misfit in the data were studied through simulation as factors that influence the asymptotic standard errors for person measures. The underestimation of the observed standard deviation of ability in simulated data is discussed. (SLD)
Descriptors: Ability, Error of Measurement, Estimation (Mathematics), Goodness of Fit
Peer reviewed Peer reviewed
Dwyer, Carol Anne – Psychological Assessment, 1996
The uses and abuses of cut scores are examined. The article demonstrates (1) that cut scores always entail judgment; (2) that cut scores inherently result in misclassification; (3) that cut scores impose an artificial dichotomy on an essentially continuous distribution of knowledge, skill, or ability; and (4) that no true cut scores exist. (SLD)
Descriptors: Classification, Cutting Scores, Educational Testing, Error of Measurement
Previous Page | Next Page »
Pages: 1  |  2