Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 13 |
Descriptor
Error of Measurement | 20 |
Measurement | 20 |
Sampling | 20 |
Computation | 5 |
Research Methodology | 5 |
Scores | 5 |
Statistical Analysis | 5 |
Test Reliability | 5 |
Correlation | 4 |
Equated Scores | 4 |
Evaluation Methods | 4 |
More ▼ |
Source
Author
Ludtke, Oliver | 3 |
Marsh, Herbert W. | 3 |
Trautwein, Ulrich | 3 |
Nagengast, Benjamin | 2 |
Robitzsch, Alexander | 2 |
Abduljabbar, Adel S. | 1 |
Asparouhov, Tihomir | 1 |
Bramley, Tom | 1 |
Brennan, Robert L. | 1 |
Chun Wang | 1 |
Duong, Minh Q. | 1 |
More ▼ |
Publication Type
Journal Articles | 10 |
Reports - Research | 7 |
Reports - Descriptive | 4 |
Reports - Evaluative | 4 |
Speeches/Meeting Papers | 2 |
Dissertations/Theses -… | 1 |
Guides - General | 1 |
Numerical/Quantitative Data | 1 |
Education Level
Elementary Secondary Education | 2 |
Grade 8 | 2 |
Higher Education | 2 |
Postsecondary Education | 2 |
Secondary Education | 2 |
Adult Education | 1 |
Grade 7 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Audience
Location
Germany | 2 |
Arizona | 1 |
California | 1 |
Missouri | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Weicong Lyu; Chun Wang; Gongjun Xu – Grantee Submission, 2024
Data harmonization is an emerging approach to strategically combining data from multiple independent studies, enabling addressing new research questions that are not answerable by a single contributing study. A fundamental psychometric challenge for data harmonization is to create commensurate measures for the constructs of interest across…
Descriptors: Data Analysis, Test Items, Psychometrics, Item Response Theory
Bramley, Tom – Research Matters, 2020
The aim of this study was to compare, by simulation, the accuracy of mapping a cut-score from one test to another by expert judgement (using the Angoff method) versus the accuracy with a small-sample equating method (chained linear equating). As expected, the standard-setting method resulted in more accurate equating when we assumed a higher level…
Descriptors: Cutting Scores, Standard Setting (Scoring), Equated Scores, Accuracy
Michaelides, Michalis P.; Haertel, Edward H. – Applied Measurement in Education, 2014
The standard error of equating quantifies the variability in the estimation of an equating function. Because common items for deriving equated scores are treated as fixed, the only source of variability typically considered arises from the estimation of common-item parameters from responses of samples of examinees. Use of alternative, equally…
Descriptors: Equated Scores, Test Items, Sampling, Statistical Inference
Duong, Minh Q.; von Davier, Alina A. – International Journal of Testing, 2012
Test equating is a statistical procedure for adjusting for test form differences in difficulty in a standardized assessment. Equating results are supposed to hold for a specified target population (Kolen & Brennan, 2004; von Davier, Holland, & Thayer, 2004) and to be (relatively) independent of the subpopulations from the target population (see…
Descriptors: Ability Grouping, Difficulty Level, Psychometrics, Statistical Analysis
Ludtke, Oliver; Marsh, Herbert W.; Robitzsch, Alexander; Trautwein, Ulrich – Psychological Methods, 2011
In multilevel modeling, group-level variables (L2) for assessing contextual effects are frequently generated by aggregating variables from a lower level (L1). A major problem of contextual analyses in the social sciences is that there is no error-free measurement of constructs. In the present article, 2 types of error occurring in multilevel data…
Descriptors: Simulation, Educational Psychology, Social Sciences, Measurement
Marsh, Herbert W.; Ludtke, Oliver; Nagengast, Benjamin; Trautwein, Ulrich; Morin, Alexandre J. S.; Abduljabbar, Adel S.; Koller, Olaf – Educational Psychologist, 2012
Classroom context and climate are inherently classroom-level (L2) constructs, but applied researchers sometimes--inappropriately--represent them by student-level (L1) responses in single-level models rather than more appropriate multilevel models. Here we focus on important conceptual issues (distinctions between climate and contextual variables;…
Descriptors: Foreign Countries, Classroom Environment, Educational Research, Research Design
Wilson, Celia M. – ProQuest LLC, 2010
Research pertaining to the distortion of the squared canonical correlation coefficient has traditionally been limited to the effects of sampling error and associated correction formulas. The purpose of this study was to compare the degree of attenuation of the squared canonical correlation coefficient under varying conditions of score reliability.…
Descriptors: Monte Carlo Methods, Measurement, Multivariate Analysis, Error of Measurement
Micceri, Theodore; Parasher, Pradnya; Waugh, Gordon W.; Herreid, Charlene – Online Submission, 2009
An extensive review of the research literature and a study comparing over 36,000 survey responses with archival true scores indicated that one should expect a minimum of at least three percent random error for the least ambiguous of self-report measures. The Gulliver Effect occurs when a small proportion of error in a sizable subpopulation exerts…
Descriptors: Error of Measurement, Minority Groups, Measurement, Computation
Olsen, Robert B.; Unlu, Fatih; Price, Cristofer; Jaciw, Andrew P. – National Center for Education Evaluation and Regional Assistance, 2011
This report examines the differences in impact estimates and standard errors that arise when these are derived using state achievement tests only (as pre-tests and post-tests), study-administered tests only, or some combination of state- and study-administered tests. State tests may yield different evaluation results relative to a test that is…
Descriptors: Achievement Tests, Standardized Tests, State Standards, Reading Achievement
Marsh, Herbert W.; Ludtke, Oliver; Robitzsch, Alexander; Trautwein, Ulrich; Asparouhov, Tihomir; Muthen, Bengt; Nagengast, Benjamin – Multivariate Behavioral Research, 2009
This article is a methodological-substantive synergy. Methodologically, we demonstrate latent-variable contextual models that integrate structural equation models (with multiple indicators) and multilevel models. These models simultaneously control for and unconfound measurement error due to sampling of items at the individual (L1) and group (L2)…
Descriptors: Educational Environment, Context Effect, Models, Structural Equation Models
Wu, Margaret – Educational Measurement: Issues and Practice, 2010
In large-scale assessments, such as state-wide testing programs, national sample-based assessments, and international comparative studies, there are many steps involved in the measurement and reporting of student achievement. There are always sources of inaccuracies in each of the steps. It is of interest to identify the source and magnitude of…
Descriptors: Testing Programs, Educational Assessment, Measures (Individuals), Program Effectiveness
Gugiu, P. Cristian – Journal of MultiDisciplinary Evaluation, 2007
The constraints of conducting evaluations in real-world settings often necessitate the implementation of less than ideal designs. Unfortunately, the standard method for estimating the precision of a result (i.e., confidence intervals [CI]) cannot be used for evaluative conclusions that are derived from multiple indicators, measures, and data…
Descriptors: Measurement, Evaluation Methods, Evaluation Problems, Error of Measurement
Lord, Frederic M. – 1971
A simple, rigorous, small-sample statistical technique is described for testing the hypothesis that two sets of measurements differ only because of errors of measurement and because of differing origins and units of measurement. (Author)
Descriptors: Error of Measurement, Hypothesis Testing, Mathematical Applications, Mathematics

Whitely, Susan E. – Journal of Educational Measurement, 1977
A debate concerning specific issues and the general usefulness of the Rasch latent trait test model is continued. Methods of estimation, necessary sample size, and the applicability of the model are discussed. (JKS)
Descriptors: Error of Measurement, Item Analysis, Mathematical Models, Measurement

Wright, Benjamin D. – Journal of Educational Measurement, 1977
Statements made in a previous article of this journal concerning the Rasch latent trait test model are questioned. Methods of estimation, necessary sample sizes, several formuli, and the general usefulness of the Rasch model are discussed. (JKS)
Descriptors: Computers, Error of Measurement, Item Analysis, Mathematical Models
Previous Page | Next Page ยป
Pages: 1 | 2