Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 5 |
Descriptor
Error of Measurement | 8 |
Test Reliability | 8 |
Weighted Scores | 8 |
Test Validity | 3 |
Evaluation Methods | 2 |
Evaluation Problems | 2 |
Item Analysis | 2 |
Item Response Theory | 2 |
Predictor Variables | 2 |
Statistical Bias | 2 |
Test Items | 2 |
More ▼ |
Source
Applied Measurement in… | 2 |
Grantee Submission | 2 |
Educational and Psychological… | 1 |
Journal of Educational and… | 1 |
Practical Assessment,… | 1 |
Author
Publication Type
Journal Articles | 6 |
Reports - Research | 6 |
Reports - Evaluative | 2 |
Speeches/Meeting Papers | 1 |
Education Level
Audience
Researchers | 1 |
Location
Virginia | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Ke-Hai Yuan; Zhiyong Zhang; Lijuan Wang – Grantee Submission, 2024
Mediation analysis plays an important role in understanding causal processes in social and behavioral sciences. While path analysis with composite scores was criticized to yield biased parameter estimates when variables contain measurement errors, recent literature has pointed out that the population values of parameters of latent-variable models…
Descriptors: Structural Equation Models, Path Analysis, Weighted Scores, Comparative Testing
Grabovsky, Irina; Wainer, Howard – Journal of Educational and Behavioral Statistics, 2017
In this article, we extend the methodology of the Cut-Score Operating Function that we introduced previously and apply it to a testing scenario with multiple independent components and different testing policies. We derive analytically the overall classification error rate for a test battery under the policy when several retakes are allowed for…
Descriptors: Cutting Scores, Weighted Scores, Classification, Testing
Luke W. Miratrix; Jasjeet S. Sekhon; Alexander G. Theodoridis; Luis F. Campos – Grantee Submission, 2018
The popularity of online surveys has increased the prominence of using weights that capture units' probabilities of inclusion for claims of representativeness. Yet, much uncertainty remains regarding how these weights should be employed in analysis of survey experiments: Should they be used or ignored? If they are used, which estimators are…
Descriptors: Online Surveys, Weighted Scores, Data Interpretation, Robustness (Statistics)
Phillips, Gary W. – Applied Measurement in Education, 2015
This article proposes that sampling design effects have potentially huge unrecognized impacts on the results reported by large-scale district and state assessments in the United States. When design effects are unrecognized and unaccounted for they lead to underestimating the sampling error in item and test statistics. Underestimating the sampling…
Descriptors: State Programs, Sampling, Research Design, Error of Measurement
Williams, Matt N.; Gomez Grajales, Carlos Alberto; Kurkiewicz, Dason – Practical Assessment, Research & Evaluation, 2013
In 2002, an article entitled "Four assumptions of multiple regression that researchers should always test" by Osborne and Waters was published in "PARE." This article has gone on to be viewed more than 275,000 times (as of August 2013), and it is one of the first results displayed in a Google search for "regression…
Descriptors: Multiple Regression Analysis, Misconceptions, Reader Response, Predictor Variables
Thompson, Bruce; Frankiewicz, Ronald G. – 1980
A procedure for estimating reliability in a factor analytic context, when reliability of the extracted factors is not an emphasis, is identified. The procedure is an extension of Dressel's work and might be applied in attitude measurement. It assesses how homogeneous the weighted original item responses are, when they are scored for pattern…
Descriptors: Attitude Measures, Error of Measurement, Factor Analysis, Measures (Individuals)

MisLevy, Robert J.; Bock, R. Darrell – Educational and Psychological Measurement, 1982
An alternative biweight estimator based on Tukey's is examined in which (1) test disturbances are not assumed to be the same for all subjects, (2) each response is utilized proportional to its value, and (3) the biweight and maximum likelihood estimate agree when no disturbances are present. Smaller mean-squared errors are shown. (Author/CM)
Descriptors: Error of Measurement, Estimation (Mathematics), Guessing (Tests), Latent Trait Theory
Sykes, Robert C.; Hou, Liling – Applied Measurement in Education, 2003
Weighting responses to Constructed-Response (CR) items has been proposed as a way to increase the contribution these items make to the test score when there is insufficient testing time to administer additional CR items. The effect of various types of weighting items of an IRT-based mixed-format writing examination was investigated.…
Descriptors: Item Response Theory, Weighted Scores, Responses, Scores