Publication Date
In 2025 | 1 |
Since 2024 | 3 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 7 |
Descriptor
Error of Measurement | 17 |
Statistical Bias | 17 |
Test Reliability | 17 |
Mathematical Models | 5 |
Test Items | 5 |
Item Analysis | 4 |
Simulation | 4 |
Adaptive Testing | 3 |
Computer Assisted Testing | 3 |
Correlation | 3 |
Latent Trait Theory | 3 |
More ▼ |
Source
Author
Publication Type
Reports - Research | 12 |
Journal Articles | 10 |
Reports - Evaluative | 3 |
Speeches/Meeting Papers | 2 |
Dissertations/Theses -… | 1 |
Information Analyses | 1 |
Non-Print Media | 1 |
Education Level
Grade 9 | 1 |
High Schools | 1 |
Higher Education | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Postsecondary Education | 1 |
Secondary Education | 1 |
Audience
Researchers | 1 |
Location
Indonesia | 1 |
South Carolina | 1 |
Virginia | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Comprehensive Tests of Basic… | 1 |
Test of Standard Written… | 1 |
What Works Clearinghouse Rating
Hsin-Yun Lee; You-Lin Chen; Li-Jen Weng – Journal of Experimental Education, 2024
The second version of Kaiser's Measure of Sampling Adequacy (MSA[subscript 2]) has been widely applied to assess the factorability of data in psychological research. The MSA[subscript 2] is developed in the population and little is known about its behavior in finite samples. If estimated MSA[subscript 2]s are biased due to sampling errors,…
Descriptors: Error of Measurement, Reliability, Sampling, Statistical Bias
Yanxuan Qu; Sandip Sinharay – ETS Research Report Series, 2024
The goal of this paper is to find better ways to estimate the internal consistency reliability of scores on tests with a specific type of design that are often encountered in practice: tests with constructed-response items clustered into sections that are not parallel or tau-equivalent, and one of the sections has only one item. To estimate the…
Descriptors: Test Reliability, Essay Tests, Construct Validity, Error of Measurement
Ole J. Kemi – Advances in Physiology Education, 2025
Students are assessed by coursework and/or exams, all of which are marked by assessors (markers). Student and marker performances are then subject to end-of-session board of examiner handling and analysis. This occurs annually and is the basis for evaluating students but also the wider learning and teaching efficiency of an academic institution.…
Descriptors: Undergraduate Students, Evaluation Methods, Evaluation Criteria, Academic Standards
Luke W. Miratrix; Jasjeet S. Sekhon; Alexander G. Theodoridis; Luis F. Campos – Grantee Submission, 2018
The popularity of online surveys has increased the prominence of using weights that capture units' probabilities of inclusion for claims of representativeness. Yet, much uncertainty remains regarding how these weights should be employed in analysis of survey experiments: Should they be used or ignored? If they are used, which estimators are…
Descriptors: Online Surveys, Weighted Scores, Data Interpretation, Robustness (Statistics)
Topczewski, Anna Marie – ProQuest LLC, 2013
Developmental score scales represent the performance of students along a continuum, where as students learn more they move higher along that continuum. Unidimensional item response theory (UIRT) vertical scaling has become a commonly used method to create developmental score scales. Research has shown that UIRT vertical scaling methods can be…
Descriptors: Item Response Theory, Scaling, Scores, Student Development
Kuo, Bor-Chen; Daud, Muslem; Yang, Chih-Wei – EURASIA Journal of Mathematics, Science & Technology Education, 2015
This paper describes a curriculum-based multidimensional computerized adaptive test that was developed for Indonesia junior high school Biology. In adherence to the Indonesian curriculum of different Biology dimensions, 300 items was constructed, and then tested to 2238 students. A multidimensional random coefficients multinomial logit model was…
Descriptors: Secondary School Science, Science Education, Science Tests, Computer Assisted Testing
Williams, Matt N.; Gomez Grajales, Carlos Alberto; Kurkiewicz, Dason – Practical Assessment, Research & Evaluation, 2013
In 2002, an article entitled "Four assumptions of multiple regression that researchers should always test" by Osborne and Waters was published in "PARE." This article has gone on to be viewed more than 275,000 times (as of August 2013), and it is one of the first results displayed in a Google search for "regression…
Descriptors: Multiple Regression Analysis, Misconceptions, Reader Response, Predictor Variables

Hunyh, Hunyh; Saunders, Joseph C. – 1979
Comparisons were made among various methods of estimating the reliability of pass-fail decisions based on mastery tests. The reliability indices that are considered are p, the proportion of agreements between two estimates, and kappa, the proportion of agreements corrected for chance. Estimates of these two indices were made on the basis of…
Descriptors: Cutting Scores, Error of Measurement, Mastery Tests, Reliability
Roberts, A. O. H. – New Directions for Testing and Measurement, 1980
When a test is made to separate a group into two distinct subgroups, as well as to provide an estimate of the mean performance of one subgroup (or both of them), such estimates will be biased due to the unreliability of the test. Suggestions for corrective procedures are presented. (RL)
Descriptors: Error of Measurement, Mathematical Formulas, Prognostic Tests, Regression (Statistics)

Huynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models

Kleinke, David J. – Applied Psychological Measurement, 1979
Lord's, Millman's and Saupe's methods of approximating the standard error of measurement are reviewed. Through an empirical demonstration involving 200 university classroom tests, all three approximations are shown to be biased. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Higher Education, Mathematical Formulas

Werts, C. E.; And Others – Educational and Psychological Measurement, 1980
Test-retest correlations can lead to biased reliability estimates when there is instability of true scores and/or when measurement errors are correlated. Using three administrations of the Test of Standard Written English and essay ratings, an analysis is demonstrated which separates true score instability and correlated errors. (Author/BW)
Descriptors: College Freshmen, Error of Measurement, Essay Tests, Higher Education
Shoemaker, David M. – 1972
Described and listed herein with concomitant sample input and output is the Fortran IV program which estimates parameters and standard errors of estimate per parameters for parameters estimated through multiple matrix sampling. The specific program is an improved and expanded version of an earlier version. (Author/BJG)
Descriptors: Computer Oriented Programs, Computer Programs, Error of Measurement, Error Patterns
Quinn, Jimmy L. – 1978
A logistic model was used to generate data to serve as a proxy for an immediate retest from item responses to a fourth grade standardized reading comprehension test of 45 items. Assuming that the actual test may be considered a pretest and the proxy data may be considered a retest, the effect of regression was investigated using a percentage of…
Descriptors: Correlation, Error of Measurement, Intermediate Grades, Item Analysis

Kim, Jwa K.; Nicewander, W. Alan – Psychometrika, 1993
Bias, standard error, and reliability of five ability estimators were evaluated using Monte Carlo estimates of the unknown conditional means and variances of the estimators. Results indicate that estimates based on Bayesian modal, expected a posteriori, and weighted likelihood estimators were reasonably unbiased with relatively small standard…
Descriptors: Ability, Bayesian Statistics, Equations (Mathematics), Error of Measurement
Previous Page | Next Page ยป
Pages: 1 | 2