Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 2 |
Descriptor
Computer Assisted Testing | 7 |
Item Response Theory | 7 |
Adaptive Testing | 6 |
Test Items | 4 |
Foreign Countries | 2 |
Scores | 2 |
Behavior Problems | 1 |
Certification | 1 |
Cheating | 1 |
Comparative Testing | 1 |
Computation | 1 |
More ▼ |
Author
Meijer, Rob R. | 7 |
van Krimpen-Stoop, Edith M.… | 3 |
Tendeiro, Jorge N. | 2 |
Egberink, Iris J. L. | 1 |
Publication Type
Journal Articles | 4 |
Reports - Research | 4 |
Reports - Evaluative | 2 |
Reports - Descriptive | 1 |
Education Level
Higher Education | 1 |
Audience
Location
Netherlands | 2 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Egberink, Iris J. L.; Meijer, Rob R.; Tendeiro, Jorge N. – Educational and Psychological Measurement, 2015
A popular method to assess measurement invariance of a particular item is based on likelihood ratio tests with all other items as anchor items. The results of this method are often only reported in terms of statistical significance, and researchers proposed different methods to empirically select anchor items. It is unclear, however, how many…
Descriptors: Personality Measures, Computer Assisted Testing, Measurement, Test Items
Tendeiro, Jorge N.; Meijer, Rob R. – Applied Psychological Measurement, 2012
This article extends the work by Armstrong and Shi on CUmulative SUM (CUSUM) person-fit methodology. The authors present new theoretical considerations concerning the use of CUSUM person-fit statistics based on likelihood ratios for the purpose of detecting cheating and random guessing by individual test takers. According to the Neyman-Pearson…
Descriptors: Cheating, Individual Testing, Adaptive Testing, Statistics
van Krimpen-Stoop, Edith M. L. A.; Meijer, Rob R. – 1999
Item scores that do not fit an assumed item response theory model may cause the latent trait value to be estimated inaccurately. Several person-fit statistics for detecting nonfitting score patterns for paper-and-pencil tests have been proposed. In the context of computerized adaptive tests (CAT), the use of person-fit analysis has hardly been…
Descriptors: Adaptive Testing, Computer Assisted Testing, Estimation (Mathematics), Item Response Theory

van Krimpen-Stoop, Edith M. L. A.; Meijer, Rob R. – Applied Psychological Measurement, 2002
Compared the nominal and empirical null distributions of the standardized log-likelihood statistic for polytomous items for paper-and-pencil (P&P) and computerized adaptive tests (CATs). Results show that the empirical distribution of the statistic differed from the assumed standard normal distribution for both P&P tests and CATs. Also…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Response Theory, Statistical Distributions
van Krimpen-Stoop, Edith M. L. A.; Meijer, Rob R. – 2000
Item scores that do not fit an assumed item response theory model may cause the latent trait value to be estimated inaccurately. For computerized adaptive tests (CAT) with dichotomous items, several person-fit statistics for detecting nonfitting item score patterns have been proposed. Both for paper-and-pencil (P&P) test and CATs, detection of…
Descriptors: Adaptive Testing, Computer Assisted Testing, Goodness of Fit, Item Response Theory
Meijer, Rob R. – 2001
Recent developments of person-fit analysis in computerized adaptive testing (CAT) are discussed. Methods from statistical process control are presented that have been proposed to classify an item score pattern as fitting or misfitting the underlying item response theory (IRT) model in a CAT. Most person-fit research in CAT is restricted to…
Descriptors: Adaptive Testing, Certification, Computer Assisted Testing, High Stakes Tests
Meijer, Rob R. – Journal of Educational Measurement, 2004
Two new methods have been proposed to determine unexpected sum scores on sub-tests (testlets) both for paper-and-pencil tests and computer adaptive tests. A method based on a conservative bound using the hypergeometric distribution, denoted p, was compared with a method where the probability for each score combination was calculated using a…
Descriptors: Probability, Adaptive Testing, Item Response Theory, Scores