Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 6 |
Descriptor
Source
Journal of Educational… | 35 |
Author
Clauser, Brian E. | 2 |
Frary, Robert B. | 2 |
Miller, Timothy R. | 2 |
Nandakumar, Ratna | 2 |
Wilson, Mark | 2 |
Zwick, Rebecca | 2 |
Ackerman, Terry A. | 1 |
Andrich, David | 1 |
Briggs, Derek C. | 1 |
Chang, Hua-Hua | 1 |
Clyman, Stephen G. | 1 |
More ▼ |
Publication Type
Journal Articles | 35 |
Reports - Research | 18 |
Reports - Evaluative | 17 |
Speeches/Meeting Papers | 3 |
Education Level
Secondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Program for International… | 1 |
What Works Clearinghouse Rating
Herborn, Katharina; Mustafic, Maida; Greiff, Samuel – Journal of Educational Measurement, 2017
Collaborative problem solving (CPS) assessment is a new academic research field with a number of educational implications. In 2015, the Programme for International Student Assessment (PISA) assessed CPS with a computer-simulated human-agent (H-A) approach that claimed to measure 12 individual CPS skills for the first time. After reviewing the…
Descriptors: Cooperative Learning, Problem Solving, Computer Simulation, Evaluation Methods
Wang, Shiyu; Lin, Haiyan; Chang, Hua-Hua; Douglas, Jeff – Journal of Educational Measurement, 2016
Computerized adaptive testing (CAT) and multistage testing (MST) have become two of the most popular modes in large-scale computer-based sequential testing. Though most designs of CAT and MST exhibit strength and weakness in recent large-scale implementations, there is no simple answer to the question of which design is better because different…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Format, Sequential Approach
Wilson, Mark; Gochyyev, Perman; Scalise, Kathleen – Journal of Educational Measurement, 2017
This article summarizes assessment of cognitive skills through collaborative tasks, using field test results from the Assessment and Teaching of 21st Century Skills (ATC21S) project. This project, sponsored by Cisco, Intel, and Microsoft, aims to help educators around the world enable students with the skills to succeed in future career and…
Descriptors: Cognitive Ability, Thinking Skills, Evaluation Methods, Educational Assessment
Zimmerman, Donald W. – Journal of Educational Measurement, 2009
This study was an investigation of the relation between the reliability of difference scores, considered as a parameter characterizing a population of examinees, and the reliability estimates obtained from random samples from the population. The parameters in familiar equations for the reliability of difference scores were redefined in such a way…
Descriptors: Computer Simulation, Reliability, Population Groups, Scores
Kim, Seonghoon; Feldt, Leonard S. – Journal of Educational Measurement, 2008
This article extends the Bonett (2003a) approach to testing the equality of alpha coefficients from two independent samples to the case of m [greater than or equal] 2 independent samples. The extended Fisher-Bonett test and its competitor, the Hakstian-Whalen (1976) test, are illustrated with numerical examples of both hypothesis testing and power…
Descriptors: Tests, Comparative Analysis, Hypothesis Testing, Error of Measurement
Briggs, Derek C.; Wilson, Mark – Journal of Educational Measurement, 2007
An approach called generalizability in item response modeling (GIRM) is introduced in this article. The GIRM approach essentially incorporates the sampling model of generalizability theory (GT) into the scaling model of item response theory (IRT) by making distributional assumptions about the relevant measurement facets. By specifying a random…
Descriptors: Markov Processes, Generalizability Theory, Item Response Theory, Computation

Oshima, Takako C.; Miller, M. David – Journal of Educational Measurement, 1990
A bidimensional 2-parameter logistic model was applied to data generated for 2 groups on a 40-item test. Item parameters were the same across groups; correlation across the 2 traits varied. Results indicate the need for caution in using item-response theory (IRT)-based invariance indexes with multidimensional data for these groups. (TJH)
Descriptors: Computer Simulation, Correlation, Discriminant Analysis, Item Response Theory

French, Ann W.; Miller, Timothy R. – Journal of Educational Measurement, 1996
A computer simulation study was conducted to determine the feasibility of using logistic regression procedures to detect differential item functioning (DIF) in polytomous items. Results indicate that logistic regression is powerful in detecting most forms of DIF, although it requires large amounts of data manipulation and careful interpretation.…
Descriptors: Computer Simulation, Identification, Item Bias, Test Interpretation

Wilcox, Rand R. – Journal of Educational Measurement, 1987
Four procedures are discussed for obtaining a confidence interval when answer-until-correct scoring is used in multiple choice tests. Simulated data show that the choice of procedure depends upon sample size. (GDC)
Descriptors: Computer Simulation, Multiple Choice Tests, Sample Size, Scoring

Oshima, T. C. – Journal of Educational Measurement, 1994
The effect of violating the assumption of nonspeededness on ability and item parameter estimates in item response theory was studied through simulation under three speededness conditions. Results indicate that ability estimation was least affected by speededness but that substantial effects on item parameter estimates were found. (SLD)
Descriptors: Ability, Computer Simulation, Estimation (Mathematics), Item Response Theory

Frary, Robert B. – Journal of Educational Measurement, 1989
Responses to a 50-item, 4-choice test were simulated for 1,000 examinees under conventional formula-scoring instructions. Based on 192 simulation runs, formula scores and expected formula scores were determined for each examinee allowing and not allowing for inappropriate omissions. (TJH)
Descriptors: Computer Simulation, Difficulty Level, Guessing (Tests), Multiple Choice Tests

Parshall, Cynthia G.; Miller, Timothy R. – Journal of Educational Measurement, 1995
Exact testing was evaluated as a method for conducting Mantel-Haenszel differential item functioning (DIF) analyses with relatively small samples. A series of computer simulations found that the asymptotic Mantel-Haenszel and the exact method yielded very similar results across sample size, levels of DIF, and data sets. (SLD)
Descriptors: Comparative Analysis, Computer Simulation, Identification, Item Bias
Kamata, Akihito; Tate, Richard – Journal of Educational Measurement, 2005
The goal of this study was the development of a procedure to predict the equating error associated with the long-term equating method of Tate (2003) for mixed-format tests. An expression for the determination of the error of an equating based on multiple links using the error for the component links was derived and illustrated with simulated data.…
Descriptors: Computer Simulation, Item Response Theory, Test Format, Evaluation Methods

Reise, Steve P.; Yu, Jiayuan – Journal of Educational Measurement, 1990
Parameter recovery in the graded-response model was investigated using the MULTILOG computer program under default conditions. Results from 36 simulated data sets suggest that at least 500 examinees are needed to achieve adequate calibration under the graded model. Sample size had little influence on the true ability parameter's recovery. (SLD)
Descriptors: Computer Assisted Testing, Computer Simulation, Computer Software, Estimation (Mathematics)

Andrich, David – Journal of Educational Measurement, 1989
The distinction between deterministic and statistical reasoning in the application of models to educational measurement is explicated. Issues addressed include the relationship between data and estimation equations, distinction between parameters and parameter estimates, and power of tests of fit of responses across the ability continuum. (TJH)
Descriptors: Computer Simulation, Equations (Mathematics), Estimation (Mathematics), Goodness of Fit