NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)2
Since 2006 (last 20 years)7
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 18 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Chun Wang; Ping Chen; Shengyu Jiang – Journal of Educational Measurement, 2020
Many large-scale educational surveys have moved from linear form design to multistage testing (MST) design. One advantage of MST is that it can provide more accurate latent trait [theta] estimates using fewer items than required by linear tests. However, MST generates incomplete response data by design; hence, questions remain as to how to…
Descriptors: Test Construction, Test Items, Adaptive Testing, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Reckase, Mark D. – Educational and Psychological Measurement, 2012
This study investigates how different rounding rules and ways of providing Angoff standard-setting judgments affect cut-scores. A simulation design based on data from the National Assessment of Education Progress was used to investigate how rounding judgments to the nearest whole number (e.g., 0, 1, 2, etc.), nearest 0.05, or nearest two decimal…
Descriptors: Standard Setting, Cutting Scores, Statistical Bias, Numbers
Peer reviewed Peer reviewed
Direct linkDirect link
Sachse, Karoline A.; Roppelt, Alexander; Haag, Nicole – Journal of Educational Measurement, 2016
Trend estimation in international comparative large-scale assessments relies on measurement invariance between countries. However, cross-national differential item functioning (DIF) has been repeatedly documented. We ran a simulation study using national item parameters, which required trends to be computed separately for each country, to compare…
Descriptors: Comparative Analysis, Measurement, Test Bias, Simulation
Hornback, Joseph E. – ProQuest LLC, 2013
This dissertation addresses two research questions: 1. Do states misrepresent their progress on their own state assessments? 2. If states do distort their progress, are their predictors to suggest why this distortion occurs? The first research question requires that distortion be defined. For the purposes of this dissertation I calculated the…
Descriptors: Standardized Tests, State Standards, Computation, Equations (Mathematics)
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Deping; Oranje, Andreas; Jiang, Yanlin – Journal of Educational and Behavioral Statistics, 2009
To find population proficiency distributions, a two-level hierarchical linear model may be applied to large-scale survey assessments such as the National Assessment of Educational Progress (NAEP). The model and parameter estimation are developed and a simulation was carried out to evaluate parameter recovery. Subsequently, both a hierarchical and…
Descriptors: Computation, National Competency Tests, Measurement, Regression (Statistics)
Chauncey, Caroline T., Ed. – Harvard Education Press, 2010
"Harvard Education Letter" is published bimonthly at the Harvard Graduate School of Education. This issue of "Harvard Education Letter" contains the following articles: (1) Online Testing, Version 1.0: Oregon's Adaptive Computer-Based Accountability Test Offers a Peek at a Brave New Future (Robert Rothman); (2) Beyond…
Descriptors: Family Programs, Homosexuality, Educational Policy, Sexual Identity
Peer reviewed Peer reviewed
Muraki, Eiji – Journal of Educational Measurement, 1999
Extended an Item Response Theory (IRT) method for detection of differential item functioning to the partial credit model and applied the method to simulated data using a stepwise procedure. Then applied the stepwise DIF analysis based on the multiple-group partial credit model to writing trend data from the National Assessment of Educational…
Descriptors: Groups, Item Bias, Item Response Theory, Simulation
Lau, C. Allen; Wang, Tianyou – 1999
A study was conducted to extend the sequential probability ratio testing (SPRT) procedure with the polytomous model under some practical constraints in computerized classification testing (CCT), such as methods to control item exposure rate, and to study the effects of other variables, including item information algorithms, test difficulties, item…
Descriptors: Algorithms, Computer Assisted Testing, Difficulty Level, Item Banks
Lau, C. Allen; Wang, Tianyou – 1998
The purposes of this study were to: (1) extend the sequential probability ratio testing (SPRT) procedure to polytomous item response theory (IRT) models in computerized classification testing (CCT); (2) compare polytomous items with dichotomous items using the SPRT procedure for their accuracy and efficiency; (3) study a direct approach in…
Descriptors: Computer Assisted Testing, Cutting Scores, Item Response Theory, Mastery Tests
Peer reviewed Peer reviewed
Muraki, Eiji – Applied Psychological Measurement, 1992
The partial credit model with a varying slope parameter is developed and called the generalized partial credit model (GPCM). Analysis results for simulated data by this and other polytomous item-response models demonstrate that the rating formulation of the GPCM is adaptable to the analysis of polytomous item responses. (SLD)
Descriptors: Algorithms, Equations (Mathematics), Generalization, Item Response Theory
Yamamoto, Kentaro; Muraki, Eiji – 1991
The extent to which properties of the ability scale and the form of the latent trait distribution influence the estimated item parameters of item response theory (IRT) was investigated using real and simulated data. Simulated data included 5,000 ability values randomly drawn from the standard normal distribution. Real data included the results for…
Descriptors: Ability, Estimation (Mathematics), Graphs, Item Response Theory
Peer reviewed Peer reviewed
Bloxom, Bruce; And Others – Journal of Educational and Behavioral Statistics, 1995
Develops and evaluates the linkage of the Armed Services Vocational Aptitude Battery to the mathematics scale of the National Assessment of Educational Progress. The accuracy of the proficiency distribution estimated from the projection was close to the accuracy of the distribution estimated from the large scale assessment. (SLD)
Descriptors: Educational Assessment, Estimation (Mathematics), Evaluation Methods, Mathematics Tests
Peer reviewed Peer reviewed
Cohen, Jon; Snow, Stephanie – Journal of Educational Measurement, 2002
Studied the impact of changes in item difficulty on National Assessment of Educational Progress (NAEP) estimates over time through a Monte Carlo study that simulated the responses of 1990 NAEP mathematics respondents to 1990 and 1996 NAEP mathematics items. Results support the idea that these changes have not affected the NAEP trend line.…
Descriptors: Change, Difficulty Level, Estimation (Mathematics), Mathematics Tests
Nandakumar, Ratna – 1991
Performance in assessing the unidimensionality of tests was examined for four methods: (1) W. F. Stout's procedure (1987); (2) the approach of P. W. Holland and P. R. Rosenbaum (1986); (3) linear factor analysis; and (4) non-linear factor analysis. Each method was examined and compared with the others using simulated and real test data. Seven data…
Descriptors: Ability Identification, Comparative Analysis, Correlation, Evaluation Methods
Lau, C. Allen; Wang, Tianyou – 2000
This paper proposes a new Information-Time index as the basis for item selection in computerized classification testing (CCT) and investigates how this new item selection algorithm can help improve test efficiency for item pools with mixed item types. It also investigates how practical constraints such as item exposure rate control, test…
Descriptors: Algorithms, Classification, Computer Assisted Testing, Elementary Secondary Education
Previous Page | Next Page ยป
Pages: 1  |  2