NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Type
Reports - Research18
Journal Articles7
Speeches/Meeting Papers5
Audience
Researchers4
Location
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 18 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Xiao, Leifeng; Hau, Kit-Tai – Applied Measurement in Education, 2023
We compared coefficient alpha with five alternatives (omega total, omega RT, omega h, GLB, and coefficient H) in two simulation studies. Results showed for unidimensional scales, (a) all indices except omega h performed similarly well for most conditions; (b) alpha is still good; (c) GLB and coefficient H overestimated reliability with small…
Descriptors: Test Theory, Test Reliability, Factor Analysis, Test Length
Peer reviewed Peer reviewed
Direct linkDirect link
DeCarlo, Lawrence T. – Journal of Educational Measurement, 2023
A conceptualization of multiple-choice exams in terms of signal detection theory (SDT) leads to simple measures of item difficulty and item discrimination that are closely related to, but also distinct from, those used in classical item analysis (CIA). The theory defines a "true split," depending on whether or not examinees know an item,…
Descriptors: Multiple Choice Tests, Test Items, Item Analysis, Test Wiseness
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kim, Sooyeon; Livingston, Samuel A. – ETS Research Report Series, 2017
The purpose of this simulation study was to assess the accuracy of a classical test theory (CTT)-based procedure for estimating the alternate-forms reliability of scores on a multistage test (MST) having 3 stages. We generated item difficulty and discrimination parameters for 10 parallel, nonoverlapping forms of the complete 3-stage test and…
Descriptors: Accuracy, Test Theory, Test Reliability, Adaptive Testing
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kogar, Hakan – International Journal of Assessment Tools in Education, 2018
The aim of this simulation study, determine the relationship between true latent scores and estimated latent scores by including various control variables and different statistical models. The study also aimed to compare the statistical models and determine the effects of different distribution types, response formats and sample sizes on latent…
Descriptors: Simulation, Context Effect, Computation, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Reimann, Peter; Kickmeier-Rust, Michael; Albert, Dietrich – Computers & Education, 2013
This paper explores the relation between problem solving learning environments (PSLEs) and assessment concepts. The general framework of evidence-centered assessment design is used to describe PSLEs in terms of assessment concepts, and to identify similarities between the process of assessment design and of PSLE design. We use a recently developed…
Descriptors: Teaching Methods, Psychometrics, Problem Solving, Test Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational Measurement, 2010
Recently, there has been an increasing level of interest in subscores for their potential diagnostic value. Haberman suggested a method based on classical test theory to determine whether subscores have added value over total scores. In this article I first provide a rich collection of results regarding when subscores were found to have added…
Descriptors: Scores, Test Theory, Simulation, Reliability
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zhang, Jinming – ETS Research Report Series, 2004
This paper extends the theory of conditional covariances to polytomous items. It has been mathematically proven that under some mild conditions, commonly assumed in the analysis of response data, the conditional covariance of two items, dichotomously or polytomously scored, is positive if the two items are dimensionally homogeneous and negative…
Descriptors: Test Items, Test Theory, Correlation, National Competency Tests
Bogan, Evelyn Doody; Yen, Wendy M. – 1983
Four multidimensional data configurations and one unidimensional data configuration were simulated for three differences in mean difficulty between two tests to be equated. Two chi-square statistics, Q1 and Q2, were examined for their ability to detect multidimensionality. Results indicated that Q1 did not discriminate between any of the…
Descriptors: Difficulty Level, Equated Scores, Goodness of Fit, Latent Trait Theory
Epstein, Kenneth I.; Knerr, Claramae S. – 1976
The literature on criterion referenced testing is full of discussions concerning whether classical measurement techniques are appropriate, whether variance is necessary, whether new indices of reliability are needed, and the like. What appears to be lacking, however, is a clear and simple discussion of why the problems occur. This paper suggests…
Descriptors: Career Development, Criterion Referenced Tests, Item Analysis, Item Sampling
Cope, Ronald T. – 1986
Comparisons were made of three Angoff Design V linear equating methods (two forms equated to a common test, two forms predicted by a common test, or two forms used to predict a common test) and Tucker's and R. Levine's linear methods, under common item linear equating with non-equivalent populations. Forms of a professional certification test…
Descriptors: Certification, Comparative Analysis, Equated Scores, Higher Education
Becker, Betsy Jane – 1986
This paper discusses distribution theory and power computations for four common "tests of combined significance." These tests are calculated using one-sided sample probabilities or p values from independent studies (or hypothesis tests), and provide an overall significance level for the series of results. Noncentral asymptotic sampling…
Descriptors: Achievement Tests, Correlation, Effect Size, Hypothesis Testing
Weitzman, R. A. – 1982
The goal of this research was to predict from a recruit's responses to the Armed Services Vocational Aptitude Battery (ASVAB) items whether the recruit would pass the Armed Forces Qualification Test (AFQT). The data consisted of the responses (correct/incorrect) of 1,020 Navy recruits to 200 items of the ASVAB together with the scores of these…
Descriptors: Adults, Armed Forces, Computer Oriented Programs, Computer Simulation
Vale, C. David; And Others – 1981
A simulation study to determine appropriate linking methods for adaptive testing items was designed. Three basic data sets for responses were created. These were randomly sampled, systematically sampled, and selected data sets. The evaluative criteria used were fidelity of parameter estimation, asymptotic ability estimates, root-mean-square error…
Descriptors: Adaptive Testing, Aptitude Tests, Armed Forces, Bayesian Statistics
Hambleton, Ronald K.; Cook, Linda L. – 1978
The purpose of the present research was to study, systematically, the "goodness-of-fit" of the one-, two-, and three-parameter logistic models. We studied, using computer-simulated test data, the effects of four variables: variation in item discrimination parameters, the average value of the pseudo-chance level parameters, test length,…
Descriptors: Career Development, Difficulty Level, Goodness of Fit, Item Analysis
Marshall, J. Laird – 1976
A summary is provided of the rationale for questioning the applicability of classical reliability measures to criterion referenced tests; an extension of the classical theory of true and error scores to incorporate a theory of dichotomous decisions; a presentation of the mean split-half coefficient of agreement, a single-administration test index…
Descriptors: Career Development, Computer Programs, Criterion Referenced Tests, Decision Making
Previous Page | Next Page ยป
Pages: 1  |  2