Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 6 |
Descriptor
Source
Applied Psychological… | 2 |
Journal of Educational and… | 2 |
International Journal of… | 1 |
Journal of Educational… | 1 |
Practical Assessment,… | 1 |
Quality Assurance in… | 1 |
Author
Publication Type
Journal Articles | 7 |
Reports - Evaluative | 7 |
Reports - Research | 6 |
Speeches/Meeting Papers | 3 |
Dissertations/Theses | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Kárász, Judit T.; Széll, Krisztián; Takács, Szabolcs – Quality Assurance in Education: An International Perspective, 2023
Purpose: Based on the general formula, which depends on the length and difficulty of the test, the number of respondents and the number of ability levels, this study aims to provide a closed formula for the adaptive tests with medium difficulty (probability of solution is p = 1/2) to determine the accuracy of the parameters for each item and in…
Descriptors: Test Length, Probability, Comparative Analysis, Difficulty Level
Doble, Christopher; Matayoshi, Jeffrey; Cosyn, Eric; Uzun, Hasan; Karami, Arash – International Journal of Artificial Intelligence in Education, 2019
A large-scale simulation study of the assessment effectiveness of a particular instantiation of knowledge space theory is described. In this study, data from more than 700,000 actual assessments in mathematics using the ALEKS (Assessment and LEarning in Knowledge Spaces) software were used to determine response probabilities for the same number of…
Descriptors: Test Reliability, Adaptive Testing, Mathematics Tests, Computer Assisted Testing
Nydick, Steven W. – Journal of Educational and Behavioral Statistics, 2014
The sequential probability ratio test (SPRT) is a common method for terminating item response theory (IRT)-based adaptive classification tests. To decide whether a classification test should stop, the SPRT compares a simple log-likelihood ratio, based on the classification bound separating two categories, to prespecified critical values. As has…
Descriptors: Probability, Item Response Theory, Models, Classification
Thompson, Nathan A. – Practical Assessment, Research & Evaluation, 2011
Computerized classification testing (CCT) is an approach to designing tests with intelligent algorithms, similar to adaptive testing, but specifically designed for the purpose of classifying examinees into categories such as "pass" and "fail." Like adaptive testing for point estimation of ability, the key component is the…
Descriptors: Adaptive Testing, Computer Assisted Testing, Classification, Probability
Finkelman, Matthew David – Applied Psychological Measurement, 2010
In sequential mastery testing (SMT), assessment via computer is used to classify examinees into one of two mutually exclusive categories. Unlike paper-and-pencil tests, SMT has the capability to use variable-length stopping rules. One approach to shortening variable-length tests is stochastic curtailment, which halts examination if the probability…
Descriptors: Mastery Tests, Computer Assisted Testing, Adaptive Testing, Test Length
Finkelman, Matthew – Journal of Educational and Behavioral Statistics, 2008
Sequential mastery testing (SMT) has been researched as an efficient alternative to paper-and-pencil testing for pass/fail examinations. One popular method for determining when to cease examination in SMT is the truncated sequential probability ratio test (TSPRT). This article introduces the application of stochastic curtailment in SMT to shorten…
Descriptors: Mastery Tests, Sequential Approach, Computer Assisted Testing, Adaptive Testing
Zwick, Rebecca – 1994
The Mantel Haenszel (MH; 1959) approach of Holland and Thayer (1988) is a well-established method for assessing differential item functioning (DIF). The formula for the variance of the MH DIF statistic is based on work by Phillips and Holland (1987) and Robins, Breslow, and Greenland (1986). Recent simulation studies showed that the MH variances…
Descriptors: Adaptive Testing, Evaluation Methods, Item Bias, Measurement Techniques
Patsula, Liane N.; Steffen, Mandred – 1997
One challenge associated with computerized adaptive testing (CAT) is the maintenance of test and item security while allowing for daily testing. An alternative to continually creating new pools containing an independent set of items would be to consider each CAT pool as a sample of items from a larger collection (referred to as a VAT) rather than…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Multiple Choice Tests
Zwick, Rebecca – 1995
This paper describes a study, now in progress, of new methods for representing the sampling variability of Mantel-Haenszel differential item functioning (DIF) results, based on the system for categorizing the severity of DIF that is now in place at the Educational Testing Service. The methods, which involve a Bayesian elaboration of procedures…
Descriptors: Adaptive Testing, Bayesian Statistics, Classification, Computer Assisted Testing
Meijer, Rob R. – Journal of Educational Measurement, 2004
Two new methods have been proposed to determine unexpected sum scores on sub-tests (testlets) both for paper-and-pencil tests and computer adaptive tests. A method based on a conservative bound using the hypergeometric distribution, denoted p, was compared with a method where the probability for each score combination was calculated using a…
Descriptors: Probability, Adaptive Testing, Item Response Theory, Scores
Kalisch, Stanley James, Jr. – 1974
The four purposes of this study were: (1) To compare two versions of a tailored testing model similar to one suggested by Kalisch (1974); (2) To identify levels of the variables within the two versions, which produce an efficient tailored testing procedures; (3) To compare, within each version, the results obtained when employing relatively small…
Descriptors: Ability, Adaptive Testing, Branching, Comparative Analysis
Kalisch, Stanley James, Jr. – 1975
Two tailored testing models, specifying procedures by which the correctness of examinees' responses to a fixed number of test items are predicted by presenting as few items as possible to the examinee, were compared for their efficiency. The models differ in that one requires reconsideration of each prediction whenever additional information is…
Descriptors: Ability, Adaptive Testing, Branching, Comparative Analysis
The Effect of Item Choice on Ability Estimation When Using a Simple Logistic Tailored Testing Model.
Reckase, Mark D. – 1975
This paper explores the effects of item choice on ability estimation when using a tailored testing procedure based on the Rasch simple logistic model. Most studies of the simple logistic model imply that ability estimates are totally independent of the items used, regardless of the testing procedure. This paper shows that the ability estimate is…
Descriptors: Ability, Achievement Tests, Adaptive Testing, Individual Differences
Papa, Frank J.; Schumacker, Randall E. – 1995
Measures of the robustness of disease class-specific diagnostic concepts could play a central role in training programs designed to assure the development of diagnostic competence. In the pilot study, the authors used disease/sign-symptom conditional probability estimates, Monte Carlo procedures, and artificial intelligence (AI) tools to create…
Descriptors: Adaptive Testing, Artificial Intelligence, Classification, Clinical Diagnosis

Samejima, Fumiko – Applied Psychological Measurement, 1977
The accuracy of estimation of the subjects' latent ability maintained by tailoring for each testee the order of item presentation and the border of item dichotomization was compared to the information provided by the original graded test items. (RC)
Descriptors: Ability, Adaptive Testing, Branching, Computer Assisted Testing
Previous Page | Next Page »
Pages: 1 | 2