NotesFAQContact Us
Collection
Advanced
Search Tips
Education Level
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 39 results Save | Export
Frisbie, David A. – 1981
The relative difficulty ratio (RDR) is used as a method of representing test difficulty. The RDR is the ratio of a test mean to the ideal mean, the point midway between the perfect score and the mean chance score for the test. The RDR tranformation is a linear scale conversion method but not a linear equating method in the classical sense. The…
Descriptors: Comparative Testing, Difficulty Level, Evaluation Methods, Raw Scores
Clauser, Brian E.; And Others – 1991
Item bias has been a major concern for test developers during recent years. The Mantel-Haenszel statistic has been among the preferred methods for identifying biased items. The statistic's performance in identifying uniform bias in simulated data modeled by producing various levels of difference in the (item difficulty) b-parameter for reference…
Descriptors: Comparative Testing, Difficulty Level, Item Bias, Item Response Theory
Mazor, Kathleen M.; And Others – 1993
The Mantel-Haenszel (MH) procedure has become one of the most popular procedures for detecting differential item functioning (DIF). One of the most troublesome criticisms of this procedure is that while detection rates for uniform DIF are very good, the procedure is not sensitive to non-uniform DIF. In this study, examinee responses were generated…
Descriptors: Comparative Testing, Computer Simulation, Item Bias, Item Response Theory
Peer reviewed Peer reviewed
Stocking, Martha L.; And Others – Applied Psychological Measurement, 1993
A method of automatically selecting items for inclusion in a test with constraints on item content and statistical properties was applied to real data. Tests constructed manually from the same data and constraints were compared to tests constructed automatically. Results show areas in which automated assembly can improve test construction. (SLD)
Descriptors: Algorithms, Automation, Comparative Testing, Computer Assisted Testing
Ebel, Robert L. – 1981
An alternate-choice test item is a simple declarative sentence, one portion of which is given with two different wordings. For example, "Foundations like Ford and Carnegie tend to be (1) eager (2) hesitant to support innovative solutions to educational problems." The examinee's task is to choose the alternative that makes the sentence…
Descriptors: Comparative Testing, Difficulty Level, Guessing (Tests), Multiple Choice Tests
Brigham, Donald; Sullivan, Edward A. – 1980
The goals of the visual arts program of the Attleboro (MA) public schools, its relationship with the rest of the curriculum, and a study of the effectiveness of the program in seventh grade are described. It is suggested that the visual conceptual skills that are developed through the visual arts program are essential to cognitive processes and…
Descriptors: Cognitive Development, Comparative Testing, Concept Formation, Elementary Secondary Education
Graham, Darol L. – 1974
The adequacy of a test developed for statewide assessment of basic mathematics skills was investigated. The test, comprised of multiple-choice items reflecting a series of behavioral objectives, was compared with a more extensive criterion measure generated from the same objectives by the application of a strict item sampling model. In many…
Descriptors: Comparative Testing, Criterion Referenced Tests, Educational Assessment, Item Sampling
Clauser, Brian E.; And Others – 1991
This paper explores the effectiveness of the Mantel-Haenszel (MH) statistic in detecting differentially functioning test items when the internal criterion is varied. Using a data set from the 1982 statewide administration of a 150-item life skills examination (the New Mexico High School Proficiency Examination), a randomly selected sample of 1,000…
Descriptors: American Indians, Anglo Americans, Comparative Testing, High School Students
Cizek, Gregory J. – 1991
A commonly accepted rule for developing equated examinations using the common-items non-equivalent groups (CINEG) design is that items common to the two examinations being equated should be identical. The CINEG design calls for two groups of examinees to respond to a set of common items that is included in two examinations. In practice, this rule…
Descriptors: Certification, Comparative Testing, Difficulty Level, Higher Education
DeAyala, R. J.; Koch, William R. – 1987
A nominal response model-based computerized adaptive testing procedure (nominal CAT) was implemented using simulated data. Ability estimates from the nominal CAT were compared to those from a CAT based upon the three-parameter logistic model (3PL CAT). Furthermore, estimates from both CAT procedures were compared with the known true abilities used…
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation
Sykes, Robert C.; And Others – 1991
To investigate the psychometric feasibility of replacing a paper-and-pencil licensing examination with a computer-administered test, a validity study was conducted. The computer-administered test (Cadm) was a common set of items for all test takers, distinct from computerized adaptive testing, in which test takers receive items appropriate to…
Descriptors: Adults, Certification, Comparative Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Ellis, Barbara B. – Intelligence, 1990
Intellectual abilities were measured for 217 German and 205 American college students using tests (in the subjects' native languages) in which equivalence was established by an item-response theory-based differential-item-functioning (DIF) analysis. Comparisons between groups were not the same before and after removal of DIF items. (SLD)
Descriptors: College Students, Comparative Testing, Cross Cultural Studies, Culture Fair Tests
De Ayala, R. J. – 1992
One important and promising application of item response theory (IRT) is computerized adaptive testing (CAT). The implementation of a nominal response model-based CAT (NRCAT) was studied. Item pool characteristics for the NRCAT as well as the comparative performance of the NRCAT and a CAT based on the three-parameter logistic (3PL) model were…
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation
Powell, Z. Emily – 1992
Little research exists on the psychological impacts of computerized adaptive testing (CAT) and how it may affect test performance. Three CAT procedures were examined, in which items were selected to match students' achievement levels, from the item pool at random, or according to student choice of item difficulty levels. Twenty-four graduate…
Descriptors: Academic Achievement, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Kim, Haeok; Plake, Barbara S. – 1993
A two-stage testing strategy is one method of adapting the difficulty of a test to an individual's ability level in an effort to achieve more precise measurement. A routing test provides an initial estimate of ability level, and a second-stage measurement test then evaluates the examinee further. The measurement accuracy and efficiency of item…
Descriptors: Ability, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Previous Page | Next Page ยป
Pages: 1  |  2  |  3