NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 8 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias
Peer reviewed Peer reviewed
Bennett, Randy Elliot; And Others – Journal of Educational Measurement, 1991
The relationship of multiple-choice and free-response items on the College Board's Advanced Placement Computer Science Examination was studied using confirmatory factor analysis. Results with 2 samples of 1,000 high school students suggested that the most parsimonious fit was achieved using a single factor. Implications for construct validity are…
Descriptors: Chi Square, College Entrance Examinations, Comparative Testing, Computer Science
Peer reviewed Peer reviewed
Skaggs, Gary; Lissitz, Robert W. – Journal of Educational Measurement, 1992
The consistency of several item bias detection methods was studied across different test administrations of the same items using data from a mathematics test given to approximately 6,600 eighth grade students in all. The Mantel Haenszel and item-response-theory-based sum-of-squares methods were the most consistent. (SLD)
Descriptors: Comparative Testing, Grade 8, Item Bias, Item Response Theory
Peer reviewed Peer reviewed
Stricker, Lawrence J. – Journal of Educational Measurement, 1991
To study whether different forms of the Scholastic Aptitude Test (SAT) used since the mid-1970s varied in their correlations with academic performance criteria, 1975 and 1985 forms were administered to 1,554 and 1,753 high school juniors, respectively. The 1975 form did not have greater validity than the 1985 form. (SLD)
Descriptors: Class Rank, College Entrance Examinations, Comparative Testing, Correlation
Peer reviewed Peer reviewed
Bolger, Niall; Kellaghan, Thomas – Journal of Educational Measurement, 1990
Gender differences in scholastic achievement as a function of measurement method were examined by comparing performance of 739 15-year-old boys and 758 15-year-old girls in Irish high schools on multiple-choice and free-response tests of mathematics, Irish, and English achievement. Method-based gender differences are discussed. (SLD)
Descriptors: Academic Achievement, Adolescents, Comparative Testing, English
Peer reviewed Peer reviewed
Bridgeman, Brent – Journal of Educational Measurement, 1992
Examinees in a regular administration of the quantitative portion of the Graduate Record Examination responded to particular items in a machine-scannable multiple-choice format. Volunteers (n=364) used a computer to answer open-ended counterparts of these items. Scores for both formats demonstrated similar correlational patterns. (SLD)
Descriptors: Answer Sheets, College Entrance Examinations, College Students, Comparative Testing
Peer reviewed Peer reviewed
Martinez, Michael E. – Journal of Educational Measurement, 1991
Figural response items (FRIs) in science were administered to 347 fourth graders, 365 eighth graders, and 322 twelfth graders. Item and test statistics from parallel FRIs and multiple-choice questions illustrate FRIs' more difficult and more discriminating nature. Relevance of guessing to FRIs and diagnostic value of the item type are highlighted.…
Descriptors: Comparative Testing, Constructed Response, Elementary School Students, Elementary Secondary Education
Peer reviewed Peer reviewed
Wise, Steven L.; And Others – Journal of Educational Measurement, 1992
Performance of 156 undergraduate and 48 graduate students on a self-adapted test (SFAT)--students choose the difficulty level of their test items--was compared with performance on a computer-adapted test (CAT). Those taking the SFAT obtained higher ability scores and reported lower posttest state anxiety than did CAT takers. (SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Difficulty Level