Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 7 |
| Since 2017 (last 10 years) | 11 |
| Since 2007 (last 20 years) | 32 |
Descriptor
Source
Author
| Badger, Elizabeth | 4 |
| Melancon, Janet G. | 3 |
| Thomas, Brenda | 3 |
| Thompson, Bruce | 3 |
| Anderson, Paul S. | 2 |
| Colliver, Jerry A. | 2 |
| Enger, John M. | 2 |
| Huntley, Renee M. | 2 |
| Lissitz, Robert W. | 2 |
| Lunz, Mary E. | 2 |
| Sykes, Robert C. | 2 |
| More ▼ | |
Publication Type
Education Level
| Higher Education | 17 |
| Postsecondary Education | 9 |
| Elementary Secondary Education | 6 |
| Elementary Education | 4 |
| High Schools | 4 |
| Secondary Education | 3 |
| Adult Education | 1 |
| Grade 3 | 1 |
| Grade 4 | 1 |
| Grade 5 | 1 |
| Grade 6 | 1 |
| More ▼ | |
Audience
| Researchers | 11 |
| Practitioners | 1 |
| Teachers | 1 |
Location
| Canada | 3 |
| United Kingdom | 3 |
| China | 2 |
| Czech Republic | 2 |
| Ireland | 2 |
| United States | 2 |
| California | 1 |
| India | 1 |
| Israel | 1 |
| Israel (Tel Aviv) | 1 |
| Jamaica | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedCrehan, Kevin D.; And Others – Educational and Psychological Measurement, 1993
Studies with 220 college students found that multiple-choice test items with 3 items are more difficult than those with 4 items, and items with the none-of-these option are more difficult than those without this option. Neither format manipulation affected item discrimination. Implications for test construction are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Difficulty Level, Distractors (Tests)
Rodriguez-Aragon, Graciela; And Others – 1993
The predictive power of the Split-Half version of the Wechsler Intelligence Scale for Children--Revised (WISC-R) Object Assembly (OA) subtest was compared to that of the full administration of the OA subtest. A cohort of 218 male and 49 female adolescent offenders detained in a Texas juvenile detention facility between 1990 and 1992 was used. The…
Descriptors: Adolescents, Cohort Analysis, Comparative Testing, Correlation
Sykes, Robert C.; And Others – 1992
A part-form methodology was used to study the effect of varying degrees of multidimensionality on the consistency of pass/fail classification decisions obtained from simulated unidimensional item response theory (IRT) based licensure examinations. A control on the degree of form multidimensionality permitted an assessment throughout the range of…
Descriptors: Classification, Comparative Testing, Computer Simulation, Decision Making
Byrne, Barbara M.; Baron, Pierre – 1991
The aims of the present study were threefold: (1) to test for the equivalency of an hierarchical three-factor structure of the Beck Depression Inventory (BDI) across English and French versions for non-clinical adolescents; (2) given evidence of poor model fit, to validate the factorial structure of the BDI French version across three independent…
Descriptors: Adolescents, Comparative Testing, English, Factor Structure
Cizek, Gregory J. – 1991
A commonly accepted rule for developing equated examinations using the common-items non-equivalent groups (CINEG) design is that items common to the two examinations being equated should be identical. The CINEG design calls for two groups of examinees to respond to a set of common items that is included in two examinations. In practice, this rule…
Descriptors: Certification, Comparative Testing, Difficulty Level, Higher Education
Peer reviewedBarnes, Janet L.; Landy, Frank J. – Applied Psychological Measurement, 1979
Although behaviorally anchored rating scales have both intuitive and empirical appeal, they have not always yielded superior results in contrast with graphic rating scales. Results indicate that the choice of an anchoring procedure will depend on the nature of the actual rating process. (Author/JKS)
Descriptors: Behavior Rating Scales, Comparative Testing, Higher Education, Rating Scales
Peer reviewedSchriesheim, Chester A.; And Others – Educational and Psychological Measurement, 1991
Effects of item wording on questionnaire reliability and validity were studied, using 280 undergraduate business students who completed a questionnaire comprising 4 item types: (1) regular; (2) polar opposite; (3) negated polar opposite; and (4) negated regular. Implications of results favoring regular and negated regular items are discussed. (SLD)
Descriptors: Business Education, Comparative Testing, Higher Education, Negative Forms (Language)
Assessing the Effects of Computer Administration on Scores and Parameter Estimates Using IRT Models.
Sykes, Robert C.; And Others – 1991
To investigate the psychometric feasibility of replacing a paper-and-pencil licensing examination with a computer-administered test, a validity study was conducted. The computer-administered test (Cadm) was a common set of items for all test takers, distinct from computerized adaptive testing, in which test takers receive items appropriate to…
Descriptors: Adults, Certification, Comparative Testing, Computer Assisted Testing
Knapp, Deirdre J.; Pliske, Rebecca M. – 1986
A study was conducted to validate the Army's Computerized Adaptive Screening Test (CAST), using data from 2,240 applicants from 60 army recruiting stations across the nation. CAST is a computer-assisted adaptive test used to predict performance on the Armed Forces Qualification Test (AFQT). AFQT scores are computed by adding four subtest scores of…
Descriptors: Adaptive Testing, Adults, Aptitude Tests, Comparative Testing
Heller, Eric S.; Rife, Frank N. – 1987
The goal of this study was to assess the relative merit of various ranges and types of response scales in terms of respondent satisfaction and comfort and the nature of the elicited information in a population of seventh grade students. Three versions of an attitudinal questionnaire, each containing the same items but employing a different…
Descriptors: Attitude Measures, Comparative Testing, Grade 7, Junior High Schools
Peer reviewedBergstrom, Betty A.; Lunz, Mary E. – Evaluation and the Health Professions, 1992
The level of confidence in pass/fail decisions obtained with computerized adaptive tests and paper-and-pencil tests was greater for 645 medical technology students when the computer adaptive test implemented a 90 percent confidence stopping rule than for paper-and-pencil tests of comparable length. (SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Confidence Testing
Peer reviewedTzeng, Oliver C. S.; And Others – Educational and Psychological Measurement, 1991
Measurement properties of two response formats (bipolar and unipolar ratings) in personality assessment were compared using data from 135 college students taking the Myers-Briggs Type Indicator (MBTI). Factorial validity and construct validity of the MBTI were supported. Reasons why the bipolar method is preferable are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Construct Validity, Factor Analysis
Peer reviewedShea, Judy A.; And Others – Evaluation and the Health Professions, 1992
Video and print formats of cardiovascular motion studies were compared for use as assessment measures of interpretive skills for 392 doctors taking a cardiovascular disease certification test. Although video studies were easier to interpret, the equivalence of both motion studies supports use of the print format in national examinations. (SLD)
Descriptors: Cardiovascular System, Comparative Testing, Graduate Medical Education, Interpretive Skills
Peer reviewedBenedict, Ralph H. B.; And Others – Psychological Assessment, 1992
The concurrent validities of 3 short forms of the Wechsler Adult Intelligence Scale (WAIS) were compared for their prediction of full-scale IQ for 145 male and 159 female psychiatric inpatients. Results support previous research showing better predictive accuracy for L. C. Ward's (1990) seven-subtest short form than the others. (SLD)
Descriptors: Adults, Comparative Testing, Concurrent Validity, Cost Effectiveness
Peer reviewedFriedman, Stephen J.; Ansley, Timothy N. – Journal of Experimental Education, 1990
To investigate the relationship between reading and listening test scores, 3 different sets of listening items accompanied by answer sheets requiring varying amounts of reading were administered to 1,200 students in grades 3 through 8. Listening scores increased as more printed information was added to the answer sheet. (SLD)
Descriptors: Answer Sheets, Comparative Testing, Elementary Education, Elementary School Students


