Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 1 |
Descriptor
Comparative Analysis | 7 |
Higher Education | 7 |
College Students | 3 |
Verbal Ability | 3 |
Computer Assisted Testing | 2 |
Graduate Study | 2 |
Item Response Theory | 2 |
Predictive Measurement | 2 |
Test Items | 2 |
Test Use | 2 |
Ability | 1 |
More ▼ |
Source
Applied Measurement in… | 7 |
Author
Bridgeman, Brent | 1 |
Chang, Lei | 1 |
Clauser, Brian E. | 1 |
Clyman, Stephen G. | 1 |
Cohen, Allan S. | 1 |
Kim, Seock-Ho | 1 |
Lawless, Rene | 1 |
Linn, Robert L. | 1 |
Martinez, Michael E. | 1 |
Oliveri, Maria Elena | 1 |
Robin, Frederic | 1 |
More ▼ |
Publication Type
Journal Articles | 7 |
Reports - Research | 5 |
Reports - Evaluative | 2 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 1 |
What Works Clearinghouse Rating
Oliveri, Maria Elena; Lawless, Rene; Robin, Frederic; Bridgeman, Brent – Applied Measurement in Education, 2018
We analyzed a pool of items from an admissions test for differential item functioning (DIF) for groups based on age, socioeconomic status, citizenship, or English language status using Mantel-Haenszel and item response theory. DIF items were systematically examined to identify its possible sources by item type, content, and wording. DIF was…
Descriptors: Test Bias, Comparative Analysis, Item Banks, Item Response Theory

Chang, Lei – Applied Measurement in Education, 1999
Compared the Nedelsky (L. Nedelsky, 1954) and Angoff (W. Angoff, 1971) standard-setting methods in three studies involving 80 graduate students as judges. Nedelsky cutscores were significantly lower than Angoff cutscores. Suggests that combining the strong features of both methods would make a better standard-setting procedure. (SLD)
Descriptors: Comparative Analysis, Cutting Scores, Graduate Students, Graduate Study

Clauser, Brian E.; Swanson, David B.; Clyman, Stephen G. – Applied Measurement in Education, 1999
Performed generalizability analyses of expert ratings and computer-produced scores for a computer-delivered performance assessment of physicians' patient management skills. The two automated scoring systems produced scores for the 200 medical students that were approximately as generalizable as those produced by the four expert raters. (SLD)
Descriptors: Comparative Analysis, Computer Assisted Testing, Generalizability Theory, Higher Education

Kim, Seock-Ho; Cohen, Allan S. – Applied Measurement in Education, 1995
Three procedures for the detection of differential item functioning under item response theory were compared. Data for 2 forms of a mathematics test taken by 1,490 college students were analyzed through F. M. Lord's chi-square, N. S. Raju's area measures, and the likelihood ratio test. (SLD)
Descriptors: Chi Square, College Students, Comparative Analysis, Higher Education

Linn, Robert L. – Applied Measurement in Education, 1993
The following ways of linking results from distinct assessments to use them for multiple purposes are reviewed: (1) equating; (2) calibration; (3) statistical moderation; (4) prediction; and (5) social moderation. The characteristics of these methods, their requirements, and the comparative inferences they support are described. (SLD)
Descriptors: Comparative Analysis, Educational Assessment, Elementary Secondary Education, Equated Scores

Vispoel, Walter P.; And Others – Applied Measurement in Education, 1994
Vocabulary fixed-item (FIT), computerized-adaptive (CAT), and self-adapted (SAT) tests were compared with 121 college students. CAT was more precise and efficient than SAT, which was more precise and efficient than FIT. SAT also yielded higher ability estimates for individuals with lower verbal self-concepts. (SLD)
Descriptors: Ability, Adaptive Testing, College Students, Comparative Analysis

Martinez, Michael E. – Applied Measurement in Education, 1993
Figural response (FR) items in architecture were compared with multiple-choice (MC) counterparts for their ability to predict architectural problem-solving proficiency of 33 practicing architects, 34 architecture interns, and 53 architecture students. Although both FR and MC predicted verbal design problem solving, only FR scores predicted…
Descriptors: Architects, Architectural Drafting, College Students, Comparative Analysis