NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 18 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kluge, Annette – Applied Psychological Measurement, 2008
The use of microworlds (MWs), or complex dynamic systems, in educational testing and personnel selection is hampered by systematic measurement errors because these new and innovative item formats are not adequately controlled for their difficulty. This empirical study introduces a way to operationalize an MW's difficulty and demonstrates the…
Descriptors: Personnel Selection, Self Efficacy, Educational Testing, Computer Uses in Education
Peer reviewed Peer reviewed
Stocking, Martha L.; And Others – Applied Psychological Measurement, 1993
A method of automatically selecting items for inclusion in a test with constraints on item content and statistical properties was applied to real data. Tests constructed manually from the same data and constraints were compared to tests constructed automatically. Results show areas in which automated assembly can improve test construction. (SLD)
Descriptors: Algorithms, Automation, Comparative Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Drasgow, Fritz; And Others – Applied Psychological Measurement, 1991
Extensions of unidimensional appropriateness indices are developed for multiunidimensional tests (multidimensional tests composed of unidimensional subtests). Simulated and real data (scores of 2,978 students on the Armed Services Vocational Aptitude Battery) were used to evaluate the indices' effectiveness in determining individuals who are…
Descriptors: Comparative Testing, Computer Simulation, Equations (Mathematics), Graphs
Peer reviewed Peer reviewed
Betz, Nancy E. – Applied Psychological Measurement, 1977
The effects of providing immediate feedback on an item-by-item basis to testees were investigated on several samples of undergraduates who were administered either adaptive or conventional tests. Results generally indicated that knowledge of results improved test performance. (JKS)
Descriptors: Academic Aptitude, Adaptive Testing, Comparative Testing, Feedback
Peer reviewed Peer reviewed
Henly, Susan J.; And Others – Applied Psychological Measurement, 1989
A group of covariance structure models was examined to ascertain the similarity between conventionally administered and computerized adaptive versions of the Differential Aptitude Test (DAT). Results for 332 students indicate that the computerized version of the DAT is an adequate representation of the conventional test battery. (TJH)
Descriptors: Ability Identification, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J. – Applied Psychological Measurement, 2006
Two local methods for observed-score equating are applied to the problem of equating an adaptive test to a linear test. In an empirical study, the methods were evaluated against a method based on the test characteristic function (TCF) of the linear test and traditional equipercentile equating applied to the ability estimates on the adaptive test…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Format, Equated Scores
Peer reviewed Peer reviewed
Reilly, Richard R.; Echternacht, Gary J. – Applied Psychological Measurement, 1979
Occupational interest scales and homogeneous scales were developed and compared on a sample of 3,072 Air Force personnel in a double cross-validation design. Results indicated that the occupational scales were slightly superior in correctly classifying satisified personnel v people in general. (Author/JKS)
Descriptors: Attitude Measures, Classification, Comparative Testing, Interest Inventories
Peer reviewed Peer reviewed
Lunz, Mary E.; And Others – Applied Psychological Measurement, 1992
The effects of reviewing items and altering responses on the efficiency of computerized adaptive tests and resultant ability estimates of the examinees were explored for medical technology students (220 students could and 492 students could not review and alter their responses). Data do not support disallowing review. (SLD)
Descriptors: Ability, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Norcini, John; And Others – Applied Psychological Measurement, 1991
Effects of numbers of experts (NOEs) and common items (CIs) on the scaling of cutting scores from expert judgments were studied for 11,917 physicians taking 2 forms of a medical specialty examination. Increasing NOEs and CIs reduced error; beyond 5 experts and 25 CIs, error differences were small. (SLD)
Descriptors: Comparative Testing, Cutting Scores, Equated Scores, Estimation (Mathematics)
Peer reviewed Peer reviewed
Bennett, Randy Elliot; And Others – Applied Psychological Measurement, 1990
The relationship of an expert-system-scored constrained free-response item type to multiple-choice and free-response items was studied using data for 614 students on the College Board's Advanced Placement Computer Science (APCS) Examination. Implications for testing and the APCS test are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Computer Assisted Testing, Computer Science
Peer reviewed Peer reviewed
Traub, Ross E.; Fisher, Charles W. – Applied Psychological Measurement, 1977
Two sets of mathematical reasoning and two sets of verbal comprehension items were cast into each of three formats--constructed response, standard multiple-choice, and Coombs multiple-choice--in order to assess whether tests with identical content but different formats measure the same attribute. (Author/CTM)
Descriptors: Comparative Testing, Confidence Testing, Constructed Response, Factor Analysis
Peer reviewed Peer reviewed
Barnes, Janet L.; Landy, Frank J. – Applied Psychological Measurement, 1979
Although behaviorally anchored rating scales have both intuitive and empirical appeal, they have not always yielded superior results in contrast with graphic rating scales. Results indicate that the choice of an anchoring procedure will depend on the nature of the actual rating process. (Author/JKS)
Descriptors: Behavior Rating Scales, Comparative Testing, Higher Education, Rating Scales
Peer reviewed Peer reviewed
De Ayala, R. J. – Applied Psychological Measurement, 1992
A computerized adaptive test (CAT) based on the nominal response model (NR CAT) was implemented, and the performance of the NR CAT and a CAT based on the three-parameter logistic model was compared. The NR CAT produced trait estimates comparable to those of the three-parameter test. (SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Equations (Mathematics)
Peer reviewed Peer reviewed
Kim, Seock-Ho; Cohen, Allan S. – Applied Psychological Measurement, 1991
The exact and closed-interval area measures for detecting differential item functioning are compared for actual data from 1,000 African-American and 1,000 white college students taking a vocabulary test with items intentionally constructed to favor 1 set of examinees. No real differences in detection of biased items were found. (SLD)
Descriptors: Black Students, College Students, Comparative Testing, Equations (Mathematics)
Peer reviewed Peer reviewed
Birenbaum, Menucha; And Others – Applied Psychological Measurement, 1992
The effect of multiple-choice (MC) or open-ended (OE) response format on diagnostic assessment of algebra test performance was investigated with 231 eighth and ninth graders in Tel Aviv (Israel) using bug or rule space analysis. Both analyses indicated closer similarity between parallel OE subsets than between stem-equivalent OE and MC subsets.…
Descriptors: Algebra, Comparative Testing, Educational Assessment, Educational Diagnosis
Previous Page | Next Page ยป
Pages: 1  |  2