NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 5 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Bengs, Daniel; Kroehne, Ulf; Brefeld, Ulf – Journal of Educational Measurement, 2021
By tailoring test forms to the test-taker's proficiency, Computerized Adaptive Testing (CAT) enables substantial increases in testing efficiency over fixed forms testing. When used for formative assessment, the alignment of task difficulty with proficiency increases the chance that teachers can derive useful feedback from assessment data. The…
Descriptors: Computer Assisted Testing, Formative Evaluation, Group Testing, Program Effectiveness
Peer reviewed Peer reviewed
Direct linkDirect link
Clauser, Brian E.; Mee, Janet; Baldwin, Su G.; Margolis, Melissa J.; Dillon, Gerard F. – Journal of Educational Measurement, 2009
Although the Angoff procedure is among the most widely used standard setting procedures for tests comprising multiple-choice items, research has shown that subject matter experts have considerable difficulty accurately making the required judgments in the absence of examinee performance data. Some authors have viewed the need to provide…
Descriptors: Standard Setting (Scoring), Program Effectiveness, Expertise, Health Personnel
Peer reviewed Peer reviewed
Marco, Gary L.; And Others – Journal of Educational Measurement, 1976
Special emphasis is given to the kinds of control that can be exercised over initial status, including the use of proxy input data. A rationale for the classification scheme is developed, based on (1) three one-shot, one cross-sectional, and two longitudinal data types and (2) two types of referencing: criterion referencing and norm referencing.…
Descriptors: Classification, Data Collection, Evaluation Methods, Methods
Peer reviewed Peer reviewed
Schmidt, William H. – Journal of Educational Measurement, 1983
A conception of invalidity as bias is related to content validity for standardized achievement tests. A method of estimating content bias for each of three content domains (a priori, curricular, and instructional) based on the specification of a content taxonomy is also proposed. (Author/CM)
Descriptors: Achievement Tests, Content Analysis, Evaluation Methods, Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Finch, Holmes; Habing, Brian – Journal of Educational Measurement, 2005
This study examines the performance of a new method for assessing and characterizing dimensionality in test data using the NOHARM model, and comparing it with DETECT. Dimensionality assessment is carried out using two goodness-of-fit statistics that are compared to reference X[2] distributions. A Monte Carlo study is used with item parameters…
Descriptors: Program Effectiveness, Monte Carlo Methods, Item Response Theory, Comparative Analysis