NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 256 to 270 of 418 results Save | Export
Peer reviewed Peer reviewed
Barnes, Janet L.; Landy, Frank J. – Applied Psychological Measurement, 1979
Although behaviorally anchored rating scales have both intuitive and empirical appeal, they have not always yielded superior results in contrast with graphic rating scales. Results indicate that the choice of an anchoring procedure will depend on the nature of the actual rating process. (Author/JKS)
Descriptors: Behavior Rating Scales, Comparative Testing, Higher Education, Rating Scales
Peer reviewed Peer reviewed
Schriesheim, Chester A.; And Others – Educational and Psychological Measurement, 1991
Effects of item wording on questionnaire reliability and validity were studied, using 280 undergraduate business students who completed a questionnaire comprising 4 item types: (1) regular; (2) polar opposite; (3) negated polar opposite; and (4) negated regular. Implications of results favoring regular and negated regular items are discussed. (SLD)
Descriptors: Business Education, Comparative Testing, Higher Education, Negative Forms (Language)
Peer reviewed Peer reviewed
Charak, David A.; Stella, Jennifer L. – Assessment for Effective Intervention, 2002
This article provides in-depth information regarding the most commonly used instruments for the screening or diagnosis of autistic spectrum disorders. Reliability, validity, format, and target population are presented to help clinicians select appropriate diagnostic measures. Future directions in the development of new instruments are discussed.…
Descriptors: Adolescents, Adults, Autism, Children
Peer reviewed Peer reviewed
Direct linkDirect link
Xu, Yuejin; Iran-Nejad, Asghar; Thoma, Stephen J. – Journal of Interactive Online Learning, 2007
The purpose of the study was to determine comparability of an online version to the original paper-pencil version of Defining Issues Test 2 (DIT2). This study employed methods from both Classical Test Theory (CTT) and Item Response Theory (IRT). Findings from CTT analyses supported the reliability and discriminant validity of both versions.…
Descriptors: Computer Assisted Testing, Test Format, Comparative Analysis, Test Theory
Henning, Grant – 1991
In order to evaluate the Test of English as a Foreign Language (TOEFL) vocabulary item format and to determine the effectiveness of alternative vocabulary test items, this study investigated the functioning of eight different multiple-choice formats that differed with regard to: (1) length and inference-generating quality of the stem; (2) the…
Descriptors: Adults, Context Effect, Difficulty Level, English (Second Language)
Ohio State Univ., Columbus. Trade and Industrial Education Instructional Materials Lab. – 1978
The Ohio Vocational Achievement Tests are specially designed instruments for use by teachers, supervisors, and administrators to evaluate and diagnose vocational achievement for improving instruction in secondary vocational programs at the 11th and 12th grade levels. This guide explains the Ohio Vocational Achievement Tests and how they are used.…
Descriptors: Academic Achievement, Achievement Tests, High Schools, Scoring Formulas
Melancon, Janet G.; Thompson, Bruce – 1988
Applied classical measurement theory was used to study the measurement characteristics of Forms A and B of the Finding Embedded Figures Test (FEFT) when the test is administered in a "no-guessing" or "supply" format. Data provided by 69 students at a private university in the southern United States were used. Both forms of the…
Descriptors: Comparative Analysis, Difficulty Level, Discriminant Analysis, Guessing (Tests)
Henk, William A. – 1983
The specific performance characteristics of eight alternative cloze test formats were examined at the fourth and sixth grade levels. At each grade, 64 subjects were randomly assigned to one of four basic treatments (every-fifth/standard, every-fifth/cued, total random/standard, and total random/cued) and tested. Responses on each of the cloze…
Descriptors: Cloze Procedure, Comparative Analysis, Grade 4, Grade 6
Milton, Ohmer – 1982
Educators are called upon to improve the quality of classroom tests to enhance the learning of content. Less faculty concern for tests than for other features of instruction, compounded by a lack of knowing how to assess different levels of learning with test questions that measure complex processes, appear to generate poor quality classroom…
Descriptors: Educational Testing, Evaluation Methods, Higher Education, Learning Activities
Kingston, Neal; Turner, Nancy – 1984
This investigation examines the impact the l98l Graduate Record Examination (GRE) General Test Format Revision had on the stability over time of the verbal, quantitative, and analytical scores. Scores were used from the self-selected group of repeaters who took the GRE General Test twice between October 1980 and June 1982. Examinees were divided…
Descriptors: College Entrance Examinations, Graduate Study, Higher Education, Multiple Regression Analysis
Marston, Doug; Deno, Stanley – 1981
The reliability of four measures of written expression was examined (total words written, mature words, words spelled correctly, and letters in sequence). Subjects included elementary-age students in several school districts, some of whom were learning disabled. Results revealed high coefficients for test-retest reliability, parallel-form…
Descriptors: Classroom Techniques, Comparative Analysis, Elementary Education, Formative Evaluation
Peer reviewed Peer reviewed
Knight, Deborah Forsyth – Journal of Reading, 1985
Reviews the Curriculum Referenced Tests of Mastery that are intended to measure achievement, with the emphasis on measuring what a student has learned rather than predicting future success in school, concluding that the tests are worthy of consideration by any district. (HOD)
Descriptors: Academic Achievement, Educational Objectives, Language Arts, Mathematics Instruction
Peer reviewed Peer reviewed
Aiken, Lewis R. – Educational and Psychological Measurement, 1989
Two alternatives to traditional item analysis and reliability estimation procedures are considered for determining the difficulty, discrimination, and reliability of optional items on essay and other tests. A computer program to compute these measures is described, and illustrations are given. (SLD)
Descriptors: College Entrance Examinations, Computer Software, Difficulty Level, Essay Tests
Peer reviewed Peer reviewed
Frisbie, David A. – Educational Measurement: Issues and Practice, 1992
Literature related to the multiple true-false (MTF) item format is reviewed. Each answer cluster of a MTF item may have several true items and the correctness of each is judged independently. MTF tests appear efficient and reliable, although they are a bit harder than multiple choice items for examinees. (SLD)
Descriptors: Achievement Tests, Difficulty Level, Literature Reviews, Multiple Choice Tests
Peer reviewed Peer reviewed
Trevisan, Michael S.; And Others – Educational and Psychological Measurement, 1994
The reliabilities of 2-, 3-, 4-, and 5-choice tests were compared through an incremental-option model on a test taken by 154 high school seniors. Creating the test forms incrementally more closely approximates actual test construction. The nonsignificant differences among the option choices support the three-option item. (SLD)
Descriptors: Distractors (Tests), Estimation (Mathematics), High School Students, High Schools
Pages: 1  |  ...  |  14  |  15  |  16  |  17  |  18  |  19  |  20  |  21  |  22  |  ...  |  28