NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 10 results Save | Export
Benton, Tom – Research Matters, 2020
This article reviews the evidence on the extent to which experts' perceptions of item difficulties, captured using comparative judgement, can predict empirical item difficulties. This evidence is drawn from existing published studies on this topic and also from statistical analysis of data held by Cambridge Assessment. Having reviewed the…
Descriptors: Test Items, Difficulty Level, Expertise, Comparative Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
David Bell; Vikki O'Neill; Vivienne Crawford – Practitioner Research in Higher Education, 2023
We compared the influence of open-book extended duration versus closed book time-limited format on reliability and validity of written assessments of pharmacology learning outcomes within our medical and dental courses. Our dental cohort undertake a mid-year test (30xfree-response short answer to a question, SAQ) and end-of-year paper (4xSAQ,…
Descriptors: Undergraduate Students, Pharmacology, Pharmaceutical Education, Test Format
Neitzel, Jennifer; Early, Diane; Sideris, John; LaForrett, Doré; Abel, Michael B.; Soli, Margaret; Davidson, Dawn L.; Haboush-Deloye, Amanda; Hestenes, Linda L.; Jenson, Denise; Johnson, Cindy; Kalas, Jennifer; Mamrak, Angela; Masterson, Marie L.; Mims, Sharon U.; Oya, Patti; Philson, Bobbi; Showalter, Megan; Warner-Richter, Mallory; Kortright Wood, Jill – Journal of Early Childhood Research, 2019
The Early Childhood Environment Rating Scales, including the "Early Childhood Environment Rating Scale--Revised" (Harms et al., 2005) and the "Early Childhood Environment Rating Scale, Third Edition" (Harms et al., 2015) are the most widely used observational assessments in early childhood learning environments. The most recent…
Descriptors: Rating Scales, Early Childhood Education, Educational Quality, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Qian, David D. – Language Assessment Quarterly, 2008
In the last 15 years or so, language testing practitioners have increasingly favored assessing vocabulary in context. The discrete-point vocabulary measure used in the old version of the Test of English as a Foreign Language (TOEFL) has long been criticized for encouraging test candidates to memorize wordlists out of context although test items…
Descriptors: Predictive Validity, Context Effect, Vocabulary, English (Second Language)
Kirisci, Levent; Hsu, Tse-Chi – 1988
The predictive analysis approach to adaptive testing originated in the idea of statistical predictive analysis suggested by J. Aitchison and I.R. Dunsmore (1975). The adaptive testing model proposed is based on parameter-free predictive distribution. Aitchison and Dunsmore define statistical prediction analysis as the use of data obtained from an…
Descriptors: Adaptive Testing, Bayesian Statistics, Comparative Analysis, Item Analysis
Peer reviewed Peer reviewed
Boldt, Robert F. – Language Testing, 1992
The assumption called PIRC (proportional item response curve) was tested in which PIRC was used to predict item scores of selected examinees on selected items. Findings show approximate accuracies of prediction for PIRC, the three-parameter logist model, and a modified Rasch model. (12 references) (Author/LB)
Descriptors: Comparative Analysis, English (Second Language), Factor Analysis, Item Response Theory
Peer reviewed Peer reviewed
DeMauro, G. – Language Testing, 1992
Several analyses are presented on the relationships among the Test of Spoken English, Test of Written English, and Test of English as a Foreign Language. The multivariate prediction of each test from the scores on the others is very accurate; variances with two prominent factors may relate to specific cognitive test-taking skills. (eight…
Descriptors: Comparative Analysis, Language Research, Language Skills, Language Tests
Downey, Ronald G.
Previous research has studied the effects of different methods of item option weighting on the reliability and concurrent and predictive validity of achievement tests. Increases in reliability are generally found, but with mixed results for validity. Several methods of producing option weights, (i.e., Guttman internal and external weights and…
Descriptors: Achievement Tests, Comparative Analysis, Correlation, Grade Point Average
Owen, K. – 1989
This study was undertaken to shed light on problems concerning the construction and use of a common test battery for various South African population groups. Specifically, the psychometric suitability for use with Indian and Black pupils, of the Junior Aptitude Tests (JAT), which were standardized for White pupils in Standard 7 was studied. The…
Descriptors: Aptitude Tests, Black Students, Comparative Analysis, Construct Validity
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Benderson, Albert, Ed. – Focus, 1988
The scores of handicapped students taking tests such as the Scholastic Aptitude Test (SAT) or the Graduate Record Examinations are flagged so that admissions officers will be aware that they were achieved under special circumstances. A series of studies was initiated to determine whether special administrations of such tests are comparable to…
Descriptors: Admission Criteria, College Admission, College Entrance Examinations, College Students