NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 42 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kentaro Fukushima; Nao Uchida; Kensuke Okada – Journal of Educational and Behavioral Statistics, 2025
Diagnostic tests are typically administered in a multiple-choice (MC) format due to their advantages of objectivity and time efficiency. The MC-deterministic input, noisy "and" gate (DINA) family of models, a representative class of cognitive diagnostic models for MC items, efficiently and parsimoniously estimates the mastery profiles of…
Descriptors: Diagnostic Tests, Cognitive Measurement, Multiple Choice Tests, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Kylie Gorney; Sandip Sinharay – Educational and Psychological Measurement, 2025
Test-takers, policymakers, teachers, and institutions are increasingly demanding that testing programs provide more detailed feedback regarding test performance. As a result, there has been a growing interest in the reporting of subscores that potentially provide such detailed feedback. Haberman developed a method based on classical test theory…
Descriptors: Scores, Test Theory, Test Items, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Sivaci, Seda – Journal on English Language Teaching, 2020
The purpose of the present study is to evaluate proficiency tests of the universities in Turkey in line with the ALTE Quality Assurance Checklists in terms of four dimensions: a. Test construction, b. Administration & Logistics c. Grading, Marking Results and d. Test Analysis & Post-examination Review. The study took place in four…
Descriptors: Foreign Countries, Language Tests, Language Proficiency, Educational Quality
Peer reviewed Peer reviewed
Direct linkDirect link
Soland, James; Kuhfeld, Megan – Educational Assessment, 2019
Considerable research has examined the use of rapid guessing measures to identify disengaged item responses. However, little is known about students who rapidly guess over the course of several tests. In this study, we use achievement test data from six administrations over three years to investigate whether rapid guessing is a stable trait-like…
Descriptors: Testing, Guessing (Tests), Reaction Time, Achievement Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Bailey, Janelle M.; Johnson, Bruce; Prather, Edward E.; Slater, Timothy F. – International Journal of Science Education, 2012
Concept inventories (CIs)--typically multiple-choice instruments that focus on a single or small subset of closely related topics--have been used in science education for more than a decade. This paper describes the development and validation of a new CI for astronomy, the "Star Properties Concept Inventory" (SPCI). Questions cover the areas of…
Descriptors: Educational Strategies, Validity, Testing, Astronomy
Peer reviewed Peer reviewed
Direct linkDirect link
Royal, Kenneth D.; Gilliland, Kurt O.; Kernick, Edward T. – Anatomical Sciences Education, 2014
Any examination that involves moderate to high stakes implications for examinees should be psychometrically sound and legally defensible. Currently, there are two broad and competing families of test theories that are used to score examination data. The majority of instructors outside the high-stakes testing arena rely on classical test theory…
Descriptors: Item Response Theory, Scoring, Evaluation Methods, Anatomy
Peer reviewed Peer reviewed
Direct linkDirect link
Daly, Anthony L.; Baird, Jo-Anne; Chamberlain, Suzanne; Meadows, Michelle – Curriculum Journal, 2012
This paper describes an exploration into a reform of the A-level qualification in England in 2008; namely, the introduction of the "stretch and challenge" policy. This policy was initiated by the exams regulator and determined that exam papers should be redesigned to encourage the application of higher order thinking skills, both in the…
Descriptors: Test Preparation, Student Evaluation, Student Attitudes, Educational Change
Peer reviewed Peer reviewed
Direct linkDirect link
Mislevy, Robert J.; Haertel, Geneva; Cheng, Britte H.; Ructtinger, Liliana; DeBarger, Angela; Murray, Elizabeth; Rose, David; Gravel, Jenna; Colker, Alexis M.; Rutstein, Daisy; Vendlinski, Terry – Educational Research and Evaluation, 2013
Standardizing aspects of assessments has long been recognized as a tactic to help make evaluations of examinees fair. It reduces variation in irrelevant aspects of testing procedures that could advantage some examinees and disadvantage others. However, recent attention to making assessment accessible to a more diverse population of students…
Descriptors: Testing Accommodations, Access to Education, Testing, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Darrah, Marjorie; Fuller, Edgar; Miller, David – Journal of Computers in Mathematics and Science Teaching, 2010
This paper discusses a possible solution to a problem frequently encountered by educators seeking to use computer-based or multiple choice-based exams for mathematics. These assessment methodologies force a discrete grading system on students and do not allow for the possibility of partial credit. The research presented in this paper investigates…
Descriptors: College Students, College Mathematics, Calculus, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Vannest, Kimberly J.; Parker, Richard I.; Davis, John L.; Soares, Denise A.; Smith, Stacey L. – Behavioral Disorders, 2012
More and more, schools are considering the use of progress monitoring data for high-stakes decisions such as special education eligibility, program changes to more restrictive environments, and major changes in educational goals. Those high-stakes types of data-based decisions will need methodological defensibility. Current practice for…
Descriptors: Decision Making, Educational Change, Regression (Statistics), Field Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Brooks, Lindsay – Language Testing, 2009
This study, framed within sociocultural theory, examines the interaction of adult ESL test-takers in two tests of oral proficiency: one in which they interacted with an examiner (the individual format) and one in which they interacted with another student (the paired format). The data for the eight pairs in this study were drawn from a larger…
Descriptors: Testing, Rating Scales, Program Effectiveness, Interaction
Braun, Henry I.; Mislevy, Robert J. – US Department of Education, 2004
Psychologist Andrea diSessa coined the term "phenomenological primitives", or p-prims, to talk about nonexperts' reasoning about physical situations. P-prims are primitive in the sense that they stand without significant explanatory substructure or explanation. Examples are "Heavy objects fall faster than light objects" and "Continuing force is…
Descriptors: Test Theory, Testing, Evaluation Methods, Scores
Peer reviewed Peer reviewed
Wilcox, Rand R. – Educational and Psychological Measurement, 1983
This article provides unbiased estimates of the proportion of items in an item domain that an examinee would answer correctly if every item were attempted, when a closed sequential testing procedure is used. (Author)
Descriptors: Estimation (Mathematics), Psychometrics, Scores, Sequential Approach
Peer reviewed Peer reviewed
Woodruff, David – Journal of Educational Statistics, 1986
The purpose of the present paper is to derive linear equating methods for the common item nonequivalent populations design from explicitly stated congeneric type test score models. The equating methods developed are compared with previously developed methods and applied to five professionally constructed examinations administered to approximately…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Scores
Burton, Robert S. – New Directions for Testing and Measurement, 1980
Although Model A, the only norm-referenced evaluation procedure in the Title I Evaluation and Reporting System, requires no data other than the test scores themselves, it introduces two sources of bias and involved three test administrations. Roberts' two-test procedure offers the advantages of less bias and less testing. (RL)
Descriptors: Comparative Analysis, Mathematical Formulas, Scores, Statistical Bias
Previous Page | Next Page ยป
Pages: 1  |  2  |  3