Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 9 |
Descriptor
Difficulty Level | 14 |
Psychometrics | 14 |
Test Items | 11 |
Models | 7 |
Evaluation Methods | 5 |
Item Response Theory | 5 |
Test Construction | 5 |
Item Analysis | 4 |
Multiple Choice Tests | 4 |
Goodness of Fit | 3 |
Adaptive Testing | 2 |
More ▼ |
Source
Author
Revuelta, Javier | 2 |
Alonzo, Julie | 1 |
Andrews, Glenda | 1 |
Bechger, Timo M. | 1 |
Birney, Damian P. | 1 |
Carolyn Maxwell | 1 |
Conejo, Ricardo | 1 |
Dahl, Laura S. | 1 |
De Champlain, Andre | 1 |
Embretson, Susan E. | 1 |
Emilio | 1 |
More ▼ |
Publication Type
Reports - Descriptive | 14 |
Journal Articles | 12 |
Numerical/Quantitative Data | 1 |
Opinion Papers | 1 |
Speeches/Meeting Papers | 1 |
Tests/Questionnaires | 1 |
Education Level
Higher Education | 3 |
Postsecondary Education | 2 |
Elementary Education | 1 |
Grade 1 | 1 |
Grade 2 | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
Grade 5 | 1 |
Kindergarten | 1 |
Audience
Teachers | 1 |
Location
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 1 |
What Works Clearinghouse Rating
Stephen Humphry; Paul Montuoro; Carolyn Maxwell – Journal of Psychoeducational Assessment, 2024
This article builds upon a proiminent definition of construct validity that focuses on variation in attributes causing variation in measurement outcomes. This article synthesizes the defintion and uses Rasch measurement modeling to explicate a modified conceptualization of construct validity for assessments of developmental attributes. If…
Descriptors: Construct Validity, Measurement Techniques, Developmental Stages, Item Analysis
Dahl, Laura S.; Staples, B. Ashley; Mayhew, Matthew J.; Rockenbach, Alyssa N. – Innovative Higher Education, 2023
Surveys with rating scales are often used in higher education research to measure student learning and development, yet testing and reporting on the longitudinal psychometric properties of these instruments is rare. Rasch techniques allow scholars to map item difficulty and individual aptitude on the same linear, continuous scale to compare…
Descriptors: Surveys, Rating Scales, Higher Education, Educational Research
Towns, Marcy H. – Journal of Chemical Education, 2014
Chemistry faculty members are highly skilled in obtaining, analyzing, and interpreting physical measurements, but often they are less skilled in measuring student learning. This work provides guidance for chemistry faculty from the research literature on multiple-choice item development in chemistry. Areas covered include content, stem, and…
Descriptors: Multiple Choice Tests, Test Construction, Psychometrics, Test Items
Kahraman, Nilufer; De Champlain, Andre; Raymond, Mark – Applied Measurement in Education, 2012
Item-level information, such as difficulty and discrimination are invaluable to the test assembly, equating, and scoring practices. Estimating these parameters within the context of large-scale performance assessments is often hindered by the use of unbalanced designs for assigning examinees to tasks and raters because such designs result in very…
Descriptors: Performance Based Assessment, Medicine, Factor Analysis, Test Items
Martineau, Joseph A.; Wyse, Adam E. – Measurement: Interdisciplinary Research and Perspectives, 2015
This article is a commentary of a paper by Derek C. Briggs and Frederick A. Peck, "Using Learning Progressions to Design Vertical Scales That Support Coherent Inferences about Student Growth," which describes an elegant potential framework for at least beginning to address three priorities in large-scale assessment that have not been…
Descriptors: Performance Factors, Barriers, Program Implementation, Group Testing
Kubinger, Klaus D. – Educational and Psychological Measurement, 2009
The linear logistic test model (LLTM) breaks down the item parameter of the Rasch model as a linear combination of some hypothesized elementary parameters. Although the original purpose of applying the LLTM was primarily to generate test items with specified item difficulty, there are still many other potential applications, which may be of use…
Descriptors: Models, Test Items, Psychometrics, Item Response Theory
Maris, Gunter; Bechger, Timo M. – Psychometrika, 2004
It is shown that in the context of the Model with Internal Restrictions on the Item Difficulties (MIRID), different componential theories about an item set may lead to equivalent models. Furthermore, we provide conditions for the identifiability of the MIRID model parameters, and it will be shown how the MIRID model relates to the Linear Logistic…
Descriptors: Difficulty Level, Test Items, Models, Theories
Birney, Damian P.; Halford, Graeme S.; Andrews, Glenda – Educational and Psychological Measurement, 2006
Relational complexity (RC) theory conceptualizes an individual's processing capacity and a task's complexity along a common ordinal metric. The authors describe the development of the Latin Square Task (LST) that assesses the influence of RC on reasoning. The LST minimizes the role of knowledge and storage capacity and thus refines the…
Descriptors: Memory, Age Differences, Cognitive Processes, Psychometrics
Revuelta, Javier – Psychometrika, 2004
Two psychometric models are presented for evaluating the difficulty of the distractors in multiple-choice items. They are based on the criterion of rising distractor selection ratios, which facilitates interpretation of the subject and item parameters. Statistical inferential tools are developed in a Bayesian framework: modal a posteriori…
Descriptors: Multiple Choice Tests, Psychometrics, Models, Difficulty Level
Revuelta, Javier – Journal of Educational and Behavioral Statistics, 2004
This article presents a psychometric model for estimating ability and item-selection strategies in self-adapted testing. In contrast to computer adaptive testing, in self-adapted testing the examinees are allowed to select the difficulty of the items. The item-selection strategy is defined as the distribution of difficulty conditional on the…
Descriptors: Psychometrics, Adaptive Testing, Test Items, Evaluation Methods
Alonzo, Julie; Liu, Kimy; Tindal, Gerald – Behavioral Research and Teaching, 2007
In this technical report, the authors describe the development and piloting of reading comprehension measures as part of a comprehensive progress monitoring literacy assessment system developed in 2006 for use with students in Kindergarten through fifth grade. They begin with a brief overview of the two conceptual frameworks underlying the…
Descriptors: Reading Comprehension, Emergent Literacy, Test Construction, Literacy Education
Smith, Richard M. – 1982
There have been many attempts to formulate a procedure for extracting information from incorrect responses to multiple choice items, i.e., the assessment of partial knowledge. The results of these attempts can be described as inconsistent at best. It is hypothesized that these inconsistencies arise from three methodological problems: the…
Descriptors: Difficulty Level, Evaluation Methods, Goodness of Fit, Guessing (Tests)
Gorin, Joanna S.; Embretson, Susan E. – Applied Psychological Measurement, 2006
Recent assessment research joining cognitive psychology and psychometric theory has introduced a new technology, item generation. In algorithmic item generation, items are systematically created based on specific combinations of features that underlie the processing required to correctly solve a problem. Reading comprehension items have been more…
Descriptors: Difficulty Level, Test Items, Modeling (Psychology), Paragraph Composition
Guzman, Eduardo; Conejo, Ricardo; Garcia-Hervas; Emilio – Educational Technology & Society, 2005
SIETTE is a web-based adaptive testing system. It implements Computerized Adaptive Tests. These tests are tailor-made, theory-based tests, where questions shown to students, finalization of the test, and student knowledge estimation is accomplished adaptively. To construct these tests, SIETTE has an authoring environment comprising a suite of…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Construction, Test Items