Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 7 |
Since 2006 (last 20 years) | 10 |
Descriptor
Models | 14 |
Evaluation Methods | 6 |
Item Response Theory | 5 |
Test Items | 5 |
Educational Assessment | 4 |
Statistical Analysis | 4 |
Accuracy | 3 |
Comparative Analysis | 3 |
Difficulty Level | 3 |
Evidence | 3 |
Knowledge Level | 3 |
More ▼ |
Source
Applied Measurement in… | 14 |
Author
Ackerman, Terry A. | 1 |
Bauer, Malcolm I. | 1 |
Bejar, Isaac I. | 1 |
Birenbaum, Menucha | 1 |
Bovaird, James A. | 1 |
Brookhart, Susan M. | 1 |
Cohen, Allan | 1 |
Ercikan, Kadriye | 1 |
Finch, Holmes | 1 |
French, Brian F. | 1 |
George, Ann Cathrice | 1 |
More ▼ |
Publication Type
Journal Articles | 14 |
Reports - Research | 7 |
Reports - Evaluative | 6 |
Information Analyses | 1 |
Reports - Descriptive | 1 |
Education Level
High Schools | 3 |
Secondary Education | 3 |
Middle Schools | 2 |
Early Childhood Education | 1 |
Elementary Education | 1 |
Elementary Secondary Education | 1 |
Grade 1 | 1 |
Grade 2 | 1 |
Grade 3 | 1 |
Grade 7 | 1 |
Junior High Schools | 1 |
More ▼ |
Audience
Laws, Policies, & Programs
Assessments and Surveys
Stanford Achievement Tests | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Pham, Duy N.; Wells, Craig S.; Bauer, Malcolm I.; Wylie, E. Caroline; Monroe, Scott – Applied Measurement in Education, 2021
Assessments built on a theory of learning progressions are promising formative tools to support learning and teaching. The quality and usefulness of those assessments depend, in large part, on the validity of the theory-informed inferences about student learning made from the assessment results. In this study, we introduced an approach to address…
Descriptors: Formative Evaluation, Mathematics Instruction, Mathematics Achievement, Middle School Students
Finch, Holmes; French, Brian F. – Applied Measurement in Education, 2019
The usefulness of item response theory (IRT) models depends, in large part, on the accuracy of item and person parameter estimates. For the standard 3 parameter logistic model, for example, these parameters include the item parameters of difficulty, discrimination, and pseudo-chance, as well as the person ability parameter. Several factors impact…
Descriptors: Item Response Theory, Accuracy, Test Items, Difficulty Level
Raczynski, Kevin; Cohen, Allan – Applied Measurement in Education, 2018
The literature on Automated Essay Scoring (AES) systems has provided useful validation frameworks for any assessment that includes AES scoring. Furthermore, evidence for the scoring fidelity of AES systems is accumulating. Yet questions remain when appraising the scoring performance of AES systems. These questions include: (a) which essays are…
Descriptors: Essay Tests, Test Scoring Machines, Test Validity, Evaluators
Hawley, Leslie R.; Bovaird, James A.; Wu, ChaoRong – Applied Measurement in Education, 2017
Value-added assessment methods have been criticized by researchers and policy makers for a number of reasons. One issue includes the sensitivity of model results across different outcome measures. This study examined the utility of incorporating multivariate latent variable approaches within a traditional value-added framework. We evaluated the…
Descriptors: Value Added Models, Reliability, Multivariate Analysis, Scaling
George, Ann Cathrice; Robitzsch, Alexander – Applied Measurement in Education, 2018
This article presents a new perspective on measuring gender differences in the large-scale assessment study Trends in International Science Study (TIMSS). The suggested empirical model is directly based on the theoretical competence model of the domain mathematics and thus includes the interaction between content and cognitive sub-competencies.…
Descriptors: Achievement Tests, Elementary Secondary Education, Mathematics Achievement, Mathematics Tests
Koziol, Natalie A. – Applied Measurement in Education, 2016
Testlets, or groups of related items, are commonly included in educational assessments due to their many logistical and conceptual advantages. Despite their advantages, testlets introduce complications into the theory and practice of educational measurement. Responses to items within a testlet tend to be correlated even after controlling for…
Descriptors: Classification, Accuracy, Comparative Analysis, Models
Ercikan, Kadriye; Oliveri, María Elena – Applied Measurement in Education, 2016
Assessing complex constructs such as those discussed under the umbrella of 21st century constructs highlights the need for a principled assessment design and validation approach. In our discussion, we made a case for three considerations: (a) taking construct complexity into account across various stages of assessment development such as the…
Descriptors: Evaluation Methods, Test Construction, Design, Scaling
Bejar, Isaac I. – Applied Measurement in Education, 2010
The foregoing articles constitute what I consider a comprehensive and clear description of the redesign process of a major assessment. The articles serve to illustrate the problems that will need to be addressed by large-scale assessments in the twenty-first century. Primary among them is how to organize the development of such assessments to meet…
Descriptors: Advanced Placement Programs, Equivalency Tests, Evidence, Test Construction
Huff, Kristen; Steinberg, Linda; Matts, Thomas – Applied Measurement in Education, 2010
The cornerstone of evidence-centered assessment design (ECD) is an evidentiary argument that requires that each target of measurement (e.g., learning goal) for an assessment be expressed as a "claim" to be made about an examinee that is relevant to the specific purpose and audience(s) for the assessment. The "observable evidence" required to…
Descriptors: Advanced Placement Programs, Equivalency Tests, Evidence, Test Construction
Lee, Hee-Sun; Liu, Ou Lydia; Linn, Marcia C. – Applied Measurement in Education, 2011
This study explores measurement of a construct called knowledge integration in science using multiple-choice and explanation items. We use construct and instructional validity evidence to examine the role multiple-choice and explanation items plays in measuring students' knowledge integration ability. For construct validity, we analyze item…
Descriptors: Knowledge Level, Construct Validity, Validity, Scaffolding (Teaching Technique)

Plake, Barbara S. – Applied Measurement in Education, 1995
This article provides a framework for the rest of the articles in this special issue comparing the utility of three standard-setting methods with complex performance assessments. The context of the standard setting study is described, and the methods are outlined. (SLD)
Descriptors: Comparative Analysis, Criteria, Decision Making, Educational Assessment

Brookhart, Susan M. – Applied Measurement in Education, 1997
Presents a theory about the role of classroom assessment in motivating student effort and achievement. The theory postulates that in any particular class the classroom assessment environment is played out in repeated assessment events through which a teacher communicates and students respond according to their perceptions. (SLD)
Descriptors: Academic Achievement, Communication (Thought Transfer), Educational Assessment, Educational Environment

Ackerman, Terry A. – Applied Measurement in Education, 1994
When item response data do not satisfy the unidimensionality assumption, multidimensional item response theory (MIRT) should be used to model the item-examinee interaction. This article presents and discusses MIRT analyses designed to give better insight into what individual items are measuring. (SLD)
Descriptors: Evaluation Methods, Item Response Theory, Measurement Techniques, Models

Birenbaum, Menucha; Tatsuoka, Kikumi K. – Applied Measurement in Education, 1993
The item-response-theory-based (IRT) rule-space model was used to diagnose student knowledge about how exponents behave in multiplication and division in a sample of 431 tenth-grade students. Implications for using feedback from the rule-space model in instruction and assessment are discussed. (SLD)
Descriptors: Cognitive Processes, Diagnostic Tests, Division, Educational Assessment