NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers2
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 24 results Save | Export
Rachel Lee – ProQuest LLC, 2024
Classical item analysis (CIA) entails summarizing items based on two key attributes: item difficulty and item discrimination, defined as the proportion of examinees answering correctly and the difference in correctness between high and low scorers. Recent insights reveal a direct link between these measures and aspects of signal detection theory…
Descriptors: Item Analysis, Knowledge Level, Difficulty Level, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Stephen Humphry; Paul Montuoro; Carolyn Maxwell – Journal of Psychoeducational Assessment, 2024
This article builds upon a proiminent definition of construct validity that focuses on variation in attributes causing variation in measurement outcomes. This article synthesizes the defintion and uses Rasch measurement modeling to explicate a modified conceptualization of construct validity for assessments of developmental attributes. If…
Descriptors: Construct Validity, Measurement Techniques, Developmental Stages, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Trace, Jonathan; Brown, James Dean; Janssen, Gerriet; Kozhevnikova, Liudmila – Language Testing, 2017
Cloze tests have been the subject of numerous studies regarding their function and use in both first language and second language contexts (e.g., Jonz & Oller, 1994; Watanabe & Koyama, 2008). From a validity standpoint, one area of investigation has been the extent to which cloze tests measure reading ability beyond the sentence level.…
Descriptors: Cloze Procedure, Language Tests, Test Items, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Stancliffe, R. J.; Wilson, N. J.; Bigby, C.; Balandin, S.; Craig, D. – Journal of Intellectual Disability Research, 2014
Background: We compared responsiveness to two self-report assessments of loneliness: the "UCLA Loneliness Scale" (UCLALS) designed for the general community, and the "Modified Worker Loneliness Questionnaire" (MWLQ) designed for people with intellectual disability (ID). Methods: Participants were 56 older adults with…
Descriptors: Measurement Techniques, Psychological Patterns, Measures (Individuals), Older Adults
Peer reviewed Peer reviewed
Direct linkDirect link
Raker, Jeffrey R.; Trate, Jaclyn M.; Holme, Thomas A.; Murphy, Kristen – Journal of Chemical Education, 2013
Experts use their domain expertise and knowledge of examinees' ability levels as they write test items. The expert test writer can then estimate the difficulty of the test items subjectively. However, an objective method for assigning difficulty to a test item would capture the cognitive demands imposed on the examinee as well as be…
Descriptors: Organic Chemistry, Test Items, Item Analysis, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Zeuch, Nina; Holling, Heinz; Kuhn, Jorg-Tobias – Learning and Individual Differences, 2011
The Latin Square Task (LST) was developed by Birney, Halford, and Andrews [Birney, D. P., Halford, G. S., & Andrews, G. (2006). Measuring the influence of cognitive complexity on relational reasoning: The development of the Latin Square Task. Educational and Psychological Measurement, 66, 146-171.] and represents a non-domain specific,…
Descriptors: Difficulty Level, Geometric Concepts, Item Response Theory, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Laprise, Shari L. – College Teaching, 2012
Successful exam composition can be a difficult task. Exams should not only assess student comprehension, but be learning tools in and of themselves. In a biotechnology course delivered to nonmajors at a business college, objective multiple-choice test questions often require students to choose the exception or "not true" choice. Anecdotal student…
Descriptors: Feedback (Response), Test Items, Multiple Choice Tests, Biotechnology
Peer reviewed Peer reviewed
Direct linkDirect link
Kingsbury, G. Gage; Wise, Steven L. – Journal of Applied Testing Technology, 2011
Development of adaptive tests used in K-12 settings requires the creation of stable measurement scales to measure the growth of individual students from one grade to the next, and to measure change in groups from one year to the next. Accountability systems like No Child Left Behind require stable measurement scales so that accountability has…
Descriptors: Elementary Secondary Education, Adaptive Testing, Academic Achievement, Measures (Individuals)
PDF pending restoration PDF pending restoration
Frisbie, David A. – 1980
The development of a new technique, the Relative Difficulty Ratio (RDR), is described, as well as how it can be used to determine the difficulty level of a test so that meaningful inter-test difficulty comparisons can be made. Assumptions made in computing RDR include: 1) each item must be scored dichotomously with only one answer choice keyed as…
Descriptors: Difficulty Level, Item Analysis, Measurement Techniques, Scores
Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2009
In this technical report, we describe the development and piloting of a series of mathematics progress monitoring measures intended for use with students in grade 1. These measures, available as part of easyCBM [TM], an online progress monitoring assessment system, were developed in 2008 and administered to approximately 2800 students from schools…
Descriptors: Academic Achievement, Research Reports, Grade 1, Outcome Measures
Smith, Donald M. – 1974
The concept of scaled achievement tests is discussed and a method of selecting those items of a test that form the most scalable (i.e., having the highest coefficient of reproducibility) subset is presented. Sometimes called a monotonic-deterministic model, this type of test assumes that the test items may be sequentially ordered. To determine the…
Descriptors: Achievement Tests, Arithmetic, Difficulty Level, Item Analysis
Peer reviewed Peer reviewed
Lord, Frederic M. – Educational and Psychological Measurement, 1971
A number of empirical studies are suggested to answer certain questions in connection with flexilevel tests. (MS)
Descriptors: Comparative Analysis, Difficulty Level, Guessing (Tests), Item Analysis
Samejima, Fumiko – 1980
Many combinations of a method and an approach for estimating the operating characteristics of the graded item responses, without assuming any mathematical forms, have been produced. In these methods, a set of items whose characteristics are known, or Old Test, is used, which has a large, constant amount of test information throughout the interval…
Descriptors: Difficulty Level, Item Analysis, Latent Trait Theory, Least Squares Statistics
Peer reviewed Peer reviewed
Lord, Frederic M. – Educational and Psychological Measurement, 1971
Descriptors: Ability, Adaptive Testing, Computer Oriented Programs, Difficulty Level
Lord, Frederic M. – 1971
A flexilevel test is found to be inferior to a peaked conventional test for measuring examinees in the middle of the ability range, superior for examinees at the extremes. Throughout the entire range of ability, a flexilevel test is much superior to any conventional test that attempts to provide accurate measurement at both extremes. See also ED…
Descriptors: Ability, Comparative Analysis, Difficulty Level, Guessing (Tests)
Previous Page | Next Page ยป
Pages: 1  |  2