Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 5 |
Descriptor
Test Construction | 30 |
True Scores | 30 |
Test Reliability | 13 |
Measurement Techniques | 11 |
Error of Measurement | 9 |
Test Items | 9 |
Criterion Referenced Tests | 8 |
Test Validity | 8 |
Item Response Theory | 7 |
Psychometrics | 7 |
Reliability | 7 |
More ▼ |
Source
Author
Livingston, Samuel A. | 2 |
Banks, Karen | 1 |
Bergquist, Constance | 1 |
Biancarosa, Gina | 1 |
Brennan, Robert L. | 1 |
Chang, Lei | 1 |
Chen, Haiwen | 1 |
Cui, Ying | 1 |
Dick, Walter | 1 |
Dimitrov, Dimiter M. | 1 |
Dulaney, Chuck | 1 |
More ▼ |
Publication Type
Reports - Research | 11 |
Journal Articles | 8 |
Reports - Evaluative | 5 |
Speeches/Meeting Papers | 4 |
Reports - Descriptive | 3 |
Dissertations/Theses -… | 1 |
Numerical/Quantitative Data | 1 |
Education Level
Elementary Secondary Education | 2 |
Secondary Education | 2 |
Elementary Education | 1 |
Grade 2 | 1 |
High Schools | 1 |
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Oregon | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Advanced Placement… | 1 |
Dynamic Indicators of Basic… | 1 |
Law School Admission Test | 1 |
SAT (College Admission Test) | 1 |
Test of English as a Foreign… | 1 |
What Works Clearinghouse Rating
Selvi, Hüseyin – Higher Education Studies, 2020
This study aimed to examine the effect of using items from previous exams on students? pass-fail rates and on the psychometric properties of the tests and items. The study included data from 115 tests and 11,500 items used in the midterm and final exams of 3,910 students in the preclinical term at the Faculty of Medicine from 2014 to 2019. Data…
Descriptors: Answer Keys, Tests, Test Items, True Scores
von Davier, Alina A.; Chen, Haiwen – ETS Research Report Series, 2013
In the framework of the observed-score equating methods for the nonequivalent groups with anchor test design, there are 3 fundamentally different ways of using the information provided by the anchor scores to equate the scores of a new form to those of an old form. One method uses the anchor scores as a conditioning variable, such as the Tucker…
Descriptors: Equated Scores, Item Response Theory, True Scores, Methods
Stoolmiller, Michael; Biancarosa, Gina; Fien, Hank – Assessment for Effective Intervention, 2013
Lack of psychometric equivalence of oral reading fluency (ORF) passages used within a grade for screening and progress monitoring has recently become an issue with calls for the use of equating methods to ensure equivalence. To investigate the nature of the nonequivalence and to guide the choice of equating method to correct for nonequivalence,…
Descriptors: School Personnel, Reading Fluency, Emergent Literacy, Psychometrics
Gierl, Mark J.; Cui, Ying; Zhou, Jiawen – Journal of Educational Measurement, 2009
The attribute hierarchy method (AHM) is a psychometric procedure for classifying examinees' test item responses into a set of structured attribute patterns associated with different components from a cognitive model of task performance. Results from an AHM analysis yield information on examinees' cognitive strengths and weaknesses. Hence, the AHM…
Descriptors: Test Items, True Scores, Psychometrics, Algebra
Hagge, Sarah Lynn – ProQuest LLC, 2010
Mixed-format tests containing both multiple-choice and constructed-response items are widely used on educational tests. Such tests combine the broad content coverage and efficient scoring of multiple-choice items with the assessment of higher-order thinking skills thought to be provided by constructed-response items. However, the combination of…
Descriptors: Test Format, True Scores, Equated Scores, Psychometrics

Krus, David J.; Helmstadter, Gerald C. – Educational and Psychological Measurement, 1993
Negative coefficients of reliability, sometimes returned by the standard formula for estimation of the internal-consistency reliability, are neither theoretically nor numerically correct. Alternative strategies for test development in this special case are suggested. (Author)
Descriptors: Estimation (Mathematics), Reliability, Test Construction, Test Use
Dimitrov, Dimiter M. – 2003
This paper provides analytic evaluations of expected (marginal) true-score measures for binary items given their item response theory (IRT) calibration. Under the assumption of normal trait distributions, marginalized true scores, error variance, true score variance, and reliability for norm-referenced and criterion-references interpretations are…
Descriptors: Item Response Theory, Reliability, Test Construction, Test Items

Lord, Frederic M. – Psychometrika, 1975
For the six available sets of empirical data, the discrimination (slope) parameter of the logistic item characteristic curve was found to have a significant positive correlation over items with the difficulty (location) parameter. This unpleasant situation can be eliminated by a suitably chosen transformation of the ability scale. (Author/RC)
Descriptors: Ability, Aptitude Tests, Correlation, Item Analysis
Yang, Wen-Ling – 1997
Using an anchor-item design of test equating, the effects of three equating methods (Tucker linear and two three-parameter item-response-theory-based (3PL-IRT) methods), and the content representativeness of anchor items on the accuracy of equating were examined; and an innovative way of evaluating equating accuracy appropriate for the particular…
Descriptors: Equated Scores, Item Response Theory, Raw Scores, Test Construction

Eakin, Richard R.; Long, Clifford A. – Educational and Psychological Measurement, 1977
A scoring technique for true-false tests is presented. The technique, paired item scoring, involves combining two statements and having the student select one of the four resultants possible: true-true, false-true, true-false, and false-false. The combined item is treated as a multiple choice item. (Author/JKS)
Descriptors: Guessing (Tests), Measurement Techniques, Multiple Choice Tests, Objective Tests
Johnson, Stephen; Dulaney, Chuck; Banks, Karen – 2000
No test, however well designed, can measure a student's true achievement because numerous factors interfere with the ability to measure achievement. These factors are sources of measurement error, and the goal in creating tests is to have as little measurement error as possible. Error can result from the test design, factors related to individual…
Descriptors: Academic Achievement, Elementary Education, Error of Measurement, Measurement Techniques

Livingston, Samuel A. – Journal of Educational Measurement, 1972
This article is a reply to a previous paper (see TM 500 488) interpreting Livingston's original article (see TM 500 487). (CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Norm Referenced Tests, Test Construction
Sullins, Walter L. – 1971
Five-hundred dichotomously scored response patterns were generated with sequentially independent (SI) items and 500 with dependent (SD) items for each of thirty-six combinations of sampling parameters (i.e., three test lengths, three sample sizes, and four item difficulty distributions). KR-20, KR-21, and Split-Half (S-H) reliabilities were…
Descriptors: Comparative Analysis, Correlation, Error of Measurement, Item Analysis
Wang, Xiang-Bo; Harris, Vincent; Roussos, Louis – 2002
Multidimensionality is known to affect the accuracy of item parameter and ability estimations, which subsequently influences the computation of item characteristic curves (ICCs) and true scores. By judiciously combining sections of a Law School Admission Test (LSAT), 11 sections of varying degrees of uni- and multidimensional structures are used…
Descriptors: Ability, College Entrance Examinations, Computer Assisted Testing, Estimation (Mathematics)

Hsu, Louis M. – Applied Psychological Measurement, 1979
A comparison of the relative ordering power of separate and grouped-items true-false tests indicated that neither type of test was uniformly superior to the other across all levels of knowledge of examinees. Grouped-item tests were found superior for examinees with low levels of knowledge. (Author/CTM)
Descriptors: Academic Ability, Knowledge Level, Multiple Choice Tests, Scores
Previous Page | Next Page »
Pages: 1 | 2