Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 5 |
Since 2006 (last 20 years) | 7 |
Descriptor
Guidelines | 9 |
Test Items | 9 |
Test Length | 9 |
Simulation | 6 |
Sample Size | 5 |
Comparative Analysis | 4 |
Item Response Theory | 4 |
Error of Measurement | 3 |
Item Analysis | 3 |
Test Format | 3 |
Correlation | 2 |
More ▼ |
Author
Andersson, Björn | 1 |
Banks, Kathleen | 1 |
Berk, Ronald A. | 1 |
Cappaert, Kevin | 1 |
Choi, Youn-Jeng | 1 |
Goodrich, J. Marc | 1 |
Guo, Wenjing | 1 |
Huang, Feifei | 1 |
Koziol, Natalie A. | 1 |
Kárász, Judit T. | 1 |
Lee, Won-Chan | 1 |
More ▼ |
Publication Type
Journal Articles | 6 |
Reports - Research | 6 |
Dissertations/Theses -… | 1 |
Guides - Non-Classroom | 1 |
Reference Materials -… | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Audience
Community | 1 |
Practitioners | 1 |
Location
Laws, Policies, & Programs
Job Training Partnership Act… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Guo, Wenjing; Choi, Youn-Jeng – Educational and Psychological Measurement, 2023
Determining the number of dimensions is extremely important in applying item response theory (IRT) models to data. Traditional and revised parallel analyses have been proposed within the factor analysis framework, and both have shown some promise in assessing dimensionality. However, their performance in the IRT framework has not been…
Descriptors: Item Response Theory, Evaluation Methods, Factor Analysis, Guidelines
Wang, Shaojie; Zhang, Minqiang; Lee, Won-Chan; Huang, Feifei; Li, Zonglong; Li, Yixing; Yu, Sufang – Journal of Educational Measurement, 2022
Traditional IRT characteristic curve linking methods ignore parameter estimation errors, which may undermine the accuracy of estimated linking constants. Two new linking methods are proposed that take into account parameter estimation errors. The item- (IWCC) and test-information-weighted characteristic curve (TWCC) methods employ weighting…
Descriptors: Item Response Theory, Error of Measurement, Accuracy, Monte Carlo Methods
Kárász, Judit T.; Széll, Krisztián; Takács, Szabolcs – Quality Assurance in Education: An International Perspective, 2023
Purpose: Based on the general formula, which depends on the length and difficulty of the test, the number of respondents and the number of ability levels, this study aims to provide a closed formula for the adaptive tests with medium difficulty (probability of solution is p = 1/2) to determine the accuracy of the parameters for each item and in…
Descriptors: Test Length, Probability, Comparative Analysis, Difficulty Level
Koziol, Natalie A.; Goodrich, J. Marc; Yoon, HyeonJin – Educational and Psychological Measurement, 2022
Differential item functioning (DIF) is often used to examine validity evidence of alternate form test accommodations. Unfortunately, traditional approaches for evaluating DIF are prone to selection bias. This article proposes a novel DIF framework that capitalizes on regression discontinuity design analysis to control for selection bias. A…
Descriptors: Regression (Statistics), Item Analysis, Validity, Testing Accommodations
Andersson, Björn – Journal of Educational Measurement, 2016
In observed-score equipercentile equating, the goal is to make scores on two scales or tests measuring the same construct comparable by matching the percentiles of the respective score distributions. If the tests consist of different items with multiple categories for each item, a suitable model for the responses is a polytomous item response…
Descriptors: Equated Scores, Item Response Theory, Error of Measurement, Tests
Walker, Cindy M.; Zhang, Bo; Banks, Kathleen; Cappaert, Kevin – Educational and Psychological Measurement, 2012
The purpose of this simulation study was to establish general effect size guidelines for interpreting the results of differential bundle functioning (DBF) analyses using simultaneous item bias test (SIBTEST). Three factors were manipulated: number of items in a bundle, test length, and magnitude of uniform differential item functioning (DIF)…
Descriptors: Test Bias, Test Length, Simulation, Guidelines
Sunnassee, Devdass – ProQuest LLC, 2011
Small sample equating remains a largely unexplored area of research. This study attempts to fill in some of the research gaps via a large-scale, IRT-based simulation study that evaluates the performance of seven small-sample equating methods under various test characteristic and sampling conditions. The equating methods considered are typically…
Descriptors: Test Length, Test Format, Sample Size, Simulation
Berk, Ronald A. – 1979
Four factors essential to determining how many items should be constructed or sampled for a set of objectives are examined: (1) importance and type of decisions to be made with the results; (2) importance and emphases assigned to the instructional and behavioral objectives; (3) number of objectives; (4) practical constraints, such as item writing…
Descriptors: Behavioral Objectives, Course Objectives, Criterion Referenced Tests, Decision Making
New York State Div. for Youth, Albany. – 1985
This guide is designed to serve as a reference to assist providers of Job Training Partnership Act-funded programs in selecting appropriate interest, aptitude, and pre-employment and job readiness tests. Descriptions of 53 interest tests, 38 aptitude tests, and 37 pre-employment and job readiness tests are provided. Each description contains…
Descriptors: Aptitude Tests, Employment Potential, Evaluation Criteria, Guidelines