Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 13 |
Descriptor
Measurement Techniques | 29 |
Test Length | 29 |
Test Construction | 13 |
Item Response Theory | 11 |
Computer Assisted Testing | 9 |
Test Items | 8 |
Adaptive Testing | 7 |
Test Reliability | 6 |
Comparative Analysis | 5 |
Educational Testing | 5 |
Item Banks | 5 |
More ▼ |
Source
Author
Hambleton, Ronald K. | 2 |
Wright, Benjamin D. | 2 |
Allan S. Cohen | 1 |
Alonso, Jordi | 1 |
Bejar, Isaac I. | 1 |
Bergstrom, Betty | 1 |
Brown, Anna | 1 |
Cliff, Norman | 1 |
Davey, Tim | 1 |
De Gruijter, Dato N. M. | 1 |
Deville, Craig | 1 |
More ▼ |
Publication Type
Education Level
Higher Education | 2 |
Postsecondary Education | 2 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
Advanced Placement… | 1 |
Preliminary Scholastic… | 1 |
SAT (College Admission Test) | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating
Lin, Yin; Brown, Anna; Williams, Paul – Educational and Psychological Measurement, 2023
Several forced-choice (FC) computerized adaptive tests (CATs) have emerged in the field of organizational psychology, all of them employing ideal-point items. However, despite most items developed historically follow dominance response models, research on FC CAT using dominance items is limited. Existing research is heavily dominated by…
Descriptors: Measurement Techniques, Computer Assisted Testing, Adaptive Testing, Industrial Psychology
Sedat Sen; Allan S. Cohen – Educational and Psychological Measurement, 2024
A Monte Carlo simulation study was conducted to compare fit indices used for detecting the correct latent class in three dichotomous mixture item response theory (IRT) models. Ten indices were considered: Akaike's information criterion (AIC), the corrected AIC (AICc), Bayesian information criterion (BIC), consistent AIC (CAIC), Draper's…
Descriptors: Goodness of Fit, Item Response Theory, Sample Size, Classification
Feher, Anita; Smith, Martin M.; Saklofske, Donald H.; Plouffe, Rachel A.; Wilson, Claire A.; Sherry, Simon B. – Journal of Psychoeducational Assessment, 2020
The Big Three Perfectionism Scale (BTPS) is a 45-item self-report measure of perfectionism with three overarching factors: rigid, self-critical, and narcissistic perfectionism. Our objective was to create a brief version of the BTPS, the Big Three Perfectionism Scale--Short Form (BTPS-SF). Sixteen items were selected, and confirmatory factor…
Descriptors: Personality Measures, Personality Traits, Test Construction, Measurement Techniques
Jacob, Brian A. – Center on Children and Families at Brookings, 2016
Contrary to popular belief, modern cognitive assessments--including the new Common Core tests--produce test scores based on sophisticated statistical models rather than the simple percent of items a student answers correctly. While there are good reasons for this, it means that reported test scores depend on many decisions made by test designers,…
Descriptors: Scores, Common Core State Standards, Test Length, Test Content
Liang, Tie; Wells, Craig S.; Hambleton, Ronald K. – Journal of Educational Measurement, 2014
As item response theory has been more widely applied, investigating the fit of a parametric model becomes an important part of the measurement process. There is a lack of promising solutions to the detection of model misfit in IRT. Douglas and Cohen introduced a general nonparametric approach, RISE (Root Integrated Squared Error), for detecting…
Descriptors: Item Response Theory, Measurement Techniques, Nonparametric Statistics, Models
Zheng, Chunmei – ProQuest LLC, 2013
Educational and psychological constructs are normally measured by multifaceted dimensions. The measured construct is defined and measured by a set of related subdomains. A bifactor model can accurately describe such data with both the measured construct and the related subdomains. However, a limitation of the bifactor model is the orthogonality…
Descriptors: Educational Testing, Measurement Techniques, Test Items, Models
Doebler, Anna; Doebler, Philipp; Holling, Heinz – Psychometrika, 2013
The common way to calculate confidence intervals for item response theory models is to assume that the standardized maximum likelihood estimator for the person parameter [theta] is normally distributed. However, this approximation is often inadequate for short and medium test lengths. As a result, the coverage probabilities fall below the given…
Descriptors: Foreign Countries, Item Response Theory, Computation, Hypothesis Testing
Kolen, Michael J.; Tong, Ye – Educational Measurement: Issues and Practice, 2010
Psychometric properties of item response theory proficiency estimates are considered in this paper. Proficiency estimators based on summed scores and pattern scores include non-Bayes maximum likelihood and test characteristic curve estimators and Bayesian estimators. The psychometric properties investigated include reliability, conditional…
Descriptors: Test Length, Psychometrics, Item Response Theory, Scores
Luecht, Richard M.; Sireci, Stephen G. – College Board, 2011
Over the past four decades, there has been incremental growth in computer-based testing (CBT) as a viable alternative to paper-and-pencil testing. However, the transition to CBT is neither easy nor inexpensive. As Drasgow, Luecht, and Bennett (2006) noted, many design engineering, test development, operations/logistics, and psychometric changes…
Descriptors: College Entrance Examinations, Computer Assisted Testing, Educational Technology, Evaluation Methods
Kim, Jiseon – ProQuest LLC, 2010
Classification testing has been widely used to make categorical decisions by determining whether an examinee has a certain degree of ability required by established standards. As computer technologies have developed, classification testing has become more computerized. Several approaches have been proposed and investigated in the context of…
Descriptors: Test Length, Computer Assisted Testing, Classification, Probability
Seo, Dong Gi – ProQuest LLC, 2011
Most computerized adaptive tests (CAT) have been studied under the framework of unidimensional item response theory. However, many psychological variables are multidimensional and might benefit from using a multidimensional approach to CAT. In addition, a number of psychological variables (e.g., quality of life, depression) can be conceptualized…
Descriptors: Test Length, Quality of Life, Item Analysis, Geometric Concepts
Woods, Carol M. – Applied Psychological Measurement, 2008
In Ramsay-curve item response theory (RC-IRT), the latent variable distribution is estimated simultaneously with the item parameters of a unidimensional item response model using marginal maximum likelihood estimation. This study evaluates RC-IRT for the three-parameter logistic (3PL) model with comparisons to the normal model and to the empirical…
Descriptors: Test Length, Computation, Item Response Theory, Maximum Likelihood Statistics

Mayer, John D. – Perceptual and Motor Skills, 1983
Kelly's formula estimates sampling variance of correlation corrected for attenuation by using split-half reliabilities. In some cases, coefficient alpha estimate of reliability is preferable. A simulation study suggests a variation of Kelly's formula can be used appropriately with coefficient alpha. Kelly's formula is modified to accept…
Descriptors: Correlation, Measurement Techniques, Reliability, Sampling

Qualls, Audrey L. – Applied Measurement in Education, 1995
Classically parallel, tau-equivalently parallel, and congenerically parallel models representing various degrees of part-test parallelism and their appropriateness for tests composed of multiple item formats are discussed. An appropriate reliability estimate for a test with multiple item formats is presented and illustrated. (SLD)
Descriptors: Achievement Tests, Estimation (Mathematics), Measurement Techniques, Test Format

Millman, Jason – Review of Educational Research, 1973
Procedures for establishing standards and determining the number of items needed in criterion referenced measures were reviewed. Discussion of setting a passing score was organized around: performance of others, item content, educational consequences, psychological and financial costs, and error due to guessing and item sampling. (Author)
Descriptors: Criterion Referenced Tests, Educational Research, Literature Reviews, Measurement Techniques
Previous Page | Next Page ยป
Pages: 1 | 2