Publication Date
In 2025 | 0 |
Since 2024 | 7 |
Since 2021 (last 5 years) | 18 |
Since 2016 (last 10 years) | 35 |
Since 2006 (last 20 years) | 412 |
Descriptor
Item Analysis | 626 |
Test Items | 184 |
Psychometrics | 123 |
Test Construction | 123 |
Foreign Countries | 121 |
Measures (Individuals) | 120 |
Test Validity | 106 |
Item Response Theory | 98 |
Test Reliability | 94 |
Evaluation Methods | 90 |
Factor Analysis | 79 |
More ▼ |
Source
Author
Tindal, Gerald | 12 |
Alonzo, Julie | 10 |
Lai, Cheng Fei | 7 |
Hambleton, Ronald K. | 4 |
Chang, Hua-Hua | 3 |
Eggen, Theo J. H. M. | 3 |
French, Brian F. | 3 |
Hills, John R. | 3 |
Raykov, Tenko | 3 |
Rogers, H. Jane | 3 |
Sireci, Stephen G. | 3 |
More ▼ |
Publication Type
Education Level
Audience
Researchers | 7 |
Teachers | 5 |
Practitioners | 4 |
Policymakers | 1 |
Students | 1 |
Location
Canada | 15 |
Australia | 13 |
Oregon | 10 |
United States | 9 |
Turkey | 8 |
United Kingdom | 8 |
California | 7 |
Hong Kong | 7 |
China | 4 |
Florida | 4 |
Texas | 4 |
More ▼ |
Laws, Policies, & Programs
Individuals with Disabilities… | 13 |
No Child Left Behind Act 2001 | 13 |
Assessments and Surveys
What Works Clearinghouse Rating
Meets WWC Standards without Reservations | 1 |
Meets WWC Standards with or without Reservations | 1 |
Pere J. Ferrando; David Navarro-González; Urbano Lorenzo-Seva – Educational and Psychological Measurement, 2024
Descriptive fit indices that do not require a formal statistical basis and do not specifically depend on a given estimation criterion are useful as auxiliary devices for judging the appropriateness of unrestricted or exploratory factor analytical (UFA) solutions, when the problem is to decide the most appropriate number of common factors. While…
Descriptors: Factor Analysis, Item Analysis, Effect Size, Goodness of Fit
Sijia Huang; Dubravka Svetina Valdivia – Educational and Psychological Measurement, 2024
Identifying items with differential item functioning (DIF) in an assessment is a crucial step for achieving equitable measurement. One critical issue that has not been fully addressed with existing studies is how DIF items can be detected when data are multilevel. In the present study, we introduced a Lord's Wald X[superscript 2] test-based…
Descriptors: Item Analysis, Item Response Theory, Algorithms, Accuracy
Zachary K. Collier; Minji Kong; Olushola Soyoye; Kamal Chawla; Ann M. Aviles; Yasser Payne – Journal of Educational and Behavioral Statistics, 2024
Asymmetric Likert-type items in research studies can present several challenges in data analysis, particularly concerning missing data. These items are often characterized by a skewed scaling, where either there is no neutral response option or an unequal number of possible positive and negative responses. The use of conventional techniques, such…
Descriptors: Likert Scales, Test Items, Item Analysis, Evaluation Methods
Hao-Jan Luh; Zachary C. LaBrot; Cagla Cobek; Ryan Sunda; Lindsay M. Fallon – Journal of Educational and Psychological Consultation, 2024
Training in school-based consultation may encompass instruction on various consultation models, aims, and work with various partners. However, it is unclear how trainers currently structure coursework and the extent to which social justice is embedded in class. Therefore, we conducted a systematic replication of Hazel et al. (2010) analyses of…
Descriptors: Course Descriptions, Literature Reviews, Social Justice, Cultural Pluralism
Semere Kiros Bitew; Amir Hadifar; Lucas Sterckx; Johannes Deleu; Chris Develder; Thomas Demeester – IEEE Transactions on Learning Technologies, 2024
Multiple-choice questions (MCQs) are widely used in digital learning systems, as they allow for automating the assessment process. However, owing to the increased digital literacy of students and the advent of social media platforms, MCQ tests are widely shared online, and teachers are continuously challenged to create new questions, which is an…
Descriptors: Multiple Choice Tests, Computer Assisted Testing, Test Construction, Test Items
Eray Selçuk; Ergül Demir – International Journal of Assessment Tools in Education, 2024
This research aims to compare the ability and item parameter estimations of Item Response Theory according to Maximum likelihood and Bayesian approaches in different Monte Carlo simulation conditions. For this purpose, depending on the changes in the priori distribution type, sample size, test length, and logistics model, the ability and item…
Descriptors: Item Response Theory, Item Analysis, Test Items, Simulation
Raykov, Tenko – Measurement: Interdisciplinary Research and Perspectives, 2023
This software review discusses the capabilities of Stata to conduct item response theory modeling. The commands needed for fitting the popular one-, two-, and three-parameter logistic models are initially discussed. The procedure for testing the discrimination parameter equality in the one-parameter model is then outlined. The commands for fitting…
Descriptors: Item Response Theory, Models, Comparative Analysis, Item Analysis
van der Linden, Wim J. – Journal of Educational and Behavioral Statistics, 2022
Two independent statistical tests of item compromise are presented, one based on the test takers' responses and the other on their response times (RTs) on the same items. The tests can be used to monitor an item in real time during online continuous testing but are also applicable as part of post hoc forensic analysis. The two test statistics are…
Descriptors: Test Items, Item Analysis, Item Response Theory, Computer Assisted Testing
Effatpanah, Farshad; Baghaei, Purya – Practical Assessment, Research & Evaluation, 2023
Item response theory (IRT) refers to a family of mathematical models which describe the relationship between latent continuous variables (attributes or characteristics) and their manifestations (dichotomous/polytomous observed outcomes or responses) with regard to a set of item characteristics. Researchers typically use parametric IRT (PIRT)…
Descriptors: Item Response Theory, Feedback (Response), Mathematical Models, Item Analysis
Camenares, Devin – International Journal for the Scholarship of Teaching and Learning, 2022
Balancing assessment of learning outcomes with the expectations of students is a perennial challenge in education. Difficult exams, in which many students perform poorly, exacerbate this problem and can inspire a wide variety of interventions, such as a grading curve. However, addressing poor performance can sometimes distort or inflate grades and…
Descriptors: College Students, Student Evaluation, Tests, Test Items
Emma Walland – Research Matters, 2024
GCSE examinations (taken by students aged 16 years in England) are not intended to be speeded (i.e. to be partly a test of how quickly students can answer questions). However, there has been little research exploring this. The aim of this research was to explore the speededness of past GCSE written examinations, using only the data from scored…
Descriptors: Educational Change, Test Items, Item Analysis, Scoring
Raykov, Tenko; Pusic, Martin – Educational and Psychological Measurement, 2023
This note is concerned with evaluation of location parameters for polytomous items in multiple-component measuring instruments. A point and interval estimation procedure for these parameters is outlined that is developed within the framework of latent variable modeling. The method permits educational, behavioral, biomedical, and marketing…
Descriptors: Item Analysis, Measurement Techniques, Computer Software, Intervals
Peabody, Michael R. – Measurement: Interdisciplinary Research and Perspectives, 2023
Many organizations utilize some form of automation in the test assembly process; either fully algorithmic or heuristically constructed. However, one issue with heuristic models is that when the test assembly problem changes the entire model may need to be re-conceptualized and recoded. In contrast, mixed-integer programming (MIP) is a mathematical…
Descriptors: Programming Languages, Algorithms, Heuristics, Mathematical Models
Wolkowitz, Amanda A.; Foley, Brett; Zurn, Jared – Practical Assessment, Research & Evaluation, 2023
The purpose of this study is to introduce a method for converting scored 4-option multiple-choice (MC) items into scored 3-option MC items without re-pretesting the 3-option MC items. This study describes a six-step process for achieving this goal. Data from a professional credentialing exam was used in this study and the method was applied to 24…
Descriptors: Multiple Choice Tests, Test Items, Accuracy, Test Format
Beisemann, Marie; Forthmann, Boris; Bürkner, Paul-Christian; Holling, Heinz – Journal of Creative Behavior, 2020
The Remote Associates Test (RAT; Mednick, 1962; Mednick & Mednick, 1967) is a commonly employed test of creative convergent thinking. The RAT is scored with a dichotomous scoring, scoring correct answers as 1 and all other answers as 0. Based on recent research into the information processing underlying RAT performance, we argued that the…
Descriptors: Psychometrics, Scoring, Tests, Semantics