Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Smith, Trevor I.; Bendjilali, Nasrine – Physical Review Physics Education Research, 2022
Several recent studies have employed item response theory (IRT) to rank incorrect responses to commonly used research-based multiple-choice assessments. These studies use Bock's nominal response model (NRM) for applying IRT to categorical (nondichotomous) data, but the response rankings only utilize half of the parameters estimated by the model.…
Descriptors: Item Response Theory, Test Items, Multiple Choice Tests, Science Tests
Ozdemir, Burhanettin; Gelbal, Selahattin – Education and Information Technologies, 2022
The computerized adaptive tests (CAT) apply an adaptive process in which the items are tailored to individuals' ability scores. The multidimensional CAT (MCAT) designs differ in terms of different item selection, ability estimation, and termination methods being used. This study aims at investigating the performance of the MCAT designs used to…
Descriptors: Scores, Computer Assisted Testing, Test Items, Language Proficiency
McClurg, Virginia M.; Codalata, Bonnie M.; Bell, Sherry M.; McCallum, R. Steve – Gifted Child Today, 2022
The psychometric integrity of a curriculum-based measure to screen for academic giftedness (Monitoring Instructional Responsiveness: Reading [MIR:R]) was evaluated by examining its ceiling, item gradient, and predictive capacity using 460 fourth grade students. Eighty fourth graders (17.39%) scored one standard deviation above the MIR:R mean. Ten…
Descriptors: Psychometrics, Screening Tests, Talent Identification, Academically Gifted
Delican, Burak – International Journal of Curriculum and Instruction, 2022
In this research, the questions in the Turkish Course (2,3,4) Worksheets were examined in terms of various classification systems. In this direction, the questions in the worksheets were evaluated with the document-material analysis technique in accordance with the structure of the qualitative research. During the research process, Turkish Course…
Descriptors: Worksheets, Elementary School Students, Turkish, Classification
Saenna, Watcharaporn; Phusee-orn, Songsak – Higher Education Studies, 2022
The purposes of the research were to: (1) create a scientific creativity measure for high school students; (2) examine the quality of the science creativity scale of the created test; (3) establish a benchmark for scientific creativity scores for high school students; and (4) study a scientific creativity level of students in the senior high…
Descriptors: Foreign Countries, Test Construction, High School Students, Creativity
Heine, Jörg-Henrik; Robitzsch, Alexander – Large-scale Assessments in Education, 2022
Research Question: This paper examines the overarching question of to what extent different analytic choices may influence the inference about country-specific cross-sectional and trend estimates in international large-scale assessments. We take data from the assessment of PISA mathematics proficiency from the four rounds from 2003 to 2012 as a…
Descriptors: Foreign Countries, International Assessment, Achievement Tests, Secondary School Students
Arikan, Serkan; Aybek, Eren Can – Educational Measurement: Issues and Practice, 2022
Many scholars compared various item discrimination indices in real or simulated data. Item discrimination indices, such as item-total correlation, item-rest correlation, and IRT item discrimination parameter, provide information about individual differences among all participants. However, there are tests that aim to select a very limited number…
Descriptors: Monte Carlo Methods, Item Analysis, Correlation, Individual Differences
Sideridis, Georgios; Tsaousis, Ioannis; Al-Harbi, Khaleel – Educational and Psychological Measurement, 2022
The goal of the present study was to address the analytical complexity of incorporating responses and response times through applying the Jeon and De Boeck mixture item response theory model in Mplus 8.7. Using both simulated and real data, we attempt to identify subgroups of responders that are rapid guessers or engage knowledge retrieval…
Descriptors: Reaction Time, Guessing (Tests), Item Response Theory, Information Retrieval
Steinmann, Isa; Sánchez, Daniel; van Laar, Saskia; Braeken, Johan – Assessment in Education: Principles, Policy & Practice, 2022
Questionnaire scales that are mixed-worded, i.e. include both positively and negatively worded items, often suffer from issues like low reliability and more complex latent structures than intended. Part of the problem might be that some responders fail to respond consistently to the mixed-worded items. We investigated the prevalence and impact of…
Descriptors: Response Style (Tests), Test Items, Achievement Tests, Foreign Countries
Moradi, Elahe; Ghabanchi, Zargham; Pishghadam, Reza – Language Testing in Asia, 2022
Given the significance of the test fairness, this study aimed to investigate a reading comprehension test for evidence of differential item functioning (DIF) based on English as a Foreign Language (EFL) learners' gender and their mode of learning (conventional vs. distance learning). To this end, 514 EFL learners were asked to take a 30-item…
Descriptors: Reading Comprehension, Test Bias, Test Items, Second Language Learning
Clements, Douglas H.; Banse, Holland; Sarama, Julie; Tatsuoka, Curtis; Joswick, Candace; Hudyma, Aaron; Van Dine, Douglas W.; Tatsuoka, Kikumi K. – Mathematical Thinking and Learning: An International Journal, 2022
Researchers often develop instruments using correctness scores (and a variety of theories and techniques, such as Item Response Theory) for validation and scoring. Less frequently, observations of children's strategies are incorporated into the design, development, and application of assessments. We conducted individual interviews of 833…
Descriptors: Item Response Theory, Computer Assisted Testing, Test Items, Mathematics Tests
Elahe Moradi; Zargham Ghabanchi – Journal of College Reading and Learning, 2025
The present study scrutinized Iranian EFL learners' mode of learning (distance vs. conventional) as a probable source of bias in employing cognitive and metacognitive reading comprehension strategies. To this end, a total of 514 Iranian distance and conventional EFL learners were asked to take a 30-item multiple-choice reading comprehension test…
Descriptors: Reading Strategies, Reading Instruction, Conventional Instruction, In Person Learning
Ika Zenita Ratnaningsih; Unika Prihatsanti; Anggun Resdasari Prasetyo; Bambang Sumintono – Journal of Applied Research in Higher Education, 2025
Purpose: The present study aimed to validate the Indonesian-language version of the psychological capital questionnaire (PCQ), specifically within the context of higher education, by utilising Rasch analysis to evaluate the reliability and validity aspect such as item-fit statistics, rating scale function, and differential item functioning of the…
Descriptors: Foreign Countries, Indonesian Languages, Test Validity, Psychological Characteristics
Nurussaniah Nurussaniah; Punaji Setyosari; Dedi Kuswandi; Saida Ulfa – Journal of Baltic Science Education, 2025
The accurate assessment of analytical thinking in physics, particularly in magnetism, poses substantial challenges due to the limitations of conventional tools in measuring higher-order cognitive skills. This study aimed to validate an analytical skills test in physics, based on Bloom's Revised Taxonomy, with an emphasis on the dimensions of…
Descriptors: Physics, Science Tests, Science Instruction, Thinking Skills
Ceylan Gündeger Kilci – International Journal of Assessment Tools in Education, 2025
This study examined the psychometric quality of multiple-choice questions generated by two AI tools, ChatGPT and DeepSeek, within the context of an undergraduate Educational Measurement and Evaluation course. Guided by ten learning outcomes (LOs) aligned with Bloom's Taxonomy, each tool was prompted to generate one five-option multiple-choice item…
Descriptors: Psychometrics, Multiple Choice Tests, Artificial Intelligence, Natural Language Processing

Peer reviewed
Direct link
