Publication Date
| In 2026 | 0 |
| Since 2025 | 215 |
| Since 2022 (last 5 years) | 1084 |
| Since 2017 (last 10 years) | 2594 |
| Since 2007 (last 20 years) | 4955 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Bolt, Daniel M.; Liao, Xiangyi – Journal of Educational Measurement, 2021
We revisit the empirically observed positive correlation between DIF and difficulty studied by Freedle and commonly seen in tests of verbal proficiency when comparing populations of different mean latent proficiency levels. It is shown that a positive correlation between DIF and difficulty estimates is actually an expected result (absent any true…
Descriptors: Test Bias, Difficulty Level, Correlation, Verbal Tests
Cum, Sait – International Journal of Assessment Tools in Education, 2021
In this study, it was claimed that ROC analysis, which is used to determine to what extent medical diagnosis tests can be differentiated between patients and non-patients, can also be used to examine the discrimination of binary scored items in cognitive tests. In order to obtain various evidence for this claim, the 2x2 contingency table used in…
Descriptors: Test Items, Item Analysis, Discriminant Analysis, Item Response Theory
Zeynep Uzun; Tuncay Ögretmen – Large-scale Assessments in Education, 2025
This study aimed to evaluate the item model fit by equating the forms of the PISA 2018 mathematics subtest with concurrent common items equating in samples from Türkiye, the UK, and Italy. The answers given in mathematics subtest Forms 2, 8, and 12 were used in this context. Analyzes were performed using the Dichotomous Rasch Model in the WINSTEPS…
Descriptors: Item Response Theory, Test Items, Foreign Countries, Mathematics Tests
Mahdi Ghorbankhani; Keyvan Salehi – SAGE Open, 2025
Academic procrastination, the tendency to delay academic tasks without reasonable justification, has significant implications for students' academic performance and overall well-being. To measure this construct, numerous scales have been developed, among which the Academic Procrastination Scale (APS) has shown promise in assessing academic…
Descriptors: Psychometrics, Measures (Individuals), Time Management, Foreign Countries
Mehmet Ramazanoglu; Tayfun Akin – Education and Information Technologies, 2025
One of the most important indicators of artificial intelligence applications used to improve quality, effectiveness, and student success by optimizing instructional processes is readiness. Readiness is the cognitive, affective, and behavioral state of performing a behavior or using a technology. In this context, the development of a data…
Descriptors: Measures (Individuals), Artificial Intelligence, Technology Uses in Education, Readiness
Martin Steinbach; Carolin Eitemüller; Marc Rodemer; Maik Walpuski – International Journal of Science Education, 2025
The intricate relationship between representational competence and content knowledge in organic chemistry has been widely debated, and the ways in which representations contribute to task difficulty, particularly in assessment, remain unclear. This paper presents a multiple-choice test instrument for assessing individuals' knowledge of fundamental…
Descriptors: Organic Chemistry, Difficulty Level, Multiple Choice Tests, Fundamental Concepts
Al Lawati, Zahra Ali – Language Testing in Asia, 2023
This study discusses the characteristics of test specifications (specs) and item writer guidelines (IWGs), their role in item development of English as a Second Language (ESL) reading tests, and the use of the CEFR for specs development. This mixed-method study analyzed specs, IWGs, tests, and the Pearson Test of English General test statistics.…
Descriptors: Language Tests, Test Items, Test Construction, English (Second Language)
Chauliac, Margot; Catrysse, Leen; Gijbels, David; De Maeyer, Sven; Donche, Vincent – International Journal of Social Research Methodology, 2023
Completing questionnaires is a complex task in which multiple cognitive processes play an important role. Despite critiques of the reliability of questionnaires, our research strived to gain more insight into the process of completing questionnaires by using eye-tracking. We investigated how both the question and the categories of answers were…
Descriptors: Measurement Techniques, Questionnaires, Eye Movements, Individual Differences
Chalmers, R. Philip; Zheng, Guoguo – Applied Measurement in Education, 2023
This article presents generalizations of SIBTEST and crossing-SIBTEST statistics for differential item functioning (DIF) investigations involving more than two groups. After reviewing the original two-group setup for these statistics, a set of multigroup generalizations that support contrast matrices for joint tests of DIF are presented. To…
Descriptors: Test Bias, Test Items, Item Response Theory, Error of Measurement
Qiao, Chen; Hu, Xiao – IEEE Transactions on Learning Technologies, 2023
Free text answers to short questions can reflect students' mastery of concepts and their relationships relevant to learning objectives. However, automating the assessment of free text answers has been challenging due to the complexity of natural language. Existing studies often predict the scores of free text answers in a "black box"…
Descriptors: Computer Assisted Testing, Automation, Test Items, Semantics
Hildenbrand, Lena; Wiley, Jennifer – Discourse Processes: A Multidisciplinary Journal, 2023
The present study examined the relationship between working memory capacity (WMC) and comprehension on a multiple text assessment from the ACT test for college preparedness in which texts are available during question answering. Specifically, it was of interest whether differences in relations would be seen across different question types. Only…
Descriptors: Short Term Memory, Reading Comprehension, Predictor Variables, College Entrance Examinations
Kane, Jesse F. – ProQuest LLC, 2023
The idea of student engagement as a predictor of student success was first introduced by Alexander Astin (1974; 1984) who studied student involvement. The connection of student involvement and student success has led to the focus on student and how we measure it to ensure that institutions are doing all they can to improve outcomes. Nothing has…
Descriptors: Learner Engagement, College Freshmen, College Seniors, Student Surveys
Pearson, Christopher; Penna, Nigel – Assessment & Evaluation in Higher Education, 2023
E-assessments are becoming increasingly common and progressively more complex. Consequently, how these longer, more complex questions are designed and marked is imperative. This article uses the NUMBAS e-assessment tool to investigate the best practice for creating longer questions and their mark schemes on surveying modules taken by engineering…
Descriptors: Automation, Scoring, Engineering Education, Foreign Countries
Bingxue Zhang; Yang Shi; Yuxing Li; Chengliang Chai; Longfeng Hou – Interactive Learning Environments, 2023
The adaptive learning environment provides learning support that suits individual characteristics of students, and the student model of the adaptive learning environment is the key element to promote individualized learning. This paper provides a systematic overview of the existing student models, consequently showing that the Elo rating system…
Descriptors: Electronic Learning, Models, Students, Individualized Instruction
Daniel Jurich; Chunyan Liu – Applied Measurement in Education, 2023
Screening items for parameter drift helps protect against serious validity threats and ensure score comparability when equating forms. Although many high-stakes credentialing examinations operate with small sample sizes, few studies have investigated methods to detect drift in small sample equating. This study demonstrates that several newly…
Descriptors: High Stakes Tests, Sample Size, Item Response Theory, Equated Scores

Peer reviewed
Direct link
