Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 7 |
Since 2016 (last 10 years) | 22 |
Since 2006 (last 20 years) | 31 |
Descriptor
Difficulty Level | 40 |
Predictor Variables | 40 |
Test Items | 40 |
Foreign Countries | 13 |
Scores | 12 |
Item Analysis | 9 |
Item Response Theory | 9 |
Mathematics Tests | 9 |
Test Construction | 9 |
Accuracy | 8 |
Elementary School Students | 8 |
More ▼ |
Source
Author
Ferrara, Steve | 2 |
Petscher, Yaacov | 2 |
Schatschneider, Christopher | 2 |
Wood, Carla | 2 |
Ali, Syed Haris | 1 |
Ann E. Harman | 1 |
Antoniou, Panayiotis | 1 |
Ardoin, Scott P. | 1 |
Baird, Jo-Anne | 1 |
Becker, Anthony | 1 |
Belton-Kocher, Evelyn | 1 |
More ▼ |
Publication Type
Education Level
Audience
Policymakers | 1 |
Researchers | 1 |
Teachers | 1 |
Location
Taiwan | 3 |
Florida | 2 |
Kansas | 2 |
Massachusetts | 2 |
Turkey | 2 |
Alabama | 1 |
Arkansas | 1 |
Bosnia and Herzegovina | 1 |
Brazil | 1 |
Cyprus | 1 |
Indiana | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Emily K. Toutkoushian; Huaping Sun; Mark T. Keegan; Ann E. Harman – Measurement: Interdisciplinary Research and Perspectives, 2024
Linear logistic test models (LLTMs), leveraging item response theory and linear regression, offer an elegant method for learning about item characteristics in complex content areas. This study used LLTMs to model single-best-answer, multiple-choice-question response data from two medical subspecialty certification examinations in multiple years…
Descriptors: Licensing Examinations (Professions), Certification, Medical Students, Test Items
Tibbits, Nicole; Lancaster, Hope Sparks; de Diego-Lázaroc, Beatriz – Language, Speech, and Hearing Services in Schools, 2023
Purpose: This study examined the effect of phonological overlap on English and Spanish expressive vocabulary accuracy as measured by the bilingual Expressive One-Word Picture Vocabulary Test--Fourth Edition (EOWPVT-IV). We hypothesized that if languages interact during an expressive vocabulary task, then higher phonological overlap will predict…
Descriptors: Phonology, English, Spanish, Bilingual Students
Petscher, Yaacov; Compton, Donald L.; Steacy, Laura; Kinnon, Hannah – Annals of Dyslexia, 2020
Models of word reading that simultaneously take into account item-level and person-level fixed and random effects are broadly known as explanatory item response models (EIRM). Although many variants of the EIRM are available, the field has generally focused on the doubly explanatory model for modeling individual differences on item responses.…
Descriptors: Item Response Theory, Reading Skills, Individual Differences, Models
Ferrara, Steve; Steedle, Jeffrey T.; Frantz, Roger S. – Applied Measurement in Education, 2022
Item difficulty modeling studies involve (a) hypothesizing item features, or item response demands, that are likely to predict item difficulty with some degree of accuracy; and (b) entering the features as independent variables into a regression equation or other statistical model to predict difficulty. In this review, we report findings from 13…
Descriptors: Reading Comprehension, Reading Tests, Test Items, Item Response Theory
Esin Yilmaz Kogar; Sumeyra Soysal – International Journal of Assessment Tools in Education, 2023
In this paper, it is aimed to evaluate different aspects of students' response time to items in the mathematics test and their test effort as an indicator of test motivation with the help of some variables at the item and student levels. The data consists of 4th-grade Singapore and Turkish students participating in the TIMSS 2019. Response time…
Descriptors: Foreign Countries, Achievement Tests, International Assessment, Mathematics Achievement
Becker, Anthony; Nekrasova-Beker, Tatiana – Educational Assessment, 2018
While previous research has identified numerous factors that contribute to item difficulty, studies involving large-scale reading tests have provided mixed results. This study examined five selected-response item types used to measure reading comprehension in the Pearson Test of English Academic: a) multiple-choice (choose one answer), b)…
Descriptors: Reading Comprehension, Test Items, Reading Tests, Test Format
Ilhan, Mustafa; Öztürk, Nagihan Boztunç; Sahin, Melek Gülsah – Participatory Educational Research, 2020
In this research, the effect of an item's type and cognitive level on its difficulty index was investigated. The data source of the study consisted of the responses of the 12535 students in the Turkey sample (6079 and 6456 students from eighth and fourth grade respectively) of TIMSS 2015. The responses were a total of 215 items at the eighth-grade…
Descriptors: Test Items, Difficulty Level, Cognitive Processes, Responses
Liou, Pey-Yan; Bulut, Okan – Research in Science Education, 2020
The purpose of this study was to examine eighth-grade students' science performance in terms of two test design components, item format, and cognitive domain. The portion of Taiwanese data came from the 2011 administration of the Trends in International Mathematics and Science Study (TIMSS), one of the major international large-scale assessments…
Descriptors: Foreign Countries, Middle School Students, Grade 8, Science Achievement
Tremblay, Kathryn A.; Binder, Katherine S.; Ardoin, Scott P.; Talwar, Amani; Tighe, Elizabeth L. – Journal of Research in Reading, 2021
Background: Of the myriad of reading comprehension (RC) assessments used in schools, multiple-choice (MC) questions continue to be one of the most prevalent formats used by educators and researchers. Outcomes from RC assessments dictate many critical factors encountered during a student's academic career, and it is crucial that we gain a deeper…
Descriptors: Grade 3, Elementary School Students, Reading Comprehension, Decoding (Reading)
Wood, Carla; Schatschneider, Christopher – Journal of Speech, Language, and Hearing Research, 2019
Purpose: This study examines the response patterns of 278 Spanish-English dual language learners (DLLs) on a standardized test of receptive English vocabulary. Method: Investigators analyzed responses to 131 items on the Peabody Picture Vocabulary Test--Fourth Edition (Dunn & Dunn, 2007) focusing on differential accuracy on items influenced by…
Descriptors: Spanish, English, Receptive Language, Vocabulary
Susan Rowe – ProQuest LLC, 2023
This dissertation explored whether unnecessary linguistic complexity (LC) in mathematics and biology assessment items changes the direction and significance of differential item functioning (DIF) between subgroups emergent bilinguals (EBs) and English proficient students (EPs). Due to inconsistencies in measuring LC in items, Study One adapted a…
Descriptors: Difficulty Level, English for Academic Purposes, Second Language Learning, Second Language Instruction
Wood, Carla; Hoge, Rachel; Schatschneider, Christopher; Castilla-Earls, Anny – International Journal of Bilingual Education and Bilingualism, 2021
This study examines the response patterns of 288 Spanish-English dual language learners on a standardized test of receptive Spanish vocabulary. Investigators analyzed responses to 54 items on the "Test de Vocabulario en Imagenes" (TVIP) [Dunn, L. M., D. E. Lugo, E. R. Padilla, and L. M. Dunn. 1986. "Test de Vocabulario en Imganes…
Descriptors: Predictor Variables, Phonology, Item Analysis, Spanish
Jensen, Isabel Nadine; Slabakova, Roumyana; Westergaard, Marit; Lundquist, Björn – Second Language Research, 2020
The Bottleneck Hypothesis (Slabakova, 2008, 2013) proposes that acquiring properties of the functional morphology is the most challenging part of learning a second language. In the experiment presented here, the predictions of this hypothesis are tested in the second language (L2) English of Norwegian native speakers. Two constructions are…
Descriptors: Linguistic Theory, English (Second Language), Second Language Learning, Morphology (Languages)
Predicting Item Difficulty of Science National Curriculum Tests: The Case of Key Stage 2 Assessments
El Masri, Yasmine H.; Ferrara, Steve; Foltz, Peter W.; Baird, Jo-Anne – Curriculum Journal, 2017
Predicting item difficulty is highly important in education for both teachers and item writers. Despite identifying a large number of explanatory variables, predicting item difficulty remains a challenge in educational assessment with empirical attempts rarely exceeding 25% of variance explained. This paper analyses 216 science items of key stage…
Descriptors: Predictor Variables, Test Items, Difficulty Level, Test Construction
Golino, Hudson F.; Gomes, Cristiano M. A. – International Journal of Research & Method in Education, 2016
This paper presents a non-parametric imputation technique, named random forest, from the machine learning field. The random forest procedure has two main tuning parameters: the number of trees grown in the prediction and the number of predictors used. Fifty experimental conditions were created in the imputation procedure, with different…
Descriptors: Item Response Theory, Regression (Statistics), Difficulty Level, Goodness of Fit