Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 6 |
Since 2006 (last 20 years) | 12 |
Descriptor
Difficulty Level | 19 |
Multiple Choice Tests | 19 |
Standardized Tests | 19 |
Test Items | 16 |
Item Analysis | 7 |
Test Construction | 7 |
Item Response Theory | 6 |
Reading Comprehension | 6 |
Mathematics Tests | 5 |
Test Format | 5 |
Foreign Countries | 4 |
More ▼ |
Source
Author
Alicia A. Stoltenberg | 1 |
Alonzo, Julie | 1 |
Apino, Ezi | 1 |
Binici, Salih | 1 |
Bolt, Daniel M. | 1 |
Bucak, S. Deniz | 1 |
Córdova, Nora | 1 |
Dartnell, Pablo | 1 |
Ehrich, John | 1 |
Eng, Lin Siew | 1 |
Godoy, María Inés | 1 |
More ▼ |
Publication Type
Reports - Research | 15 |
Journal Articles | 9 |
Guides - Non-Classroom | 2 |
Speeches/Meeting Papers | 2 |
Dissertations/Theses -… | 1 |
Numerical/Quantitative Data | 1 |
Reports - Evaluative | 1 |
Education Level
Elementary Education | 4 |
Secondary Education | 4 |
Early Childhood Education | 2 |
Grade 3 | 2 |
Grade 7 | 2 |
Higher Education | 2 |
Middle Schools | 2 |
Primary Education | 2 |
Elementary Secondary Education | 1 |
Grade 12 | 1 |
Grade 4 | 1 |
More ▼ |
Audience
Administrators | 1 |
Laws, Policies, & Programs
Elementary and Secondary… | 1 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Alicia A. Stoltenberg – ProQuest LLC, 2024
Multiple-select multiple-choice items, or multiple-choice items with more than one correct answer, are used to quickly assess content on standardized assessments. Because there are multiple keys to these item types, there are also multiple ways to score student responses to these items. The purpose of this study was to investigate how changing the…
Descriptors: Scoring, Evaluation Methods, Multiple Choice Tests, Standardized Tests
Lions, Séverin; Dartnell, Pablo; Toledo, Gabriela; Godoy, María Inés; Córdova, Nora; Jiménez, Daniela; Lemarié, Julie – Educational and Psychological Measurement, 2023
Even though the impact of the position of response options on answers to multiple-choice items has been investigated for decades, it remains debated. Research on this topic is inconclusive, perhaps because too few studies have obtained experimental data from large-sized samples in a real-world context and have manipulated the position of both…
Descriptors: Multiple Choice Tests, Test Items, Item Analysis, Responses
Rafi, Ibnu; Retnawati, Heri; Apino, Ezi; Hadiana, Deni; Lydiati, Ida; Rosyada, Munaya Nikma – Pedagogical Research, 2023
This study describes the characteristics of the test and its items used in the national-standardized school examination by applying classical test theory and focusing on the item difficulty, item discrimination, test reliability, and distractor analysis. We analyzed response data of 191 12th graders from one of public senior high schools in…
Descriptors: Foreign Countries, National Competency Tests, Standardized Tests, Mathematics Tests
Liao, Xiangyi; Bolt, Daniel M. – Journal of Educational and Behavioral Statistics, 2021
Four-parameter models have received increasing psychometric attention in recent years, as a reduced upper asymptote for item characteristic curves can be appealing for measurement applications such as adaptive testing and person-fit assessment. However, applications can be challenging due to the large number of parameters in the model. In this…
Descriptors: Test Items, Models, Mathematics Tests, Item Response Theory
Woodcock, Stuart; Howard, Steven J.; Ehrich, John – School Psychology, 2020
Standardized testing is ubiquitous in educational assessment, but questions have been raised about the extent to which these test scores accurately reflect students' genuine knowledge and skills. To more rigorously investigate this issue, the current study employed a within-subject experimental design to examine item format effects on primary…
Descriptors: Elementary School Students, Grade 3, Test Items, Test Format
Susanti, Yuni; Tokunaga, Takenobu; Nishikawa, Hitoshi; Obari, Hiroyuki – Research and Practice in Technology Enhanced Learning, 2017
The present study investigates the best factor for controlling the item difficulty of multiple-choice English vocabulary questions generated by an automatic question generation system. Three factors are considered for controlling item difficulty: (1) reading passage difficulty, (2) semantic similarity between the correct answer and distractors,…
Descriptors: Test Items, Difficulty Level, Computer Assisted Testing, Vocabulary Development
Stoffel, Heather; Raymond, Mark R.; Bucak, S. Deniz; Haist, Steven A. – Practical Assessment, Research & Evaluation, 2014
Previous research on the impact of text and formatting changes on test-item performance has produced mixed results. This matter is important because it is generally acknowledged that "any" change to an item requires that it be recalibrated. The present study investigated the effects of seven classes of stylistic changes on item…
Descriptors: Test Construction, Test Items, Standardized Tests, Physicians
Lim, Chang Kuan; Eng, Lin Siew; Mohamed, Abdul Rashid – English Language Teaching, 2014
Reading and understanding a written text is one of the most important skills in English learning.This study attempts to benchmark Year Five students' reading abilities of fifteen rural schools in a district in Malaysia. The objectives of this study are to develop a set of standardised written reading comprehension and a set of indicators to inform…
Descriptors: Foreign Countries, Reading Ability, Benchmarking, Rural Schools
Irvin, P. Shawn; Alonzo, Julie; Lai, Cheng-Fei; Park, Bitnara Jasmine; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the seventh-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Reading Comprehension, Testing Programs, Statistical Analysis, Grade 7
Kobrin, Jennifer L.; Kim, YoungKoung; Sackett, Paul R. – Educational and Psychological Measurement, 2012
There is much debate on the merits and pitfalls of standardized tests for college admission, with questions regarding the format (multiple-choice vs. constructed response), cognitive complexity, and content of these assessments (achievement vs. aptitude) at the forefront of the discussion. This study addressed these questions by investigating the…
Descriptors: Grade Point Average, Standardized Tests, Predictive Validity, Predictor Variables
Powers, Sonya; Turhan, Ahmet; Binici, Salih – Pearson, 2012
The population sensitivity of vertical scaling results was evaluated for a state reading assessment spanning grades 3-10 and a state mathematics test spanning grades 3-8. Subpopulations considered included males and females. The 3-parameter logistic model was used to calibrate math and reading items and a common item design was used to construct…
Descriptors: Scaling, Equated Scores, Standardized Tests, Reading Tests
Rieck, William A. – Principal Leadership, 2006
Student assessment has long been a major component of the tasks that teachers perform. As such, it is important that school leaders consider teachers' assessment strategies as part of the normal supervisory process. In a political climate ruled by the No Child Left Behind Act, one important consideration is how well teachers' assessments prepare…
Descriptors: Federal Legislation, Test Items, Academic Achievement, Standardized Tests
Roberts, Sarah Jane – 1978
This guide explains the concept of out-of-level testing and suggests a formula for estimating the occurrence of floor effects and ceiling effects, within the context of models for evaluating Elementary Secondary Education Act (ESEA) Title I programs. An analogy explains floor and ceiling effects as if test items are stored in different levels in a…
Descriptors: Achievement Tests, Difficulty Level, Educational Testing, Elementary Education
Pike, Gary – 1989
Responses to American College Test College Outcome Measures Program (ACT-COMP) items by 481 black and 9,237 white students at the University of Tennessee (Knoxville) were analyzed using F. Samejima's graded model to determine the level of differential item functioning (DIF). Students had been tested using Form 8 of the ACT-COMP objective test…
Descriptors: Black Students, College Entrance Examinations, College Students, Comparative Testing
Huntley, Renee M.; Plake, Barbara S. – 1988
The combinational-format item (CFI)--multiple-choice item with combinations of alternatives presented as response choices--was studied to determine whether CFIs were different from regular multiple-choice items in item characteristics or in cognitive processing demands. Three undergraduate Foundations of Education classes (consisting of a total of…
Descriptors: Cognitive Processes, Computer Assisted Testing, Difficulty Level, Educational Psychology
Previous Page | Next Page »
Pages: 1 | 2