Publication Date
In 2025 | 1 |
Since 2024 | 4 |
Since 2021 (last 5 years) | 15 |
Since 2016 (last 10 years) | 32 |
Since 2006 (last 20 years) | 51 |
Descriptor
Responses | 104 |
Test Format | 104 |
Test Items | 64 |
Multiple Choice Tests | 37 |
Test Construction | 24 |
Difficulty Level | 16 |
Computer Assisted Testing | 15 |
Foreign Countries | 15 |
Mathematics Tests | 15 |
Scoring | 15 |
Higher Education | 14 |
More ▼ |
Source
Author
Kim, Sooyeon | 5 |
Walker, Michael E. | 4 |
McHale, Frederick | 3 |
DeMars, Christine E. | 2 |
Guo, Wenjing | 2 |
Katz, Irvin R. | 2 |
Wind, Stefanie A. | 2 |
A. Lopez, Alexis | 1 |
Achim Goerres | 1 |
Ackerman, Terry | 1 |
Ackermann, Nicole | 1 |
More ▼ |
Publication Type
Education Level
Audience
Researchers | 2 |
Practitioners | 1 |
Teachers | 1 |
Location
Germany | 3 |
Turkey | 2 |
California | 1 |
Canada | 1 |
Chile (Santiago) | 1 |
Europe | 1 |
Florida | 1 |
Illinois | 1 |
Illinois (Chicago) | 1 |
Israel (Tel Aviv) | 1 |
Nebraska | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Stefanie A. Wind; Yuan Ge – Measurement: Interdisciplinary Research and Perspectives, 2024
Mixed-format assessments made up of multiple-choice (MC) items and constructed response (CR) items that are scored using rater judgments include unique psychometric considerations. When these item types are combined to estimate examinee achievement, information about the psychometric quality of each component can depend on that of the other. For…
Descriptors: Interrater Reliability, Test Bias, Multiple Choice Tests, Responses
Brian E. Clauser; Victoria Yaneva; Peter Baldwin; Le An Ha; Janet Mee – Applied Measurement in Education, 2024
Multiple-choice questions have become ubiquitous in educational measurement because the format allows for efficient and accurate scoring. Nonetheless, there remains continued interest in constructed-response formats. This interest has driven efforts to develop computer-based scoring procedures that can accurately and efficiently score these items.…
Descriptors: Computer Uses in Education, Artificial Intelligence, Scoring, Responses
Jiawei Xiong; George Engelhard; Allan S. Cohen – Measurement: Interdisciplinary Research and Perspectives, 2025
It is common to find mixed-format data results from the use of both multiple-choice (MC) and constructed-response (CR) questions on assessments. Dealing with these mixed response types involves understanding what the assessment is measuring, and the use of suitable measurement models to estimate latent abilities. Past research in educational…
Descriptors: Responses, Test Items, Test Format, Grade 8
Jan Karem Höhne; Achim Goerres – International Journal of Social Research Methodology, 2024
The measurement of political solidarities and related concepts is an important endeavor in numerous scientific disciplines, such as political and social science research. European surveys, such as the Eurobarometer, frequently measure these concepts for people's home country and Europe raising questions with respect to the order of precedence.…
Descriptors: Surveys, Attitude Measures, Political Attitudes, Foreign Countries
Gerd Kortemeyer – Physical Review Physics Education Research, 2023
Solving problems is crucial for learning physics, and not only final solutions but also their derivations are important. Grading these derivations is labor intensive, as it generally involves human evaluation of handwritten work. AI tools have not been an alternative, since even for short answers, they needed specific training for each problem or…
Descriptors: Artificial Intelligence, Problem Solving, Physics, Introductory Courses
Glazer, Nancy; Wolfe, Edward W. – Applied Measurement in Education, 2020
This introductory article describes how constructed response scoring is carried out, particularly the rater monitoring processes and illustrates three potential designs for conducting rater monitoring in an operational scoring project. The introduction also presents a framework for interpreting research conducted by those who study the constructed…
Descriptors: Scoring, Test Format, Responses, Predictor Variables
Guo, Wenjing; Wind, Stefanie A. – Journal of Educational Measurement, 2021
The use of mixed-format tests made up of multiple-choice (MC) items and constructed response (CR) items is popular in large-scale testing programs, including the National Assessment of Educational Progress (NAEP) and many district- and state-level assessments in the United States. Rater effects, or raters' scoring tendencies that result in…
Descriptors: Test Format, Multiple Choice Tests, Scoring, Test Items
Wind, Stefanie A.; Guo, Wenjing – Educational Assessment, 2021
Scoring procedures for the constructed-response (CR) items in large-scale mixed-format educational assessments often involve checks for rater agreement or rater reliability. Although these analyses are important, researchers have documented rater effects that persist despite rater training and that are not always detected in rater agreement and…
Descriptors: Scoring, Responses, Test Items, Test Format
Smyth, Jolene D.; Israel, Glenn D.; Newberry, Milton G.; Hull, Richard G. – Field Methods, 2019
Considerable research has examined the effect of response option order in ordinal bipolar questions such as satisfaction questions. However, no research we know of has examined the effect of the order of presentation of concepts in the question stem or whether stem order moderates response option order. In this article, we experimentally test the…
Descriptors: Satisfaction, Responses, Test Items, Attitude Measures
Gurdil Ege, Hatice; Demir, Ergul – Eurasian Journal of Educational Research, 2020
Purpose: The present study aims to evaluate how the reliabilities computed using a, Stratified a, Angoff-Feldt, and Feldt-Raju estimators may differ when sample size (500, 1000, and 2000) and item type ratio of dichotomous to polytomous items (2:1; 1:1, 1:2) included in the scale are varied. Research Methods: In this study, Cronbach's a,…
Descriptors: Test Format, Simulation, Test Reliability, Sample Size
Mo, Ya; Carney, Michele; Cavey, Laurie; Totorica, Tatia – Applied Measurement in Education, 2021
There is a need for assessment items that assess complex constructs but can also be efficiently scored for evaluation of teacher education programs. In an effort to measure the construct of teacher attentiveness in an efficient and scalable manner, we are using exemplar responses elicited by constructed-response item prompts to develop…
Descriptors: Protocol Analysis, Test Items, Responses, Mathematics Teachers
Wang, Yu; Chiu, Chia-Yi; Köhn, Hans Friedrich – Journal of Educational and Behavioral Statistics, 2023
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Nonparametric Statistics, Test Format, Educational Assessment
Baral, Sami; Botelho, Anthony; Santhanam, Abhishek; Gurung, Ashish; Cheng, Li; Heffernan, Neil – International Educational Data Mining Society, 2023
Teachers often rely on the use of a range of open-ended problems to assess students' understanding of mathematical concepts. Beyond traditional conceptions of student open-ended work, commonly in the form of textual short-answer or essay responses, the use of figures, tables, number lines, graphs, and pictographs are other examples of open-ended…
Descriptors: Mathematics Instruction, Mathematical Concepts, Problem Solving, Test Format
Arslan, Burcu; Jiang, Yang; Keehner, Madeleine; Gong, Tao; Katz, Irvin R.; Yan, Fred – Educational Measurement: Issues and Practice, 2020
Computer-based educational assessments often include items that involve drag-and-drop responses. There are different ways that drag-and-drop items can be laid out and different choices that test developers can make when designing these items. Currently, these decisions are based on experts' professional judgments and design constraints, rather…
Descriptors: Test Items, Computer Assisted Testing, Test Format, Decision Making
Bingölbali, Erhan; Bingölbali, Ferhan – International Journal of Progressive Education, 2021
This study explores the affordances that the open-ended questions hold in comparison with those of closed-ended questions through examining 6th grade students' performance on a mathematics test. For this purpose, a questionnaire including 2 open-ended and 2 closed-ended questions was applied to 36 6th grade students. The questions were prepared in…
Descriptors: Questioning Techniques, Test Format, Test Items, Responses