Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 38 |
Since 2006 (last 20 years) | 83 |
Descriptor
Multiple Choice Tests | 104 |
Statistical Analysis | 104 |
Test Items | 104 |
Foreign Countries | 37 |
Difficulty Level | 35 |
Test Construction | 28 |
Item Analysis | 27 |
Comparative Analysis | 22 |
Item Response Theory | 22 |
Test Format | 21 |
Correlation | 16 |
More ▼ |
Source
Author
Alonzo, Julie | 3 |
Frary, Robert B. | 3 |
Tindal, Gerald | 3 |
Malau-Aduli, Bunmi S. | 2 |
Meijer, Rob R. | 2 |
Sotaridona, Leonardo S. | 2 |
Sunbul, Onder | 2 |
Yormaz, Seha | 2 |
Zimitat, Craig | 2 |
van der Linden, Wim J. | 2 |
AbManan, Norhafizah | 1 |
More ▼ |
Publication Type
Education Level
Higher Education | 33 |
Postsecondary Education | 30 |
Secondary Education | 16 |
Elementary Education | 14 |
Middle Schools | 9 |
Grade 5 | 6 |
High Schools | 6 |
Junior High Schools | 6 |
Elementary Secondary Education | 4 |
Grade 8 | 4 |
Grade 9 | 4 |
More ▼ |
Audience
Researchers | 2 |
Location
Turkey | 6 |
Australia | 5 |
Japan | 4 |
Netherlands | 4 |
Nigeria | 3 |
California | 2 |
Taiwan | 2 |
Bosnia and Herzegovina | 1 |
Bosnia and Herzegovina… | 1 |
Canada | 1 |
China | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Kim, Seonghoon; Kolen, Michael J. – Applied Measurement in Education, 2019
In applications of item response theory (IRT), fixed parameter calibration (FPC) has been used to estimate the item parameters of a new test form on the existing ability scale of an item pool. The present paper presents an application of FPC to multiple examinee groups test data that are linked to the item pool via anchor items, and investigates…
Descriptors: Item Response Theory, Item Banks, Test Items, Computation
Sunbul, Onder; Yormaz, Seha – International Journal of Evaluation and Research in Education, 2018
In this study Type I Error and the power rates of omega (?) and GBT (generalized binomial test) indices were investigated for several nominal alpha levels and for 40 and 80-item test lengths with 10,000-examinee sample size under several test level restrictions. As a result, Type I error rates of both indices were found to be below the acceptable…
Descriptors: Difficulty Level, Cheating, Duplication, Test Length
Sunbul, Onder; Yormaz, Seha – Eurasian Journal of Educational Research, 2018
Purpose: Several studies can be found in the literature that investigate the performance of ? under various conditions. However no study for the effects of item difficulty, item discrimination, and ability restrictions on the performance of ? could be found. The current study aims to investigate the performance of ? for the conditions given below.…
Descriptors: Test Items, Difficulty Level, Ability, Cheating
Paul J. Walter; Edward Nuhfer; Crisel Suarez – Numeracy, 2021
We introduce an approach for making a quantitative comparison of the item response curves (IRCs) of any two populations on a multiple-choice test instrument. In this study, we employ simulated and actual data. We apply our approach to a dataset of 12,187 participants on the 25-item Science Literacy Concept Inventory (SLCI), which includes ample…
Descriptors: Item Analysis, Multiple Choice Tests, Simulation, Data Analysis
Quesen, Sarah; Lane, Suzanne – Applied Measurement in Education, 2019
This study examined the effect of similar vs. dissimilar proficiency distributions on uniform DIF detection on a statewide eighth grade mathematics assessment. Results from the similar- and dissimilar-ability reference groups with an SWD focal group were compared for four models: logistic regression, hierarchical generalized linear model (HGLM),…
Descriptors: Test Items, Mathematics Tests, Grade 8, Item Response Theory
Kalkan, Ömür Kaya; Kara, Yusuf; Kelecioglu, Hülya – International Journal of Assessment Tools in Education, 2018
Missing data is a common problem in datasets that are obtained by administration of educational and psychological tests. It is widely known that existence of missing observations in data can lead to serious problems such as biased parameter estimates and inflation of standard errors. Most of the missing data imputation methods are focused on…
Descriptors: Item Response Theory, Statistical Analysis, Data, Test Items
Schultz, Madeleine; Lawrie, Gwendolyn A.; Bailey, Chantal H.; Bedford, Simon B.; Dargaville, Tim R.; O'Brien, Glennys; Tasker, Roy; Thompson, Christopher D.; Williams, Mark; Wright, Anthony H. – International Journal of Science Education, 2017
A multi-institution collaborative team of Australian chemistry education researchers, teaching a total of over 3000 first year chemistry students annually, has explored a tool for diagnosing students' prior conceptions as they enter tertiary chemistry courses. Five core topics were selected and clusters of diagnostic items were assembled linking…
Descriptors: Foreign Countries, College Freshmen, Chemistry, Diagnostic Tests
Answer Changing in Testing Situations: The Role of Metacognition in Deciding Which Answers to Review
Stylianou-Georgiou, Agni; Papanastasiou, Elena C. – Educational Research and Evaluation, 2017
The purpose of our study was to examine the issue of answer changing in relation to students' abilities to monitor their behaviour accurately while responding to multiple-choice tests. The data for this study were obtained from the final examination administered to students in an educational psychology course. The results of the study indicate…
Descriptors: Role, Metacognition, Testing, Multiple Choice Tests
Çakir, Sinan – Journal of Language and Linguistic Studies, 2017
The present study is a follow-up study of Çakir (2016b) which focused on the wh-adverbial & which NP constructions asymmetry within island structures in Turkish. The characteristics of wh-adverbial nasil "how" is compared with the which-NP constructions "hangisekilde" "in what way" and "hangihalde"…
Descriptors: Nouns, Phrase Structure, Grammar, Turkish
Bulut, Okan; Quo, Qi; Gierl, Mark J. – Large-scale Assessments in Education, 2017
Position effects may occur in both paper--pencil tests and computerized assessments when examinees respond to the same test items located in different positions on the test. To examine position effects in large-scale assessments, previous studies often used multilevel item response models within the generalized linear mixed modeling framework.…
Descriptors: Structural Equation Models, Educational Assessment, Measurement, Test Items
Pawade, Yogesh R.; Diwase, Dipti S. – Journal of Educational Technology, 2016
Item analysis of Multiple Choice Questions (MCQs) is the process of collecting, summarizing and utilizing information from students' responses to evaluate the quality of test items. Difficulty Index (p-value), Discrimination Index (DI) and Distractor Efficiency (DE) are the parameters which help to evaluate the quality of MCQs used in an…
Descriptors: Test Items, Item Analysis, Multiple Choice Tests, Curriculum Development
Liu, Ming; Rus, Vasile; Liu, Li – IEEE Transactions on Learning Technologies, 2018
Automatic question generation can help teachers to save the time necessary for constructing examination papers. Several approaches were proposed to automatically generate multiple-choice questions for vocabulary assessment or grammar exercises. However, most of these studies focused on generating questions in English with a certain similarity…
Descriptors: Multiple Choice Tests, Regression (Statistics), Test Items, Natural Language Processing
Ting, Mu Yu – EURASIA Journal of Mathematics, Science & Technology Education, 2017
Using the capabilities of expert knowledge structures, the researcher prepared test questions on the university calculus topic of "finding the area by integration." The quiz is divided into two types of multiple choice items (one out of four and one out of many). After the calculus course was taught and tested, the results revealed that…
Descriptors: Calculus, Mathematics Instruction, College Mathematics, Multiple Choice Tests
Büyükturan, Esin Bagcan; Sireci, Ayse – Journal of Education and Training Studies, 2018
Item discrimination index, which indicates the ability of the item to distinguish whether or not the individuals have acquired the qualities that are evaluated, is basically a validity measure and it is estimated by examining the fit between item score and the test score. Based on the definition of item discrimination index, classroom observation…
Descriptors: Foreign Countries, Classroom Observation Techniques, Scores, Test Items
Ganzfried, Sam; Yusuf, Farzana – Education Sciences, 2018
A problem faced by many instructors is that of designing exams that accurately assess the abilities of the students. Typically, these exams are prepared several days in advance, and generic question scores are used based on rough approximation of the question difficulty and length. For example, for a recent class taught by the author, there were…
Descriptors: Weighted Scores, Test Construction, Student Evaluation, Multiple Choice Tests