Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Pawade, Yogesh R.; Diwase, Dipti S. – Journal of Educational Technology, 2016
Item analysis of Multiple Choice Questions (MCQs) is the process of collecting, summarizing and utilizing information from students' responses to evaluate the quality of test items. Difficulty Index (p-value), Discrimination Index (DI) and Distractor Efficiency (DE) are the parameters which help to evaluate the quality of MCQs used in an…
Descriptors: Test Items, Item Analysis, Multiple Choice Tests, Curriculum Development
Qudah, Ahmad Hassan – Journal of Education and Practice, 2016
The research aims to reveal the specific way to evaluate learning mathematics, so that we get the "measuring tool" for the achievement of learners in mathematics that reflect their level of understanding by score (mark), which we trust it with high degree. The behavior of the learner can be measured by a professional way to build the…
Descriptors: Mathematics Instruction, Mathematics Teachers, Student Evaluation, Evaluation Methods
Zhang, Tan; Chen, Ang – AERA Online Paper Repository, 2016
Based on the Job Demands-Resources model, the study developed and validated an instrument that measures physical education teachers' job demands/resources perception. Expert review established content validity with the average item rating of 3.6/5.0. Construct validity and reliability were determined with a teacher sample (n=193). Exploratory…
Descriptors: Physical Education Teachers, Teaching Load, Resources, Measures (Individuals)
Schindler, Julia; Richter, Tobias; Isberner, Maj-Britt; Naumann, Johannes; Neeb, Yvonne – Language Assessment Quarterly, 2018
Reading comprehension is based on the efficient accomplishment of several cognitive processes at the word, sentence, and text level. To the extent that each of these processes contributes to reading comprehension, it can cause reading difficulties if it is deficient. To identify individual sources of reading difficulties, tools are required that…
Descriptors: Construct Validity, Language Tests, Grammar, Task Analysis
Yang, Ji Seung; Zheng, Xiaying – Journal of Educational and Behavioral Statistics, 2018
The purpose of this article is to introduce and review the capability and performance of the Stata item response theory (IRT) package that is available from Stata v.14, 2015. Using a simulated data set and a publicly available item response data set extracted from Programme of International Student Assessment, we review the IRT package from…
Descriptors: Item Response Theory, Item Analysis, Computer Software, Statistical Analysis
Traxler, Adrienne; Henderson, Rachel; Stewart, John; Stewart, Gay; Papak, Alexis; Lindell, Rebecca – Physical Review Physics Education Research, 2018
Research on the test structure of the Force Concept Inventory (FCI) has largely ignored gender, and research on FCI gender effects (often reported as "gender gaps") has seldom interrogated the structure of the test. These rarely crossed streams of research leave open the possibility that the FCI may not be structurally valid across…
Descriptors: Physics, Science Instruction, Sex Fairness, Gender Differences
Liu, Ming; Rus, Vasile; Liu, Li – IEEE Transactions on Learning Technologies, 2018
Automatic question generation can help teachers to save the time necessary for constructing examination papers. Several approaches were proposed to automatically generate multiple-choice questions for vocabulary assessment or grammar exercises. However, most of these studies focused on generating questions in English with a certain similarity…
Descriptors: Multiple Choice Tests, Regression (Statistics), Test Items, Natural Language Processing
Karlin, Omar; Karlin, Sayaka – InSight: A Journal of Scholarly Teaching, 2018
This study had two aims. The first was to explain the process of using the Rasch measurement model to validate tests in an easy-to-understand way for those unfamiliar with the Rasch measurement model. The second was to validate two final exams with several shared items. The exams were given to two groups of students with slightly differing English…
Descriptors: Item Response Theory, Test Validity, Test Items, Accuracy
Reardon, Sean F.; Kalogrides, Demetra; Fahle, Erin M.; Podolsky, Anne; Zárate, Rosalía C. – Educational Researcher, 2018
Prior research suggests that males outperform females, on average, on multiple-choice items compared to their relative performance on constructed-response items. This paper characterizes the extent to which gender achievement gaps on state accountability tests across the United States are associated with those tests' item formats. Using roughly 8…
Descriptors: Test Items, Test Format, Gender Differences, Achievement Gap
Kalthoff, Britta; Theyssen, Heike; Schreiber, Nico – International Journal of Science Education, 2018
Experimental skills should be acquired by learners at school and university alike. To promote experimental skills, various approaches exist within a spectrum between implicit and explicit instruction. Regarding these instructional approaches, numerous findings are available which predominantly relate to pupils. It is an open question whether it is…
Descriptors: Physics, Intervention, Science Instruction, Pretests Posttests
Carrió-Pastor, María Luisa; Martín Marchante, Beatriz – International Journal of English Studies, 2018
The work at hand is part of a wider study the aim of which was to determine what kind of factors influence failures in pragmatic items of an online adaptive placement test taken by a group of 34 Spanish students in their first year at university. A preceding analysis (Carrió & Martín, 2016) showed the type of personal factors, such as lack of…
Descriptors: Pragmatics, Test Items, Language Tests, College Freshmen
Schoen, Robert C.; Yang, Xiaotong; Paek, Insu – Grantee Submission, 2018
This report provides evidence of the substantive and structural validity of the Knowledge for Teaching Elementary Fractions Test. Field-test data were gathered with a sample of 241 elementary educators, including teachers, administrators, and instructional support personnel, in spring 2017, as part of a larger study involving a multisite…
Descriptors: Psychometrics, Pedagogical Content Knowledge, Mathematics Tests, Mathematics Instruction
Foghahaee, Zahra – Language Teaching Research Quarterly, 2019
Reverse engineering (RE) can play an important role in the re-designing tests in L2 English. It can also enrich the aim of teaching the same as raising children through academic achievement. In addition, it can play a key role in helping students understand how much their test is valid by using Standard reverse engineering (SRE). This paper is a…
Descriptors: Language Tests, Second Language Learning, Second Language Instruction, English (Second Language)
Van den Eynde, Sofie; van Kampen, Paul; Van Dooren, Wim; De Cock, Mieke – Physical Review Physics Education Research, 2019
We report on a study investigating the influence of context, direction of translation, and function type on undergraduate students' ability to translate between graphical and symbolic representations of mathematical relations. Students from an algebra-based and a calculus-based physics course were asked to solve multiple-choice items in which they…
Descriptors: Graphs, Equations (Mathematics), Mathematics Instruction, Physics
Nagy, Gabriel; Nagengast, Benjamin; Frey, Andreas; Becker, Michael; Rose, Norman – Assessment in Education: Principles, Policy & Practice, 2019
Position effects (PE) cause decreasing probabilities of correct item responses towards the end of a test. We analysed PEs in science, mathematics and reading tests administered in the German extension to the PISA 2006 study with respect to their variability at the student- and school-level. PEs were strongest in reading and weakest in mathematics.…
Descriptors: Achievement Tests, Foreign Countries, Secondary School Students, International Assessment

Peer reviewed
Direct link
