Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Gierl, Mark J.; Lai, Hollis – Educational Measurement: Issues and Practice, 2016
Testing organization needs large numbers of high-quality items due to the proliferation of alternative test administration methods and modern test designs. But the current demand for items far exceeds the supply. Test items, as they are currently written, evoke a process that is both time-consuming and expensive because each item is written,…
Descriptors: Test Items, Test Construction, Psychometrics, Models
Bridgeman, Brent – Educational Measurement: Issues and Practice, 2016
Scores on essay-based assessments that are part of standardized admissions tests are typically given relatively little weight in admissions decisions compared to the weight given to scores from multiple-choice assessments. Evidence is presented to suggest that more weight should be given to these assessments. The reliability of the writing scores…
Descriptors: Multiple Choice Tests, Scores, Standardized Tests, Comparative Analysis
Couchman, Justin J.; Miller, Noelle E.; Zmuda, Shaun J.; Feather, Kathryn; Schwartzmeyer, Tina – Metacognition and Learning, 2016
Students often gauge their performance before and after an exam, usually in the form of rough grade estimates or general feelings. Are these estimates accurate? Should they form the basis for decisions about study time, test-taking strategies, revisions, subject mastery, or even general competence? In two studies, undergraduates took a real…
Descriptors: Higher Education, College Students, Tests, Metacognition
Rutkowski, Leslie; Rutkowski, David; Zhou, Yan – International Journal of Testing, 2016
Using an empirically-based simulation study, we show that typically used methods of choosing an item calibration sample have significant impacts on achievement bias and system rankings. We examine whether recent PISA accommodations, especially for lower performing participants, can mitigate some of this bias. Our findings indicate that standard…
Descriptors: Simulation, International Programs, Adolescents, Student Evaluation
Vadivu, P. Pandia; Sridhar, R.; Kumar, B. Mohan – Online Submission, 2016
The main objective of the present pilot study was to construct and standardize a Scientific Aptitude Test (SAT) for the secondary school science students (Grade 9 and 10). The test items (30 questions) were prepared by the researchers based on the five important aspects in scientific thinking such as Analogy, Scientific reasoning, Numerical…
Descriptors: Secondary School Students, Pilot Projects, Science Tests, Grade 9
O'Keeffe, Lisa – Mathematics Education Research Group of Australasia, 2016
Language is frequently discussed as barrier to mathematics word problems. Hence this paper presents the initial findings of a linguistic analysis of numeracy skills test sample items. The theoretical perspective of multi-modal text analysis underpinned this study, in which data was extracted from the ten sample numeracy test items released by the…
Descriptors: Numeracy, Mathematics Skills, Test Items, Preservice Teachers
De Backer, Fauve; Baele, Judith; van Avermaet, Piet; Slembrouck, Stef – Language Assessment Quarterly, 2019
One of the greatest assessment challenges is providing fair and valid tests for multilingual pupils. Since language proficiency impacts their results on content tests, accommodations are suggested to help in solving the validity issues. In this mixed methods study on multilingual assessment, pupils were divided according to three testing…
Descriptors: Multilingualism, Testing Accommodations, Student Attitudes, Language Proficiency
Hosseini, Mohammad Baqerzadeh; Pourghasemian, Hossein – Cogent Education, 2019
The present paper explored how plays could contribute to pragmatic development when employed as a medium of implicit or explicit instruction. 80 undergraduate English-major students were divided into four experimental groups two of which were literary and two nonliterary. Implicit Literary, as one of the literary groups was exposed to…
Descriptors: Comparative Analysis, Teaching Methods, Language Tests, Speech Acts
Baghaei, Purya; Ravand, Hamdollah – SAGE Open, 2019
In many reading comprehension tests, different test formats are employed. Two commonly used test formats to measure reading comprehension are sustained passages followed by some questions and cloze items. Individual differences in handling test format peculiarities could constitute a source of score variance. In this study, a bifactor Rasch model…
Descriptors: Cloze Procedure, Test Bias, Individual Differences, Difficulty Level
Zhang, Hanmu; Zhang, Hanmu – Journal of Education and Learning, 2019
Since understanding reading assignments is important to succeeding in school, improving the way that text is arranged in books would be an efficient way to help students better understand the material and perform well on tests. In this study, we asked students to read two original and two rearranged historical passages, in which rephrased…
Descriptors: Test Items, Textbook Preparation, Retention (Psychology), Recall (Psychology)
Li, Sylvia; Meyer, Patrick – NWEA, 2019
This simulation study examines the measurement precision, item exposure rates, and the depth of the MAP® Growth™ item pools under various grade-level restrictions. Unlike most summative assessments, MAP Growth allows examinees to see items from any grade level, regardless of the examinee's actual grade level. It does not limit the test to items…
Descriptors: Achievement Tests, Item Banks, Test Items, Instructional Program Divisions
Mohr, Doris, Ed.; Walcott, Crystal, Ed.; Kloosterman, Peter, Ed. – National Council of Teachers of Mathematics, 2019
"Mathematical Thinking: From Assessment Items to Challenging Tasks" is a compilation of 36 problem-based lessons that encourage students to engage in productive struggle and deep thinking. Its 36 full-length lessons for grades 2-8 are each inspired by an actual test item from the National Assessment of Educational Progress (NAEP).…
Descriptors: Problem Based Learning, Test Items, Elementary School Mathematics, Middle School Mathematics
Gierl, Mark J.; Lai, Hollis – Educational Measurement: Issues and Practice, 2013
Changes to the design and development of our educational assessments are resulting in the unprecedented demand for a large and continuous supply of content-specific test items. One way to address this growing demand is with automatic item generation (AIG). AIG is the process of using item models to generate test items with the aid of computer…
Descriptors: Educational Assessment, Test Items, Automation, Computer Assisted Testing
Kahraman, Nilufer – Journal of Educational Measurement, 2013
This article considers potential problems that can arise in estimating a unidimensional item response theory (IRT) model when some test items are multidimensional (i.e., show a complex factorial structure). More specifically, this study examines (1) the consequences of model misfit on IRT item parameter estimates due to unintended minor item-level…
Descriptors: Test Items, Item Response Theory, Computation, Models
Setzer, J. Carl; Wise, Steven L.; van den Heuvel, Jill R.; Ling, Guangming – Applied Measurement in Education, 2013
Assessment results collected under low-stakes testing situations are subject to effects of low examinee effort. The use of computer-based testing allows researchers to develop new ways of measuring examinee effort, particularly using response times. At the item level, responses can be classified as exhibiting either rapid-guessing behavior or…
Descriptors: Testing, Guessing (Tests), Reaction Time, Test Items

Peer reviewed
Direct link
