Publication Date
| In 2026 | 0 |
| Since 2025 | 197 |
| Since 2022 (last 5 years) | 1067 |
| Since 2017 (last 10 years) | 2577 |
| Since 2007 (last 20 years) | 4938 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Liu, Jinghua; Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2014
Maintaining score interchangeability and scale consistency is crucial for any testing programs that administer multiple forms across years. The use of a multiple linking design, which involves equating a new form to multiple old forms and averaging the conversions, has been proposed to control scale drift. However, the use of multiple linking…
Descriptors: Comparative Analysis, Reliability, Test Construction, Equated Scores
Joan Herman; Robert Linn – Educational Leadership, 2014
Researching. Synthesizing. Reasoning with evidence. The PARCC and Smarter Balanced assessments are clearly setting their sights on complex thinking skills. Researchers Joan Herman and Robert Linn look at the new assessments to see how they stack up against Norman Webb's depth of knowledge framework as well as against current state tests. The…
Descriptors: Student Evaluation, Evaluation Methods, Test Items, Knowledge Level
Kopriva, Rebecca J. – Applied Measurement in Education, 2014
In this commentary, Rebecca Kopriva examines the articles in this special issue by drawing on her experience from three series of investigations examining how English language learners (ELLs) and other students perceive what test items ask and how they can successfully represent what they know. The first series examined the effect of different…
Descriptors: English Language Learners, Test Items, Educational Assessment, Access to Education
Brandriet, Alexandra; Holme, Thomas – Journal of Chemical Education, 2015
The American Chemical Society Examinations Institute (ACS-EI) has recently developed the Exams Data Analysis Spread (EDAS) as a tool to help instructors conduct customizable analyses of their student data from ACS exams. The EDAS calculations allow instructors to analyze their students' performances both at the total score and individual item…
Descriptors: Spreadsheets, Data Analysis, Chemistry, Science Tests
Pachai, Matthew V.; DiBattista, David; Kim, Joseph A. – Canadian Journal for the Scholarship of Teaching and Learning, 2015
Multiple choice writing guidelines are decidedly split on the use of "none of the above" (NOTA), with some authors discouraging and others advocating its use. Moreover, empirical studies of NOTA have produced mixed results. Generally, these studies have utilized NOTA as either the correct response or a distractor and assessed its effect…
Descriptors: Multiple Choice Tests, Test Items, Introductory Courses, Psychology
Gierl, Mark J.; Lai, Hollis; Hogan, James B.; Matovinovic, Donna – Journal of Applied Testing Technology, 2015
The demand for test items far outstrips the current supply. This increased demand can be attributed, in part, to the transition to computerized testing, but, it is also linked to dramatic changes in how 21st century educational assessments are designed and administered. One way to address this growing demand is with automatic item generation.…
Descriptors: Common Core State Standards, Test Items, Alignment (Education), Test Construction
Benjamin, Shanti Isabelle – Education Research and Perspectives, 2015
This study investigates the typical textual features that are most frequently targeted in short-answer reading comprehension questions of the Cambridge University "O" Level English Paper 2. Test writers' awareness of how textual features impact on understanding of meanings in text decisions will determine to great extent their decisions…
Descriptors: Reading Comprehension, Reading Tests, Taxonomy, Text Structure
Zumbo, Bruno D.; Liu, Yan; Wu, Amery D.; Shear, Benjamin R.; Olvera Astivia, Oscar L.; Ark, Tavinder K. – Language Assessment Quarterly, 2015
Methods for detecting differential item functioning (DIF) and item bias are typically used in the process of item analysis when developing new measures; adapting existing measures for different populations, languages, or cultures; or more generally validating test score inferences. In 2007 in "Language Assessment Quarterly," Zumbo…
Descriptors: Test Bias, Test Items, Holistic Approach, Models
Han, Kyung T.; Wells, Craig S.; Hambleton, Ronald K. – Practical Assessment, Research & Evaluation, 2015
In item response theory test scaling/equating with the three-parameter model, the scaling coefficients A and B have no impact on the c-parameter estimates of the test items since the cparameter estimates are not adjusted in the scaling/equating procedure. The main research question in this study concerned how serious the consequences would be if…
Descriptors: Item Response Theory, Monte Carlo Methods, Scaling, Test Items
Reynolds, Matthew R.; Niileksela, Christopher R. – Journal of Psychoeducational Assessment, 2015
"The Woodcock-Johnson IV Tests of Cognitive Abilities" (WJ IV COG) is an individually administered measure of psychometric intellectual abilities designed for ages 2 to 90+. The measure was published by Houghton Mifflin Harcourt-Riverside in 2014. Frederick Shrank, Kevin McGrew, and Nancy Mather are the authors. Richard Woodcock, the…
Descriptors: Cognitive Tests, Testing, Scoring, Test Interpretation
Dodeen, Hamzeh – Journal of Psychoeducational Assessment, 2015
The purpose of this study was to evaluate the factor structure of the University of California, Los Angeles (UCLA) Loneliness Scale and examine possible wording effects on a sample of 1,429 students from the United Arab Emirates University. Correlated traits-correlated uniqueness as well as correlated traits-correlated methods were used to examine…
Descriptors: Affective Measures, Test Items, Factor Structure, College Students
Seligin, Davy; Ishak, Mohd. Zaki; Goropos, Jominin – Journal of Education and Training Studies, 2015
This preliminary study was conducted in interior of Malaysia. The study employed the Body Coordination Two Tier Concept Test to assess students' understandings of body coordination topic. The test consisted of 14 questions as two tier question style. The students needed to select the answer and write down the justifications of each answer.…
Descriptors: Foreign Countries, Interviews, Misconceptions, Secondary School Students
Leis, Micela; Schmidt, Karen M.; Rimm-Kaufman, Sara E. – Grantee Submission, 2015
The Student Engagement in Mathematics Scale (SEMS) is a self-report measure that was created to assess three dimensions of student engagement (social, emotional, and cognitive) in mathematics based on a single day of class. In the current study, the SEMS was administered to a sample of 360 fifth graders from a large Mid-Atlantic district. The…
Descriptors: Learner Engagement, Measures (Individuals), Psychometrics, Test Items
Retnawati, Heri; Kartowagiran, Badrun; Arlinwibowo, Janu; Sulistyaningsih, Eny – International Journal of Instruction, 2017
The quality of national examination items plays an enormous role in identifying students' competencies mastery and their difficulties. This study aims to identify the difficult items in the Junior High School Mathematics National Examination, to find the factors that cause students' difficulty and to reveal the strategies that the teachers and the…
Descriptors: Standardized Tests, National Standards, Mathematics Tests, Test Items
Undersander, Molly A.; Kettler, Richard M.; Stains, Marilyne – Journal of Geoscience Education, 2017
Concept inventories have been determined to be useful assessment tools for evaluating students' knowledge, particularly in the sciences. However, these assessment tools must be validated to reflect as accurately as possible students' understanding of concepts. One possible threat to this validation is what previous literature calls the item order…
Descriptors: Earth Science, Test Items, Scientific Concepts, Science Tests

Peer reviewed
Direct link
