Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Krell, Moritz; Samia Khan; Jan van Driel – Education Sciences, 2021
The development and evaluation of valid assessments of scientific reasoning are an integral part of research in science education. In the present study, we used the linear logistic test model (LLTM) to analyze how item features related to text complexity and the presence of visual representations influence the overall item difficulty of an…
Descriptors: Cognitive Processes, Difficulty Level, Science Tests, Logical Thinking
Jiang, Yang; Gong, Tao; Saldivia, Luis E.; Cayton-Hodges, Gabrielle; Agard, Christopher – Large-scale Assessments in Education, 2021
In 2017, the mathematics assessments that are part of the National Assessment of Educational Progress (NAEP) program underwent a transformation shifting the administration from paper-and-pencil formats to digitally-based assessments (DBA). This shift introduced new interactive item types that bring rich process data and tremendous opportunities to…
Descriptors: Data Use, Learning Analytics, Test Items, Measurement
Nese, Joseph F. T.; Kamata, Akihito – School Psychology, 2021
Curriculum-based measurement of oral reading fluency (CBM-R) is widely used across the United States as a strong indicator of comprehension and overall reading achievement, but has several limitations including errors in administration and large standard errors of measurement. The purpose of this study is to compare scoring methods and passage…
Descriptors: Curriculum Based Assessment, Oral Reading, Reading Fluency, Reading Tests
He, Wei – NWEA, 2021
New MAP® Growth™ assessments are being developed that administer items more closely matched to the grade level of the student. However, MAP Growth items are calibrated with samples that typically consist of students from a variety of grades, including the target grade to which an item is aligned. While this choice of calibration sample is…
Descriptors: Achievement Tests, Test Items, Instructional Program Divisions, Difficulty Level
Liotino, Marica; Fedeli, Monica; Garone, Anja; Knorn, Steffi; Varagnolo, Damiano; Garone, Emanuele – Commission for International Adult Education, 2021
Formally describing and assessing the difficulty of learning and teaching material is important for quality assurance in university teaching, for aligning teaching and learning activities, and for easing communications among stakeholders such as teachers and students. This paper proposes a novel taxonomy to describe and quantify the difficulty…
Descriptors: Taxonomy, Student Evaluation, Engineering Education, Student Projects
Emily Tucker – ProQuest LLC, 2021
To better understand Tennessee's new standardized science assessments, this quantitative study utilized a nonexperimental, descriptive-comparative design to compare the readability of the long-used science TCAP assessment with the newly created science TNReady assessment in grades three, four, and five. As new standards in the state boast higher…
Descriptors: Science Tests, Standardized Tests, Achievement Tests, Readability
Stephanie M. Werner; Ying Chen; Mike Stieff – Grantee Submission, 2021
The Chemistry Self-Concept Inventory (CSCI) is a widely used instrument within chemistry education research. Yet, agreement on its overall reliability and validity is lacking, and psychometric analyses of the instrument remain outstanding. This study examined the psychometric properties of the subscale and item function of the CSCI on 1,140 high…
Descriptors: Self Concept Measures, Chemistry, Psychometrics, Item Response Theory
Villarreal, Victor – Journal of Psychoeducational Assessment, 2019
The "Rating Scale of Impairment" (RSI; Goldstein & Naglieri, 2016b) is a norm-referenced measure of functional impairment. The RSI measures impairment in six domains, as well as overall impairment, based in part on the International Classification of Functioning, Disability, and Health. Functional impairment, as defined by the ICF…
Descriptors: Rating Scales, Norm Referenced Tests, Disabilities, Test Construction
Han, Kyung T.; Dimitrov, Dimiter M.; Al-Mashary, Faisal – Educational and Psychological Measurement, 2019
The "D"-scoring method for scoring and equating tests with binary items proposed by Dimitrov offers some of the advantages of item response theory, such as item-level difficulty information and score computation that reflects the item difficulties, while retaining the merits of classical test theory such as the simplicity of number…
Descriptors: Test Construction, Scoring, Test Items, Adaptive Testing
Kim, Seonghoon; Kolen, Michael J. – Applied Measurement in Education, 2019
In applications of item response theory (IRT), fixed parameter calibration (FPC) has been used to estimate the item parameters of a new test form on the existing ability scale of an item pool. The present paper presents an application of FPC to multiple examinee groups test data that are linked to the item pool via anchor items, and investigates…
Descriptors: Item Response Theory, Item Banks, Test Items, Computation
Rigney, Alexander M. – Journal of Psychoeducational Assessment, 2019
The "Detroit Tests of Learning Aptitude" has been in use for more than three quarters of a century (Baker & Leland, 1935). Its longevity in the field speaks to its popularity as a broad measure of cognitive abilities. Its most recent iteration, in the form of the "Detroit Tests of Learning Abilities--Fifth Edition" (DTLA-5;…
Descriptors: Aptitude Tests, Cognitive Ability, Test Construction, Test Items
Tulek, Onder Kamil; Kose, Ibrahim Alper – Eurasian Journal of Educational Research, 2019
Purpose: This research investigates Tests that include DIF items and which are purified from DIF items. While doing this, the ability estimations and purified DIF items are compared to understand whether there is a correlation between the estimations. Method: The researcher used to R 3.4.1 in order to compare the items and after this situation;…
Descriptors: Test Items, Item Analysis, Item Response Theory, Test Length
Wang, Lin – ETS Research Report Series, 2019
Rearranging response options in different versions of a test of multiple-choice items can be an effective strategy against cheating on the test. This study investigated if rearranging response options would affect item performance and test score comparability. A study test was assembled as the base version from which 3 variant versions were…
Descriptors: Multiple Choice Tests, Test Items, Test Format, Scores
Ip, Edward H.; Strachan, Tyler; Fu, Yanyan; Lay, Alexandra; Willse, John T.; Chen, Shyh-Huei; Rutkowski, Leslie; Ackerman, Terry – Journal of Educational Measurement, 2019
Test items must often be broad in scope to be ecologically valid. It is therefore almost inevitable that secondary dimensions are introduced into a test during test development. A cognitive test may require one or more abilities besides the primary ability to correctly respond to an item, in which case a unidimensional test score overestimates the…
Descriptors: Test Items, Test Bias, Test Construction, Scores
Leo, J.; Kurdi, G.; Matentzoglu, N.; Parsia, B.; Sattler, U.; Forge, S.; Donato, G.; Dowling, W. – International Journal of Artificial Intelligence in Education, 2019
Designing good multiple choice questions (MCQs) for education and assessment is time consuming and error-prone. An abundance of structured and semi-structured data has led to the development of automatic MCQ generation methods. Recently, ontologies have emerged as powerful tools to enable the automatic generation of MCQs. However, current question…
Descriptors: Multiple Choice Tests, Test Items, Automation, Test Construction

Peer reviewed
Direct link
