Publication Date
In 2025 | 2 |
Since 2024 | 7 |
Since 2021 (last 5 years) | 14 |
Since 2016 (last 10 years) | 23 |
Since 2006 (last 20 years) | 45 |
Descriptor
Educational Assessment | 138 |
Test Format | 138 |
Test Construction | 44 |
Student Evaluation | 36 |
Test Items | 34 |
Elementary Secondary Education | 31 |
Test Use | 26 |
Evaluation Methods | 24 |
Multiple Choice Tests | 23 |
Testing Problems | 20 |
Computer Assisted Testing | 19 |
More ▼ |
Source
Author
Publication Type
Education Level
Location
California | 3 |
Canada | 3 |
Louisiana | 3 |
United Kingdom | 2 |
United States | 2 |
Asia | 1 |
Australia | 1 |
Bhutan | 1 |
Cambodia | 1 |
China | 1 |
Ethiopia | 1 |
More ▼ |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 4 |
Individuals with Disabilities… | 2 |
Improving Americas Schools… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Yang Du; Susu Zhang – Journal of Educational and Behavioral Statistics, 2025
Item compromise has long posed challenges in educational measurement, jeopardizing both test validity and test security of continuous tests. Detecting compromised items is therefore crucial to address this concern. The present literature on compromised item detection reveals two notable gaps: First, the majority of existing methods are based upon…
Descriptors: Item Response Theory, Item Analysis, Bayesian Statistics, Educational Assessment
Jennifer Darling-Aduana; Carolyn J. Heinrich; Jeremy Noonan; Jialing Wu; Kathryn Enriquez – Annenberg Institute for School Reform at Brown University, 2024
Online credit recovery (OCR) courses are the most common means through which students retake courses required for high school graduation. Yet a growing body of research has raised concerns regarding student learning in these courses, with low quality assessments posited as one contributing factor. To address this concern, we reviewed every…
Descriptors: Online Courses, Required Courses, Repetition, Credits
Kyung-Mi O. – Language Testing in Asia, 2024
This study examines the efficacy of artificial intelligence (AI) in creating parallel test items compared to human-made ones. Two test forms were developed: one consisting of 20 existing human-made items and another with 20 new items generated with ChatGPT assistance. Expert reviews confirmed the content parallelism of the two test forms.…
Descriptors: Comparative Analysis, Artificial Intelligence, Computer Software, Test Items
Yang, Li-Ping; Xin, Tao – Educational Measurement: Issues and Practice, 2022
The upgrade educational information technology triggered by COVID-19 has shaped a new educational order and new educational forms. As a result, traditional educational measurement is now facing a systematic transformation, that is, from the Assessment of Learning (AoL) to Assessment for Learning (AfL), and finally to Assessment as Learning (AaL).…
Descriptors: Educational Assessment, Information Technology, Educational Technology, COVID-19
Goran Trajkovski; Heather Hayes – Digital Education and Learning, 2025
This book explores the transformative role of artificial intelligence in educational assessment, catering to researchers, educators, administrators, policymakers, and technologists involved in shaping the future of education. It delves into the foundations of AI-assisted assessment, innovative question types and formats, data analysis techniques,…
Descriptors: Artificial Intelligence, Educational Assessment, Computer Uses in Education, Test Format
Wind, Stefanie A.; Guo, Wenjing – Educational Assessment, 2021
Scoring procedures for the constructed-response (CR) items in large-scale mixed-format educational assessments often involve checks for rater agreement or rater reliability. Although these analyses are important, researchers have documented rater effects that persist despite rater training and that are not always detected in rater agreement and…
Descriptors: Scoring, Responses, Test Items, Test Format
Yi-Hsuan Lee; Yue Jia – Applied Measurement in Education, 2024
Test-taking experience is a consequence of the interaction between students and assessment properties. We define a new notion, rapid-pacing behavior, to reflect two types of test-taking experience -- disengagement and speededness. To identify rapid-pacing behavior, we extend existing methods to develop response-time thresholds for individual items…
Descriptors: Adaptive Testing, Reaction Time, Item Response Theory, Test Format
Wang, Yu; Chiu, Chia-Yi; Köhn, Hans Friedrich – Journal of Educational and Behavioral Statistics, 2023
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Nonparametric Statistics, Test Format, Educational Assessment
Rios, Joseph A.; Ihlenfeldt, Samuel D.; Dosedel, Michael; Riegelman, Amy – Educational Measurement: Issues and Practice, 2020
This systematic review investigated the topics studied and reporting practices of published meta-analyses in educational measurement. Our findings indicated that meta-analysis is not a highly utilized methodological tool in educational measurement; on average, less than one meta-analysis has been published per year over the past 30 years (28…
Descriptors: Meta Analysis, Educational Assessment, Test Format, Testing Accommodations
Cari F. Herrmann Abell – Grantee Submission, 2021
In the last twenty-five years, the discussion surrounding validity evidence has shifted both in language and scope, from the work of Messick and Kane to the updated Standards for Educational and Psychological Testing. However, these discussions haven't necessarily focused on best practices for different types of instruments or assessments, taking…
Descriptors: Test Format, Measurement Techniques, Student Evaluation, Rating Scales
Stephen G. Sireci; Javier Suárez-Álvarez; April L. Zenisky; Maria Elena Oliveri – Grantee Submission, 2024
The goal in personalized assessment is to best fit the needs of each individual test taker, given the assessment purposes. Design-In-Real-Time (DIRTy) assessment reflects the progressive evolution in testing from a single test, to an adaptive test, to an adaptive assessment "system." In this paper, we lay the foundation for DIRTy…
Descriptors: Educational Assessment, Student Needs, Test Format, Test Construction
Stephen G. Sireci; Javier Suárez-Álvarez; April L. Zenisky; Maria Elena Oliveri – Educational Measurement: Issues and Practice, 2024
The goal in personalized assessment is to best fit the needs of each individual test taker, given the assessment purposes. Design-in-Real-Time (DIRTy) assessment reflects the progressive evolution in testing from a single test, to an adaptive test, to an adaptive assessment "system." In this article, we lay the foundation for DIRTy…
Descriptors: Educational Assessment, Student Needs, Test Format, Test Construction
Care, Esther; Vista, Alvin; Kim, Helyn – UNESCO Bangkok, 2019
UNESCO's Asia-Pacific Regional Bureau for Education has been working on education quality under the name of 'transversal competencies' (TVC) since 2013. Many of these competencies have been included in national education policy and curricula of countries in the region, but now the importance accorded them is increasingly gaining attention. As…
Descriptors: Foreign Countries, Educational Quality, 21st Century Skills, Competence
Jimenez, Laura; Modaffari, Jamil – Center for American Progress, 2021
Assessments are a way for stakeholders in education to understand what students know and can do. They can take many forms, including but not limited to paper and pencil or computer-adaptive formats. However, assessments do not have to be tests in the traditional sense at all; rather, they can be carried out through teacher observations of students…
Descriptors: Equal Education, Elementary Secondary Education, Futures (of Society), Computer Assisted Testing
Lance M. Kruse – ProQuest LLC, 2019
This study explores six item-reduction methodologies used to shorten an existing complex problem-solving non-objective test by evaluating how each shortened form performs across three sources of validity evidence (i.e., test content, internal structure, and relationships with other variables). Two concerns prompted the development of the present…
Descriptors: Educational Assessment, Comparative Analysis, Test Format, Test Length