Publication Date
In 2025 | 8 |
Since 2024 | 23 |
Descriptor
Source
Author
Guher Gorgun | 2 |
K. Kawena Begay | 2 |
Miranda Kucera | 2 |
Okan Bulut | 2 |
Alicia A. Stoltenberg | 1 |
Alina Lutsyk-King | 1 |
Ann M. Aviles | 1 |
Arif Cem Topuz | 1 |
Brian F. French | 1 |
Bridget McHugh | 1 |
Bryan R. Drost | 1 |
More ▼ |
Publication Type
Journal Articles | 22 |
Reports - Research | 15 |
Reports - Descriptive | 4 |
Information Analyses | 2 |
Dissertations/Theses -… | 1 |
Reports - Evaluative | 1 |
Education Level
Elementary Education | 5 |
Secondary Education | 4 |
Higher Education | 3 |
Postsecondary Education | 3 |
Early Childhood Education | 2 |
Grade 11 | 1 |
Grade 3 | 1 |
High Schools | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Primary Education | 1 |
More ▼ |
Audience
Location
Australia | 1 |
Bosnia and Herzegovina | 1 |
Indonesia | 1 |
New Zealand | 1 |
South Africa (Johannesburg) | 1 |
United Kingdom (England) | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Tia M. Fechter; Heeyeon Yoon – Language Testing, 2024
This study evaluated the efficacy of two proposed methods in an operational standard-setting study conducted for a high-stakes language proficiency test of the U.S. government. The goal was to seek low-cost modifications to the existing Yes/No Angoff method to increase the validity and reliability of the recommended cut scores using a convergent…
Descriptors: Standard Setting, Language Proficiency, Language Tests, Evaluation Methods
Kylie Gorney; Mark D. Reckase – Journal of Educational Measurement, 2025
In computerized adaptive testing, item exposure control methods are often used to provide a more balanced usage of the item pool. Many of the most popular methods, including the restricted method (Revuelta and Ponsoda), use a single maximum exposure rate to limit the proportion of times that each item is administered. However, Barrada et al.…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Item Banks
Guher Gorgun; Okan Bulut – Education and Information Technologies, 2024
In light of the widespread adoption of technology-enhanced learning and assessment platforms, there is a growing demand for innovative, high-quality, and diverse assessment questions. Automatic Question Generation (AQG) has emerged as a valuable solution, enabling educators and assessment developers to efficiently produce a large volume of test…
Descriptors: Computer Assisted Testing, Test Construction, Test Items, Automation
Arif Cem Topuz; Kinshuk – Educational Technology Research and Development, 2024
Online assessments of learning, or online exams, have become increasingly widespread with the rise of distance learning. Online exams are preferred by many students and are perceived as a quick and easy tool to measure knowledge. On the contrary, some students are concerned about the possibility of cheating and technological difficulties in online…
Descriptors: Computer Assisted Testing, Student Evaluation, Evaluation Methods, Student Attitudes
Zachary K. Collier; Minji Kong; Olushola Soyoye; Kamal Chawla; Ann M. Aviles; Yasser Payne – Journal of Educational and Behavioral Statistics, 2024
Asymmetric Likert-type items in research studies can present several challenges in data analysis, particularly concerning missing data. These items are often characterized by a skewed scaling, where either there is no neutral response option or an unequal number of possible positive and negative responses. The use of conventional techniques, such…
Descriptors: Likert Scales, Test Items, Item Analysis, Evaluation Methods
Joakim Wallmark; James O. Ramsay; Juan Li; Marie Wiberg – Journal of Educational and Behavioral Statistics, 2024
Item response theory (IRT) models the relationship between the possible scores on a test item against a test taker's attainment of the latent trait that the item is intended to measure. In this study, we compare two models for tests with polytomously scored items: the optimal scoring (OS) model, a nonparametric IRT model based on the principles of…
Descriptors: Item Response Theory, Test Items, Models, Scoring
Fu Chen; Ying Cui; Alina Lutsyk-King; Yizhu Gao; Xiaoxiao Liu; Maria Cutumisu; Jacqueline P. Leighton – Education and Information Technologies, 2024
Post-secondary data literacy education is critical to students' academic and career success. However, the literature has not adequately addressed the conceptualization and assessment of data literacy for post-secondary students. In this study, we introduced a novel digital performance-based assessment for teaching and evaluating post-secondary…
Descriptors: Performance Based Assessment, College Students, Information Literacy, Evaluation Methods
Santi Lestari – Research Matters, 2025
The ability to draw visual representations such as diagrams and graphs is considered fundamental to science learning. Science exams therefore often include questions which require students to draw a visual representation, or to augment a partially provided one. The design features of such questions (e.g., layout of diagrams, amount of answer…
Descriptors: Science Education, Secondary Education, Visual Aids, Foreign Countries
Maristela Petrovic-Dzerdz – Collected Essays on Learning and Teaching, 2024
Large introductory classes, with their expansive curriculum, demand assessment strategies that blend efficiency with reliability, prompting the consideration of multiple-choice (MC) tests as a viable option. Crafting a high-quality MC test, however, necessitates a meticulous process involving reflection on assessment format appropriateness, test…
Descriptors: Multiple Choice Tests, Test Construction, Test Items, Alignment (Education)
Edinalda Jakubovic; Haris Memisevic – Journal of Research in Special Educational Needs, 2024
The Teacher Efficacy for Inclusive Practices (TEIP) scale is a widely used instrument for assessing teachers' effectiveness in implementing inclusive practices. The TEIP has not been validated in Bosnia and Herzegovina (BIH). The goal of the present study was to conduct a confirmatory factor analysis (CFA) of the TEIP in a sample of teachers in…
Descriptors: Teacher Effectiveness, Inclusion, Teaching Methods, Foreign Countries
Guher Gorgun; Okan Bulut – Educational Measurement: Issues and Practice, 2025
Automatic item generation may supply many items instantly and efficiently to assessment and learning environments. Yet, the evaluation of item quality persists to be a bottleneck for deploying generated items in learning and assessment settings. In this study, we investigated the utility of using large-language models, specifically Llama 3-8B, for…
Descriptors: Artificial Intelligence, Quality Control, Technology Uses in Education, Automation
Bryan R. Drost; Char Shryock – Phi Delta Kappan, 2025
Creating assessment questions aligned to standards is a time-consuming task for teachers, but large language models such as ChatGPT can help. Bryan Drost & Char Shryock describe a three-step process for using ChatGPT to create assessments: 1) Ask ChatGPT to break standards into measurable targets. 2) Determine how much time to spend on each…
Descriptors: Artificial Intelligence, Computer Software, Technology Integration, Teaching Methods
Umi Laili Yuhana; Eko Mulyanto Yuniarno; Wenny Rahayu; Eric Pardede – Education and Information Technologies, 2024
In an online learning environment, it is important to establish a suitable assessment approach that can be adapted on the fly to accommodate the varying learning paces of students. At the same time, it is essential that assessment criteria remain compliant with the expected learning outcomes of the relevant education standard which predominantly…
Descriptors: Adaptive Testing, Electronic Learning, Elementary School Students, Student Evaluation
Traditional vs Intersectional DIF Analysis: Considerations and a Comparison Using State Testing Data
Tony Albano; Brian F. French; Thao Thu Vo – Applied Measurement in Education, 2024
Recent research has demonstrated an intersectional approach to the study of differential item functioning (DIF). This approach expands DIF to account for the interactions between what have traditionally been treated as separate grouping variables. In this paper, we compare traditional and intersectional DIF analyses using data from a state testing…
Descriptors: Test Items, Item Analysis, Data Use, Standardized Tests
Previous Page | Next Page »
Pages: 1 | 2