Publication Date
In 2025 | 1 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 5 |
Since 2016 (last 10 years) | 10 |
Since 2006 (last 20 years) | 17 |
Descriptor
Automation | 21 |
Test Construction | 21 |
Test Items | 12 |
Computer Assisted Testing | 10 |
Psychometrics | 6 |
Test Validity | 6 |
Artificial Intelligence | 5 |
Scoring | 5 |
Computer Software | 4 |
Quality Control | 4 |
Test Reliability | 4 |
More ▼ |
Source
Author
Anita Pásztor-Kovács | 1 |
Attila Pásztor | 1 |
Barghaus, Katherine M. | 1 |
Bateson, Gordon | 1 |
Bejar, Isaac I. | 1 |
Bourda, Yolaine | 1 |
Bruillard, Éric | 1 |
Chung, Hyewon | 1 |
Diao, Qi | 1 |
Dodd, Barbara G. | 1 |
Embretson, Susan E. | 1 |
More ▼ |
Publication Type
Reports - Descriptive | 21 |
Journal Articles | 18 |
Numerical/Quantitative Data | 2 |
Tests/Questionnaires | 2 |
Opinion Papers | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Early Childhood Education | 2 |
Elementary Education | 2 |
Elementary Secondary Education | 2 |
Grade 3 | 2 |
Grade 4 | 2 |
Grade 5 | 2 |
Grade 6 | 2 |
Grade 7 | 2 |
Grade 9 | 2 |
High Schools | 2 |
Higher Education | 2 |
More ▼ |
Audience
Practitioners | 1 |
Location
Canada | 1 |
Japan | 1 |
United States | 1 |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
International English… | 1 |
Test of English for… | 1 |
What Works Clearinghouse Rating
Yanyan Fu – Educational Measurement: Issues and Practice, 2024
The template-based automated item-generation (TAIG) approach that involves template creation, item generation, item selection, field-testing, and evaluation has more steps than the traditional item development method. Consequentially, there is more margin for error in this process, and any template errors can be cascaded to the generated items.…
Descriptors: Error Correction, Automation, Test Items, Test Construction
Jonathan Seiden – Annenberg Institute for School Reform at Brown University, 2025
Direct assessments of early childhood development (ECD) are a cornerstone of research in developmental psychology and are increasingly used to evaluate programs and policies in lower- and middle-income countries. Despite strong psychometric properties, these assessments are too expensive and time consuming for use in large-scale monitoring or…
Descriptors: Young Children, Child Development, Performance Based Assessment, Developmental Psychology
Anita Pásztor-Kovács; Attila Pásztor; Gyöngyvér Molnár – Interactive Learning Environments, 2023
In this paper, we present an agenda for the research directions we recommend in addressing the issues of realizing and evaluating communication in CPS instruments. We outline our ideas on potential ways to improve: (1) generalizability in Human-Human assessment tools and ecological validity in Human-Agent ones; (2) flexible and convenient use of…
Descriptors: Cooperation, Problem Solving, Evaluation Methods, Teamwork
Advancing Language Assessment with AI and ML--Leaning into AI Is Inevitable, but Can Theory Keep Up?
Xiaoming Xi – Language Assessment Quarterly, 2023
Following the burgeoning growth of artificial intelligence (AI) and machine learning (ML) applications in language assessment in recent years, the meteoric rise of ChatGPT and its sweeping applications in almost every sector have left us in awe, scrambling to catch up by developing theories and best practices. This special issue features studies…
Descriptors: Artificial Intelligence, Theory Practice Relationship, Language Tests, Man Machine Systems
Tran, Tich Phuoc; Meacheam, David – IEEE Transactions on Learning Technologies, 2020
The use of learning management systems (LMSs) for learning and knowledge sharing has accelerated quickly both in education and corporate worlds. Despite the benefits brought by LMSs, the current systems still face significant challenges, including the lack of automation in generating quiz questions and managing courses. Over the past decade, more…
Descriptors: Integrated Learning Systems, Test Construction, Test Items, Automation
Vie, Jill-Jênn; Popineau, Fabrice; Bruillard, Éric; Bourda, Yolaine – International Journal of Artificial Intelligence in Education, 2018
In large-scale assessments such as the ones encountered in MOOCs, a lot of usage data is available because of the number of learners involved. Newcomers, that just arrive on a MOOC, have various backgrounds in terms of knowledge, but the platform hardly knows anything about them. Therefore, it is crucial to elicit their knowledge fast, in order to…
Descriptors: Automation, Test Construction, Measurement, Online Courses
Bateson, Gordon – International Journal of Computer-Assisted Language Learning and Teaching, 2021
As a result of the Japanese Ministry of Education's recent edict that students' written and spoken English should be assessed in university entrance exams, there is an urgent need for tools to help teachers and students prepare for these exams. Although some commercial tools already exist, they are generally expensive and inflexible. To address…
Descriptors: Test Construction, Computer Assisted Testing, Internet, Writing Tests
Rupp, André A. – Applied Measurement in Education, 2018
This article discusses critical methodological design decisions for collecting, interpreting, and synthesizing empirical evidence during the design, deployment, and operational quality-control phases for automated scoring systems. The discussion is inspired by work on operational large-scale systems for automated essay scoring but many of the…
Descriptors: Design, Automation, Scoring, Test Scoring Machines
Park, Ryoungsun; Kim, Jiseon; Dodd, Barbara G.; Chung, Hyewon – Applied Psychological Measurement, 2011
JPLEX, short for Java simPLEX, is an automated test assembly (ATA) program. It is a mixed integer linear programming (MILP) solver written in Java. It reads in a configuration file, solves the minimization problem, and produces an output file for postprocessing. It implements the simplex algorithm to create a fully relaxed solution and…
Descriptors: Test Construction, Automation, Computer Software
Veldkamp, Bernard P.; Matteucci, Mariagiulia; de Jong, Martijn G. – Applied Psychological Measurement, 2013
Item response theory parameters have to be estimated, and because of the estimation process, they do have uncertainty in them. In most large-scale testing programs, the parameters are stored in item banks, and automated test assembly algorithms are applied to assemble operational test forms. These algorithms treat item parameters as fixed values,…
Descriptors: Test Construction, Test Items, Item Banks, Automation
New Meridian Corporation, 2020
The purpose of this report is to describe the technical qualities of the 2018-2019 operational administration of the English language arts/literacy (ELA/L) and mathematics summative assessments in grades 3 through 8 and high school. The ELA/L assessments focus on reading and comprehending a range of sufficiently complex texts independently and…
Descriptors: Language Arts, Literacy Education, Mathematics Education, Summative Evaluation
New Meridian Corporation, 2020
The purpose of this report is to describe the technical qualities of the 2018-2019 operational administration of the English language arts/literacy (ELA/L) and mathematics assessments in grades 3 through 8 and high school. New Meridian, in coordination with multiple states and vendors, developed an alternate form of the summative assessment to…
Descriptors: Language Arts, Literacy Education, Mathematics Education, Summative Evaluation
van der Linden, Wim J.; Diao, Qi – Journal of Educational Measurement, 2011
In automated test assembly (ATA), the methodology of mixed-integer programming is used to select test items from an item bank to meet the specifications for a desired test form and optimize its measurement accuracy. The same methodology can be used to automate the formatting of the set of selected items into the actual test form. Three different…
Descriptors: Test Items, Test Format, Test Construction, Item Banks
Porter, Andrew; Polikoff, Morgan S.; Barghaus, Katherine M.; Yang, Rui – Educational Researcher, 2013
We describe an innovative automated test construction algorithm for building aligned achievement tests. By incorporating the algorithm into the test construction process, along with other test construction procedures for building reliable and unbiased assessments, the result is much more valid tests than result from current test construction…
Descriptors: Achievement Tests, Automation, Test Construction, Alignment (Education)
Gierl, Mark J.; Lai, Hollis – International Journal of Testing, 2012
Automatic item generation represents a relatively new but rapidly evolving research area where cognitive and psychometric theories are used to produce tests that include items generated using computer technology. Automatic item generation requires two steps. First, test development specialists create item models, which are comparable to templates…
Descriptors: Foreign Countries, Psychometrics, Test Construction, Test Items
Previous Page | Next Page »
Pages: 1 | 2