Publication Date
In 2025 | 2 |
Since 2024 | 18 |
Since 2021 (last 5 years) | 45 |
Since 2016 (last 10 years) | 72 |
Since 2006 (last 20 years) | 103 |
Descriptor
Test Construction | 470 |
Test Validity | 137 |
Elementary Secondary Education | 100 |
Test Items | 83 |
Literature Reviews | 80 |
Test Reliability | 72 |
Evaluation Methods | 58 |
Higher Education | 54 |
Test Use | 54 |
Test Format | 53 |
Student Evaluation | 52 |
More ▼ |
Source
Author
Hambleton, Ronald K. | 5 |
Baker, Eva L. | 4 |
Ellington, Henry | 4 |
Haladyna, Tom | 4 |
Roid, Gale | 4 |
Downing, Steven M. | 3 |
Haladyna, Thomas M. | 3 |
O'Neil, Harold F., Jr. | 3 |
Quellmalz, Edys S. | 3 |
Reckase, Mark D. | 3 |
Roid, Gale H. | 3 |
More ▼ |
Publication Type
Education Level
Audience
Practitioners | 30 |
Researchers | 27 |
Teachers | 12 |
Policymakers | 3 |
Administrators | 2 |
Counselors | 1 |
Location
Australia | 8 |
United States | 6 |
China | 5 |
Turkey | 4 |
Canada | 3 |
France | 3 |
United Kingdom | 3 |
United Kingdom (England) | 3 |
California | 2 |
Florida | 2 |
Iran | 2 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Simen Hjellvik; Steven Mallam; Marte Fannelø Giskeødegård; Salman Nazir – Technology, Knowledge and Learning, 2024
Computer-based simulation is utilised across various educational fields, employing diverse technologies to facilitate practical understanding of content and the acquisition of skills that can help close the gap between theory and practice. The possibility of providing scenarios that resemble on-the-job tasks, enables instructors to both train and…
Descriptors: Computer Simulation, Competence, Evaluation Methods, Test Construction
Séverin Lions; María Paz Blanco; Pablo Dartnell; Carlos Monsalve; Gabriel Ortega; Julie Lemarié – Applied Measurement in Education, 2024
Multiple-choice items are universally used in formal education. Since they should assess learning, not test-wiseness or guesswork, they must be constructed following the highest possible standards. Hundreds of item-writing guides have provided guidelines to help test developers adopt appropriate strategies to define the distribution and sequence…
Descriptors: Test Construction, Multiple Choice Tests, Guidelines, Test Items
Welsandt, Nina Charlotte Johanna; Abs, Hermann Josef – Journal of Social Science Education, 2023
Purpose: This paper analyses and classifies currently available English- and German-language measurement instruments for assessing economic literacy. It shows the content-related focuses and gaps of the extracted test instruments, the cognitive level of demand that characterises the instruments, the technical forms of implementation, and the…
Descriptors: Economics, Knowledge Level, Measures (Individuals), German
Kiliç, Abdullah Faruk; Koyuncu, Ilhan; Uysal, Ibrahim – International Journal of Psychology and Educational Studies, 2023
Classical test theory (CTT) and item response theory (IRT) are two fundamental approaches used in scale development research. Although CTT is the preferred methodology in scale development research in the Republic of Turkey, the IRT methodology has started to gain traction in recent years. In this study, we used the systematic review methodology…
Descriptors: Test Construction, Item Response Theory, Literature Reviews, Foreign Countries
Samah AlKhuzaey; Floriana Grasso; Terry R. Payne; Valentina Tamma – International Journal of Artificial Intelligence in Education, 2024
Designing and constructing pedagogical tests that contain items (i.e. questions) which measure various types of skills for different levels of students equitably is a challenging task. Teachers and item writers alike need to ensure that the quality of assessment materials is consistent, if student evaluations are to be objective and effective.…
Descriptors: Test Items, Test Construction, Difficulty Level, Prediction
Hongwen Guo; Matthew S. Johnson; Daniel F. McCaffrey; Lixong Gu – ETS Research Report Series, 2024
The multistage testing (MST) design has been gaining attention and popularity in educational assessments. For testing programs that have small test-taker samples, it is challenging to calibrate new items to replenish the item pool. In the current research, we used the item pools from an operational MST program to illustrate how research studies…
Descriptors: Test Items, Test Construction, Sample Size, Scaling
Deny Kurniawan; Datuk Ary Adriansyah Samsura; A. M. A. van Deemen – Journal of Social Studies Education Research, 2023
Given the absence of expertise criteria of academics or lecturers, we sought to explore relevant studies to formulate an informed framework of academic expertise. Academics at higher education institutions are often considered experts. Usually, academic roles comprise teaching, research, and community service. Therefore, academics' expertise…
Descriptors: College Faculty, Expertise, Test Construction, Measures (Individuals)
Eray Selçuk; Ergül Demir – International Journal of Assessment Tools in Education, 2024
This research aims to compare the ability and item parameter estimations of Item Response Theory according to Maximum likelihood and Bayesian approaches in different Monte Carlo simulation conditions. For this purpose, depending on the changes in the priori distribution type, sample size, test length, and logistics model, the ability and item…
Descriptors: Item Response Theory, Item Analysis, Test Items, Simulation
Chan, Cecilia Ka Yuk; Chen, Siaw Wee – Assessment & Evaluation in Higher Education, 2023
This systematic review aims to explore how student partnership is enacted in higher education assessment using community of practice and liminality of student roles as the conceptual framework. Forty-three empirical studies were selected, and extracted data were synthesised using thematic analysis. The results show that student partnership occurs…
Descriptors: Student Participation, Evaluation, Test Construction, Feedback (Response)
Ediyanto, Ediyanto; Sunandar, Asep; Ramadhani, Risa Safira; Aqilah, Thalsa Syahda – Discourse and Communication for Sustainable Education, 2022
The current study presents an overview of studies investigating educational research instruments development. This study only discusses similar topics in instrument development. The instrument discussed is the assessment of perspective on inclusive education. This research adopted a literature review approach. There are seven research stages:…
Descriptors: Test Construction, Educational Research, Literature Reviews, Writing Processes
Mattar, João; Ramos, Daniela Karine; Lucas, Margarida Rocha – Education and Information Technologies, 2022
The purpose of this article is to compare digital competence assessment instruments based on DigComp related frameworks. The study aims to answer four questions: (a) What types of instruments based on these frameworks are available? (b) How were these instruments created from these frameworks? (c) What procedures were used to guarantee the…
Descriptors: Evaluation Methods, Literature Reviews, Test Construction, Competence
Christian X. Navarro-Cota; Ana I. Molina; Miguel A. Redondo; Carmen Lacave – IEEE Transactions on Education, 2024
Contribution: This article describes the process used to create a questionnaire to evaluate the usability of mobile learning applications (CECAM). The questionnaire includes specific questions to assess user interface usability and pedagogical usability. Background: Nowadays, mobile applications are expanding rapidly and are commonly used in…
Descriptors: Usability, Questionnaires, Electronic Learning, Computer Oriented Programs
Dongkwang Shin; Jang Ho Lee – ELT Journal, 2024
Although automated item generation has gained a considerable amount of attention in a variety of fields, it is still a relatively new technology in ELT contexts. Therefore, the present article aims to provide an accessible introduction to this powerful resource for language teachers based on a review of the available research. Particularly, it…
Descriptors: Language Tests, Artificial Intelligence, Test Items, Automation
Daniel M. Settlage; Jim R. Wollscheid – Journal of the Scholarship of Teaching and Learning, 2024
The examination of the testing mode effect has received increased attention as higher education has shifted to remote testing during the COVID-19 pandemic. We believe the testing mode effect consists of four components: the ability to physically write on the test, the method of answer recording, the proctoring/testing environment, and the effect…
Descriptors: College Students, Macroeconomics, Tests, Answer Sheets
Jessica Herring Watson; Amanda J. Rockinson-Szapkiw – Journal of Research on Technology in Education, 2023
This article reports on the development and initial validation of the Intention to Use Technology-enabled Learning (I-TEL) Scale. Data from 313 preservice teachers were used to examine the instrument's psychometric properties. Principal axis factoring (PAF) indicated a five-factor structure consisting of 22 items. The scale, as well as the…
Descriptors: Test Construction, Test Validity, Preservice Teachers, Psychometrics