NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 46 to 60 of 136,279 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hung Tan Ha; Duyen Thi Bich Nguyen; Tim Stoeckel – Language Assessment Quarterly, 2025
This article compares two methods for detecting local item dependence (LID): residual correlation examination and Rasch testlet modeling (RTM), in a commonly used 3:6 matching format and an extended matching test (EMT) format. The two formats are hypothesized to facilitate different levels of item dependency due to differences in the number of…
Descriptors: Comparative Analysis, Language Tests, Test Items, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Raudlah Melinda Sidik; Ana Ratna Wulan; K. Kusnadi – Journal of Biological Education Indonesia (Jurnal Pendidikan Biologi Indonesia), 2025
The research developed and validated EKSAI (Epistemic Knowledge Science Assessment Instrument), an assessment tool for epistemic knowledge in science education. The background is that 21st-century challenges demand a transformation in science education, with a focus on understanding how scientific knowledge is developed and evaluated, which is…
Descriptors: Science Tests, Knowledge Level, Biology, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Yuriko K. Sosa Paredes; Björn Andersson – Educational Assessment, Evaluation and Accountability, 2025
In international large-scale assessments, student performance comparisons across educational systems are frequently done to assess the state and development in different domains. These results often have a large impact on educational policy and on the perceptions of an educational system's performance. Early assessments, such as the First and…
Descriptors: Test Interpretation, International Assessment, Science Tests, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Shangchao Min; Kyoungwon Bishop – Language Testing, 2024
This paper evaluates the multistage adaptive test (MST) design of a large-scale academic language assessment (ACCESS) for Grades 1-12, with an aim to simplify the current MST design, using both operational and simulated test data. Study 1 explored the operational population data (1,456,287 test-takers) of the listening and reading tests of MST…
Descriptors: Adaptive Testing, Test Construction, Language Tests, English Language Learners
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Alaa Eldin A. Ayoub; Muneera R. Ghablan; Eid G. Abo Hamza; Ahmed M. Abdulla Alabbasi – European Journal of STEM Education, 2025
This study describes the development of the science, technology, engineering, and mathematics (STEM) Scale, intended to assess parental attitudes toward school programs designed to deliver STEM, and evaluates its psychometric properties. The study group included 400 parents of students (138 males and 262 females) enrolled in STEM programs…
Descriptors: STEM Education, Test Construction, Parent Attitudes, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Jacqueline Raymond; David Wei Dai; Sue McAllister – Advances in Health Sciences Education, 2025
There is increasing interest in health professions education (HPE) in applying argument-based validity approaches, such as Kane's, to assessment design. The critical first step in employing Kane's approach is to specify the interpretation-use argument (IUA). However, in the HPE literature, this step is often poorly articulated. This article…
Descriptors: Allied Health Occupations Education, Test Interpretation, Test Construction, Inferences
Peer reviewed Peer reviewed
Direct linkDirect link
Helen Zhang; Anthony Perry; Irene Lee – International Journal of Artificial Intelligence in Education, 2025
The rapid expansion of Artificial Intelligence (AI) in our society makes it urgent and necessary to develop young students' AI literacy so that they can become informed citizens and critical consumers of AI technology. Over the past decade many efforts have focused on developing curricular materials that make AI concepts accessible and engaging to…
Descriptors: Test Construction, Test Validity, Measures (Individuals), Artificial Intelligence
Peer reviewed Peer reviewed
Direct linkDirect link
Hung-Yu Huang – Educational and Psychological Measurement, 2025
The use of discrete categorical formats to assess psychological traits has a long-standing tradition that is deeply embedded in item response theory models. The increasing prevalence and endorsement of computer- or web-based testing has led to greater focus on continuous response formats, which offer numerous advantages in both respondent…
Descriptors: Response Style (Tests), Psychological Characteristics, Item Response Theory, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Jun-ichiro Yasuda; Michael M. Hull; Naohiro Mae; Kentaro Kojima – Physical Review Physics Education Research, 2025
Although conceptual assessment tests are commonly administered at the beginning and end of a semester, this pre-post approach has inherent limitations. Specifically, education researchers and instructors have limited ability to observe the progression of students' conceptual understanding throughout the course. Furthermore, instructors are limited…
Descriptors: Computer Assisted Testing, Adaptive Testing, Science Tests, Scientific Concepts
Peer reviewed Peer reviewed
Direct linkDirect link
Filiz Arzu Yalin; Ahmet Özbay; Safak Oguz – European Journal of Education, 2025
This study developed and validated a Decision-Making Skill Test (DMST) for Turkish adolescents to address the lack of culturally appropriate assessment tools for multi-criteria decision-making skills. A cross-sectional design was employed with 427 participants aged 11-17 years from diverse socioeconomic backgrounds across Turkey. Following…
Descriptors: Test Construction, Test Validity, Student Evaluation, Decision Making
Peer reviewed Peer reviewed
Direct linkDirect link
Janet Mee; Ravi Pandian; Justin Wolczynski; Amy Morales; Miguel Paniagua; Polina Harik; Peter Baldwin; Brian E. Clauser – Advances in Health Sciences Education, 2024
Recent advances in automated scoring technology have made it practical to replace multiple-choice questions (MCQs) with short-answer questions (SAQs) in large-scale, high-stakes assessments. However, most previous research comparing these formats has used small examinee samples testing under low-stakes conditions. Additionally, previous studies…
Descriptors: Multiple Choice Tests, High Stakes Tests, Test Format, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Marzieh Haghayeghi; Ali Moghadamzadeh; Hamdollah Ravand; Mohamad Javadipour; Hossein Kareshki – Journal of Psychoeducational Assessment, 2025
This study aimed to address the need for a comprehensive assessment tool to evaluate the mathematical abilities of first-grade students through cognitive diagnostic assessment (CDA). The primary challenge involved in this endeavor was to delineate the specific cognitive skills and sub-skills pertinent to first-grade mathematics (FG-M) and to…
Descriptors: Test Construction, Cognitive Measurement, Check Lists, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Eder Hernandez; Esmeralda Campos; Pablo Barniol; Genaro Zavala – Physical Review Physics Education Research, 2025
This study presents the development and validation of a novel multiple-choice test designed to assess university students' conceptual understanding of electric field, force, and flux. The test of understanding of electric field, force, and flux was constructed based on the results of previous studies using a phenomenographic approach to classify…
Descriptors: Physics, Scientific Concepts, Science Tests, Multiple Choice Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Lyniesha Ward; Fridah Rotich; Jeffrey R. Raker; Regis Komperda; Sachin Nedungadi; Maia Popova – Chemistry Education Research and Practice, 2025
This paper describes the design and evaluation of the Organic chemistry Representational Competence Assessment (ORCA). Grounded in Kozma and Russell's representational competence framework, the ORCA measures the learner's ability to "interpret," "translate," and "use" six commonly used representations of molecular…
Descriptors: Organic Chemistry, Science Tests, Test Construction, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Andrew P. Jaciw – American Journal of Evaluation, 2025
By design, randomized experiments (XPs) rule out bias from confounded selection of participants into conditions. Quasi-experiments (QEs) are often considered second-best because they do not share this benefit. However, when results from XPs are used to generalize causal impacts, the benefit from unconfounded selection into conditions may be offset…
Descriptors: Elementary School Students, Elementary School Teachers, Generalization, Test Bias
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  9086