NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 166 to 180 of 3,982 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Perry, Lindsey – AERA Online Paper Repository, 2017
Before an assessment is used to make decisions, the validity of the intended interpretation must be evaluated. The purpose of this paper is to describe how the argument-based approach and an interpretation/use argument (IUA) (Kane, 2013) were used to validate the interpretations made from the new Early Grade Mathematics Assessment (EGMA)…
Descriptors: Student Evaluation, Mathematics Tests, Test Interpretation, Inferences
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chiavaroli, Neville – Practical Assessment, Research & Evaluation, 2017
Despite the majority of MCQ writing guides discouraging the use of negatively-worded multiple choice questions (NWQs), they continue to be regularly used both in locally produced examinations and commercially available questions. There are several reasons why the use of NWQs may prove resistant to sound pedagogical advice. Nevertheless, systematic…
Descriptors: Multiple Choice Tests, Test Construction, Test Items, Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Klemencic, Eva; Mirazchiyski, Plamen Vladkov – Comparative Education, 2018
International large-scale student assessments (ILSAs) in education represent a valuable source of information for policy-makers, not only on student achievements, but also on their relationship with different contextual factors. The results are partly described in the official studies' reports; more can be derived from the publicly released data…
Descriptors: Evidence Based Practice, Educational Administration, International Assessment, Policy Formation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Leonard, Jack – Education Policy Analysis Archives, 2018
This paper introduces the new Massachusetts Performance Assessment for Leaders (PAL) and uses critical policy analysis to re-examine the validity evidence (using the 2014 Standards for Educational and Psychological Testing and a theory of multicultural validity) for the use and interpretation of the PAL in regards to emerging school leadership.…
Descriptors: Performance Based Assessment, Test Validity, High Stakes Tests, School Administration
Peer reviewed Peer reviewed
Direct linkDirect link
Aloisi, Cesare; Callaghan, A. – Higher Education Pedagogies, 2018
The University of Reading Learning Gain project is a three-year longitudinal project to test and evaluate a range of available methodologies and to draw conclusions on what might be the right combination of instruments for the measurement of Learning Gain in higher education. This paper analyses the validity of a measure of critical thinking…
Descriptors: Foreign Countries, Cognitive Tests, Critical Thinking, Thinking Skills
Boyer, Michelle; Landl, Erika – National Center on Educational Outcomes, 2021
This Brief contains a scan of the interim assessment landscape, and is focused on the availability of documentation supporting the appropriateness of these assessments for students with disabilities. The purpose of this Brief is to advise the development of guidance that facilitates improved practices related to the use of interim assessments for…
Descriptors: Students with Disabilities, Student Evaluation, Formative Evaluation, Inclusion
Peer reviewed Peer reviewed
Direct linkDirect link
Basu, Jayanti – International Journal of School & Educational Psychology, 2016
Intelligence testing was one of the earliest interests of psychologists in India. Adaptation of Western intelligence tests has been a focus of psychologists in the first half of the last century. Indigenous development of intelligence tests has been attempted, but diversity of language and culture, complexity of school systems, and infrastructural…
Descriptors: Intelligence Tests, Foreign Countries, School Psychology, Test Interpretation
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L. – Educational Measurement: Issues and Practice, 2017
The rise of computer-based testing has brought with it the capability to measure more aspects of a test event than simply the answers selected or constructed by the test taker. One behavior that has drawn much research interest is the time test takers spend responding to individual multiple-choice items. In particular, very short response…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Items, Reaction Time
Chamoy, Waritsa – ProQuest LLC, 2018
The main purpose of this study was to conduct a validation analysis of student surveys of teaching effectiveness implemented at Bangkok University, Thailand. This study included three phases; survey development, a pilot study, and a full implementation study. Four sources of validity evidence were collected to support intended interpretations and…
Descriptors: Foreign Countries, Psychometrics, Student Surveys, College Students
Peer reviewed Peer reviewed
Direct linkDirect link
Marion, Scott F. – Measurement: Interdisciplinary Research and Perspectives, 2015
The measurement industry is in crisis. The public outcry against "over testing" and the opt-out movement are symptoms of a larger sociopolitical battle being fought over Common Core, teacher evaluation, federal intrusion, and a host of other issues, but much of the vitriol is directed at the tests and the testing industry. If we, as…
Descriptors: Test Interpretation, Scores, Educational Assessment, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Faulkner-Bond, Molly; Wolf, Mikyung Kim; Wells, Craig S.; Sireci, Stephen G. – Language Assessment Quarterly, 2018
In this study we investigated the internal factor structure of a large-scale K--12 assessment of English language proficiency (ELP) using samples of fourth- and eighth-grade English learners (ELs) in one state. While U.S. schools are mandated to measure students' ELP in four language domains (listening, reading, speaking, and writing), some ELP…
Descriptors: Factor Structure, Language Tests, Language Proficiency, Grade 4
Hopfenbeck, Therese N.; Lenkeit, Jenny – International Association for the Evaluation of Educational Achievement, 2018
International large-scale assessments (ILSAs) have had an increasing influence on the discourse surrounding education systems around the world. However, the results of these studies tend to have less impact on pedagogy in the classroom than would be expected. For example, a recent review of 114 published peer-reviewed articles on the IEA's…
Descriptors: Foreign Countries, Achievement Tests, Grade 4, Reading Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Willse, John T. – Measurement and Evaluation in Counseling and Development, 2017
This article provides a brief introduction to the Rasch model. Motivation for using Rasch analyses is provided. Important Rasch model concepts and key aspects of result interpretation are introduced, with major points reinforced using a simulation demonstration. Concrete guidelines are provided regarding sample size and the evaluation of items.
Descriptors: Item Response Theory, Test Results, Test Interpretation, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Ying; Gierl, Mark; Guo, Qi – Educational Psychology, 2016
The purpose of the current investigation was to describe how the artificial neural networks (ANNs) can be used to interpret student performance on cognitive diagnostic assessments (CDAs) and evaluate the performances of ANNs using simulation results. CDAs are designed to measure student performance on problem-solving tasks and provide useful…
Descriptors: Cognitive Tests, Diagnostic Tests, Classification, Artificial Intelligence
Peer reviewed Peer reviewed
Direct linkDirect link
Hidalgo, Ma Dolores; Benítez, Isabel; Padilla, Jose-Luis; Gómez-Benito, Juana – Sociological Methods & Research, 2017
The growing use of scales in survey questionnaires warrants the need to address how does polytomous differential item functioning (DIF) affect observed scale score comparisons. The aim of this study is to investigate the impact of DIF on the type I error and effect size of the independent samples t-test on the observed total scale scores. A…
Descriptors: Test Items, Test Bias, Item Response Theory, Surveys
Pages: 1  |  ...  |  8  |  9  |  10  |  11  |  12  |  13  |  14  |  15  |  16  |  ...  |  266