NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 181 to 195 of 3,984 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Ying; Gierl, Mark; Guo, Qi – Educational Psychology, 2016
The purpose of the current investigation was to describe how the artificial neural networks (ANNs) can be used to interpret student performance on cognitive diagnostic assessments (CDAs) and evaluate the performances of ANNs using simulation results. CDAs are designed to measure student performance on problem-solving tasks and provide useful…
Descriptors: Cognitive Tests, Diagnostic Tests, Classification, Artificial Intelligence
Peer reviewed Peer reviewed
Direct linkDirect link
Hidalgo, Ma Dolores; Benítez, Isabel; Padilla, Jose-Luis; Gómez-Benito, Juana – Sociological Methods & Research, 2017
The growing use of scales in survey questionnaires warrants the need to address how does polytomous differential item functioning (DIF) affect observed scale score comparisons. The aim of this study is to investigate the impact of DIF on the type I error and effect size of the independent samples t-test on the observed total scale scores. A…
Descriptors: Test Items, Test Bias, Item Response Theory, Surveys
Peer reviewed Peer reviewed
Direct linkDirect link
Oliveri, Maria; McCaffrey, Daniel; Ezzo, Chelsea; Holtzman, Steven – Applied Measurement in Education, 2017
The assessment of noncognitive traits is challenging due to possible response biases, "subjectivity" and "faking." Standardized third-party evaluations where an external evaluator rates an applicant on their strengths and weaknesses on various noncognitive traits are a promising alternative. However, accurate score-based…
Descriptors: Factor Analysis, Decision Making, College Admission, Likert Scales
Peer reviewed Peer reviewed
Direct linkDirect link
He, Qingping; Stockford, Ian; Meadows, Michelle – Oxford Review of Education, 2018
Results from Rasch analysis of GCSE and GCE A level data over a period of four years suggest that the standards of examinations in different subjects are not consistent in terms of the levels of the latent trait specified in the Rasch model required to achieve the same grades. Variability in statistical standards between subjects exists at both…
Descriptors: Foreign Countries, Exit Examinations, Intellectual Disciplines, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Hua, Anh N.; Keenan, Janice M. – Scientific Studies of Reading, 2017
One of the most important findings to emerge from recent reading comprehension research is that there are large differences between tests in what they assess--specifically, the extent to which performance depends on word recognition versus listening comprehension skills. Because this research used ordinary least squares regression, it is not clear…
Descriptors: Reading Comprehension, Reading Tests, Test Interpretation, Regression (Statistics)
Peer reviewed Peer reviewed
Direct linkDirect link
Zapata-Rivera, Juan Diego; Katz, Irvin R. – Assessment in Education: Principles, Policy & Practice, 2014
Score reports have one or more intended audiences: the people who use the reports to make decisions about test takers, including teachers, administrators, parents and test takers. Attention to audience when designing a score report supports assessment validity by increasing the likelihood that score users will interpret and use assessment results…
Descriptors: Audience Analysis, Scores, Reports, Test Interpretation
Talan, Teri N.; Bloom, Paula Jorde – Teachers College Press, 2018
The "Business Administration Scale for Family Child Care" (BAS) is the first valid and reliable tool for measuring and improving the overall quality of business and professional practices in family child care settings. It is applicable for multiple uses, including program self-improvement, technical assistance and monitoring, training,…
Descriptors: Business Administration, Child Care, Rating Scales, Qualifications
Peer reviewed Peer reviewed
Direct linkDirect link
Newton, Paul E. – Journal of Educational Measurement, 2013
Kane distinguishes between two kinds of argument: the interpretation/use argument and the validity argument. This commentary considers whether there really are two kinds of argument, two arguments, or just one. It concludes that there is just one argument: the validity argument. (Contains 2 figures and 5 notes.)
Descriptors: Validity, Test Interpretation, Test Use
Peer reviewed Peer reviewed
Direct linkDirect link
Popham, W. James – Educational Leadership, 2014
Fifty years ago, Robert Glaser introduced the concept of criterion-referenced measurement in an article in American Psychologist. Its early proponents predicted that this measurement strategy would revolutionize education. But has it lived up to its promise? W. James Popham explores this question by looking at the history of criterion-referenced…
Descriptors: Criterion Referenced Tests, Program Effectiveness, Misconceptions, Test Interpretation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Monroe, Scott; Cai, Li – Grantee Submission, 2015
This research is concerned with two topics in assessing model fit for categorical data analysis. The first topic involves the application of a limited-information overall test, introduced in the item response theory literature, to Structural Equation Modeling (SEM) of categorical outcome variables. Most popular SEM test statistics assess how well…
Descriptors: Structural Equation Models, Test Interpretation, Goodness of Fit, Item Response Theory
Michelle M. Neumann; Jason L. Anthony; Noé A. Erazo; David L. Neumann – Grantee Submission, 2019
The framework and tools used for classroom assessment can have significant impacts on teacher practices and student achievement. Getting assessment right is an important component in creating positive learning experiences and academic success. Recent government reports (e.g., United States, Australia) call for the development of systems that use…
Descriptors: Early Childhood Education, Futures (of Society), Educational Assessment, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Schmidgall, Jonathan – Applied Measurement in Education, 2017
This study utilizes an argument-based approach to validation to examine the implications of reliability in order to further differentiate the concepts of score and decision consistency. In a methodological example, the framework of generalizability theory was used to estimate appropriate indices of score consistency and evaluations of the…
Descriptors: Scores, Reliability, Validity, Generalizability Theory
Hasbrouck, Jan; Tindal, Gerald – Behavioral Research and Teaching, 2017
This paper describes the origins of the widely used curriculum-based measure of oral reading fluency (ORF) and how the creation and use of ORF norms has evolved over time. Norms for ORF can be used to help educators make decisions about which students might need intervention in reading and to help monitor students' progress once instruction has…
Descriptors: Oral Reading, Reading Fluency, Curriculum Based Assessment, Emergent Literacy
Peer reviewed Peer reviewed
Direct linkDirect link
Tengberg, Michael – Language Assessment Quarterly, 2018
Reading comprehension is often treated as a multidimensional construct. In many reading tests, items are distributed over reading process categories to represent the subskills expected to constitute comprehension. This study explores (a) the extent to which specified subskills of reading comprehension tests are conceptually conceivable to…
Descriptors: Reading Tests, Reading Comprehension, Scores, Test Results
Peer reviewed Peer reviewed
Direct linkDirect link
Rupp, André A. – Applied Measurement in Education, 2018
This article discusses critical methodological design decisions for collecting, interpreting, and synthesizing empirical evidence during the design, deployment, and operational quality-control phases for automated scoring systems. The discussion is inspired by work on operational large-scale systems for automated essay scoring but many of the…
Descriptors: Design, Automation, Scoring, Test Scoring Machines
Pages: 1  |  ...  |  9  |  10  |  11  |  12  |  13  |  14  |  15  |  16  |  17  |  ...  |  266