NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 46 to 60 of 3,982 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Viola Merhof; Caroline M. Böhm; Thorsten Meiser – Educational and Psychological Measurement, 2024
Item response tree (IRTree) models are a flexible framework to control self-reported trait measurements for response styles. To this end, IRTree models decompose the responses to rating items into sub-decisions, which are assumed to be made on the basis of either the trait being measured or a response style, whereby the effects of such person…
Descriptors: Item Response Theory, Test Interpretation, Test Reliability, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Micir, Ian; Swygert, Kimberly; D'Angelo, Jean – Journal of Applied Testing Technology, 2022
The interpretations of test scores in secure, high-stakes environments are dependent on several assumptions, one of which is that examinee responses to items are independent and no enemy items are included on the same forms. This paper documents the development and implementation of a C#-based application that uses Natural Language Processing…
Descriptors: Artificial Intelligence, Man Machine Systems, Accuracy, Efficiency
Peer reviewed Peer reviewed
Direct linkDirect link
Leventhal, Brian C.; Gregg, Nikole; Ames, Allison J. – Measurement: Interdisciplinary Research and Perspectives, 2022
Response styles introduce construct-irrelevant variance as a result of respondents systematically responding to Likert-type items regardless of content. Methods to account for response styles through data analysis as well as approaches to mitigating the effects of response styles during data collection have been well-documented. Recent approaches…
Descriptors: Response Style (Tests), Item Response Theory, Test Items, Likert Scales
Peer reviewed Peer reviewed
Direct linkDirect link
Shadi Noroozi; Hossein Karami – Language Testing in Asia, 2024
Recently, psychometricians and researchers have voiced their concern over the exploration of language test items in light of Messick's validation framework. Validity has been central to test development and use; however, it has not received due attention in language tests having grave consequences for test takers. The present study sought to…
Descriptors: Foreign Countries, Doctoral Students, Graduate Students, Language Proficiency
Peer reviewed Peer reviewed
Direct linkDirect link
Kranzler, John H.; Maki, Kathrin E.; Benson, Nicholas F.; Eckert, Tanya L.; Floyd, Randy G.; Fefer, Sarah A. – Contemporary School Psychology, 2020
Although intelligence tests are among the most widely used psychological instruments in school psychology, at the current time, little is known about how practitioners interpret them. The primary purpose of this study, therefore, was to determine how intelligence tests are interpreted by school psychologists, particularly for the identification of…
Descriptors: School Counselors, Test Interpretation, Intelligence Tests, Disability Identification
Peer reviewed Peer reviewed
Direct linkDirect link
Stephen M. Leach; Jason C. Immekus; Jeffrey C. Valentine; Prathiba Batley; Dena Dossett; Tamara Lewis; Thomas Reece – Assessment for Effective Intervention, 2025
Educators commonly use school climate survey scores to inform and evaluate interventions for equitably improving learning and reducing educational disparities. Unfortunately, validity evidence to support these (and other) score uses often falls short. In response, Whitehouse et al. proposed a collaborative, two-part validity testing framework for…
Descriptors: School Surveys, Measurement, Hierarchical Linear Modeling, Educational Environment
Sascha Skucek – ProQuest LLC, 2022
When you look at an image, what do you see? What does the image say to you? What do you think about? What meaning do you infer? These questions may blur together, but they can be expanded individually and uniquely into a multitude of responses. Your initial thoughts are yours. You are silently debating meaning within yourself. If I interject a new…
Descriptors: Rhetoric, Listening, Freehand Drawing, Notetaking
Peer reviewed Peer reviewed
Direct linkDirect link
Eirini M. Mitropoulou; Leonidas A. Zampetakis; Ioannis Tsaousis – Evaluation Review, 2024
Unfolding item response theory (IRT) models are important alternatives to dominance IRT models in describing the response processes on self-report tests. Their usage is common in personality measures, since they indicate potential differentiations in test score interpretation. This paper aims to gain a better insight into the structure of trait…
Descriptors: Foreign Countries, Adults, Item Response Theory, Personality Traits
Peer reviewed Peer reviewed
Direct linkDirect link
Kho, Shermaine Qi En; Aryadoust, Vahid; Foo, Stacy – Education and Information Technologies, 2023
Studies have shown that test-takers tend to use keyword-matching strategies when taking listening tests. Keyword-matching involves matching content words in the written modality (test items) against those heard in the audio text. However, no research has investigated the effect of such keywords in listening tests, or the impact of gazing upon…
Descriptors: Eye Movements, Test Wiseness, Information Retrieval, Listening Comprehension Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Xu; Ouyang, Fan; Liu, Jianwen; Wei, Chengkun; Chen, Wenzhi – Journal of Educational Computing Research, 2023
The computer-supported writing assessment (CSWA) has been widely used to reduce instructor workload and provide real-time feedback. Interpretability of CSWA draws extensive attention because it can benefit the validity, transparency, and knowledge-aware feedback of academic writing assessments. This study proposes a novel assessment tool,…
Descriptors: Computer Assisted Testing, Writing Evaluation, Feedback (Response), Natural Language Processing
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A. – Educational Measurement: Issues and Practice, 2020
Researchers have documented the impact of rater effects, or raters' tendencies to give different ratings than would be expected given examinee achievement levels, in performance assessments. However, the degree to which rater effects influence person fit, or the reasonableness of test-takers' achievement estimates given their response patterns,…
Descriptors: Performance Based Assessment, Evaluators, Achievement, Influences
Peer reviewed Peer reviewed
Direct linkDirect link
Clark, Amy K.; Karvonen, Meagan – Educational Assessment, 2020
Alternate assessments based on alternate achievement standards (AA-AAS) have historically lacked broad validity evidence and an overall evaluation of the extent to which evidence supports intended uses of results. An expanding body of validation literature, the funding of two AA-AAS consortia, and advances in computer-based assessment have…
Descriptors: Alternative Assessment, Test Validity, Test Use, Students with Disabilities
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kannan, Priya; Zapata-Rivera, Diego; Bryant, Andrew D. – Practical Assessment, Research & Evaluation, 2021
Individual-student score reports sometimes include information about precision of scores (i.e., measurement error). In this study, we specifically investigated if parents understand this information when presented. We conducted an online experimental study where 196 parents of middle school children, from various parts of the country, were…
Descriptors: Comprehension, Parents, Error of Measurement, Test Interpretation
Peer reviewed Peer reviewed
Direct linkDirect link
An, Lily Shiao; Ho, Andrew Dean; Davis, Laurie Laughlin – Educational Measurement: Issues and Practice, 2022
Technical documentation for educational tests focuses primarily on properties of individual scores at single points in time. Reliability, standard errors of measurement, item parameter estimates, fit statistics, and linking constants are standard technical features that external stakeholders use to evaluate items and individual scale scores.…
Descriptors: Documentation, Scores, Evaluation Methods, Longitudinal Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Smith, Leann V.; Graves, Scott L. – Contemporary School Psychology, 2021
The purpose of this paper is to examine the factorial invariance of the Wechsler Intelligence Scale for Children--Fifth Edition (WISC-V) between genders in a sample of Black students in an urban, public school district. Few researchers test the validity of cognitive assessments on Black samples and even fewer do so utilizing samples other than…
Descriptors: Children, Intelligence Tests, African American Students, Urban Schools
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  266