NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 39 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Abashidze, Dato; McDonough, Kim; Gao, Yang – Second Language Research, 2022
Recent research that explored how input exposure and learner characteristics influence novel L2 morphosyntactic pattern learning has exposed participants to either text or static images rather than dynamic visual events. Furthermore, it is not known whether incorporating eye gaze cues into dynamic visual events enhances dual pattern learning.…
Descriptors: Second Language Learning, Second Language Instruction, Language Patterns, Morphology (Languages)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Stewart, Gail; Strachan, Andrea – TESL Canada Journal, 2022
Since its implementation in 2004, the Canadian English Language Benchmark Assessment for Nurses (CELBAN) has been accepted as evidence of language ability for licensure of internationally educated nurses (IENs) in Canada. This article focuses on the complexities of sustaining an occupation-specific assessment over time. The authors reference the…
Descriptors: Language Tests, English for Special Purposes, Benchmarking, Nurses
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chen, Michelle Y.; Flasko, Jennifer J. – Canadian Journal of Applied Linguistics / Revue canadienne de linguistique appliquée, 2020
Seeking evidence to support content validity is essential to test validation. This is especially the case in contexts where test scores are interpreted in relation to external proficiency standards and where new test content is constantly being produced to meet test administration and security demands. In this paper, we describe a modified…
Descriptors: Foreign Countries, Reading Tests, Language Tests, English (Second Language)
Peer reviewed Peer reviewed
Direct linkDirect link
Daub, Olivia; Cardy, Janis Oram; Johnson, Andrew M.; Bagatto, Marlene P. – Journal of Speech, Language, and Hearing Research, 2019
Purpose: This study reports validity evidence for an English translation of the LittlEARS Early Speech Production Questionnaire (LEESPQ). The LEESPQ was designed to support early spoken language outcome monitoring in young children who are deaf/hard of hearing. Methods: Data from 90 children with normal hearing, ages 0-18 months, are reported.…
Descriptors: Young Children, Deafness, Hearing Impairments, Oral Language
Peer reviewed Peer reviewed
Direct linkDirect link
Buono, Stephanie; Jang, Eunice Eunhee – Educational Assessment, 2021
Increasing linguistic diversity in classrooms has led researchers to examine the validity and fairness of standardized achievement tests, specifically concerning whether test score interpretations are free of bias and score use is fair for all students. This study examined whether mathematics achievement test items that contain complex language…
Descriptors: English Language Learners, Standardized Tests, Achievement Tests, Culture Fair Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Kam, Chester Chun Seng – Sociological Methods & Research, 2018
The item wording (or keying) effect is respondents' differential response style to positively and negatively worded items. Despite decades of research, the nature of the effect is still unclear. This article proposes a potential reason; namely, that the item wording effect is scale-specific, and thus findings are applicable only to a particular…
Descriptors: Response Style (Tests), Test Items, Language Usage, College Students
Peer reviewed Peer reviewed
Direct linkDirect link
Guo, Hongwen; Ercikan, Kadriye – Educational Research and Evaluation, 2020
Rapid response behaviour, a type of test disengagement, cannot be interpreted as a true indicator of the targeted constructs and may compromise score accuracy as well as score validity for interpretation. Rapid responding may be due to multiple factors for diverse populations. In this study, using Programme for International Student Assessment…
Descriptors: Response Style (Tests), Foreign Countries, International Assessment, Achievement Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Reckase, Mark D.; Xu, Jing-Ru – Educational and Psychological Measurement, 2015
How to compute and report subscores for a test that was originally designed for reporting scores on a unidimensional scale has been a topic of interest in recent years. In the research reported here, we describe an application of multidimensional item response theory to identify a subscore structure in a test designed for reporting results using a…
Descriptors: English, Language Skills, English Language Learners, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Olney, Andrew M.; Pavlik, Philip I., Jr.; Maass, Jaclyn K. – Grantee Submission, 2017
This study investigated the effect of cloze item practice on reading comprehension, where cloze items were either created by humans, by machine using natural language processing techniques, or randomly. Participants from Amazon Mechanical Turk (N = 302) took a pre-test, read a text, and took part in one of five conditions, Do-Nothing, Re-Read,…
Descriptors: Reading Improvement, Reading Comprehension, Prior Learning, Cloze Procedure
Peer reviewed Peer reviewed
Direct linkDirect link
Roth, Wolff-Michael; Oliveri, Maria Elena; Sandilands, Debra Dallie; Lyons-Thomas, Juliette; Ercikan, Kadriye – International Journal of Science Education, 2013
Even if national and international assessments are designed to be comparable, subsequent psychometric analyses often reveal differential item functioning (DIF). Central to achieving comparability is to examine the presence of DIF, and if DIF is found, to investigate its sources to ensure differentially functioning items that do not lead to bias.…
Descriptors: Test Bias, Evaluation Methods, Protocol Analysis, Science Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Goldhammer, Frank; Martens, Thomas; Lüdtke, Oliver – Large-scale Assessments in Education, 2017
Background: A potential problem of low-stakes large-scale assessments such as the Programme for the International Assessment of Adult Competencies (PIAAC) is low test-taking engagement. The present study pursued two goals in order to better understand conditioning factors of test-taking disengagement: First, a model-based approach was used to…
Descriptors: Student Evaluation, International Assessment, Adults, Competence
Peer reviewed Peer reviewed
Direct linkDirect link
Rogers, W. Todd; Lin, Jie; Rinaldi, Christia M. – Applied Measurement in Education, 2011
The evidence gathered in the present study supports the use of the simultaneous development of test items for different languages. The simultaneous approach used in the present study involved writing an item in one language (e.g., French) and, before moving to the development of a second item, translating the item into the second language (e.g.,…
Descriptors: Test Items, Item Analysis, Achievement Tests, French
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Youn-Hee; Jang, Eunice Eunhee – Language Learning, 2009
The increasing numbers of English language learners (ELLs) in Canadian schools pose a significant challenge to the standards-based provincial tests used to measure proficiency levels of all students from various linguistic and cultural backgrounds. This study investigated the extent to which reading item bundles or items on the Ontario Secondary…
Descriptors: Literacy Education, Multilingualism, Second Language Learning, Foreign Countries
Huang, Xiaoting – ProQuest LLC, 2010
In recent decades, the use of large-scale standardized international assessments has increased drastically as a way to evaluate and compare the quality of education across countries. In order to make valid international comparisons, the primary requirement is to ensure the measurement equivalence between the different language versions of these…
Descriptors: Test Bias, Comparative Testing, Foreign Countries, Measurement
Sawaki, Yasuyo; Nissan, Susan – Educational Testing Service, 2009
The study investigated the criterion-related validity of the "Test of English as a Foreign Language"[TM] Internet-based test (TOEFL[R] iBT) Listening section by examining its relationship to a criterion measure designed to reflect language-use tasks that university students encounter in everyday academic life: listening to academic…
Descriptors: Test Validity, Language Tests, English (Second Language), Computer Assisted Testing
Previous Page | Next Page »
Pages: 1  |  2  |  3