NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
No Child Left Behind Act 20011
What Works Clearinghouse Rating
Showing 1 to 15 of 26 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Anja Riemenschneider; Zarah Weiss; Pauline Schröter; Detmar Meurers – TESOL Quarterly: A Journal for Teachers of English to Speakers of Other Languages and of Standard English as a Second Dialect, 2024
The linguistic characteristics of text productions depend on various factors, including individual language proficiency as well as the tasks used to elicit the production. To date, little attention has been paid to whether some writing tasks are more suitable than others to represent and differentiate students' proficiency levels. This issue is…
Descriptors: English (Second Language), Writing (Composition), Difficulty Level, Language Proficiency
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Inga Laukaityte; Marie Wiberg – Practical Assessment, Research & Evaluation, 2024
The overall aim was to examine effects of differences in group ability and features of the anchor test form on equating bias and the standard error of equating (SEE) using both real and simulated data. Chained kernel equating, Postratification kernel equating, and Circle-arc equating were studied. A college admissions test with four different…
Descriptors: Ability Grouping, Test Items, College Entrance Examinations, High Stakes Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Rafatbakhsh, Elaheh; Ahmadi, Alireza – Practical Assessment, Research & Evaluation, 2022
The purpose of this study was to investigate the validity of the vocabulary subsection of a high-stakes university entrance exam for Ph.D. programs using the argument-based approach. All the three different versions of the test administered in a period of five years and the responses of 12,500 test-takers were studied. The study focused on four…
Descriptors: Vocabulary, College Entrance Examinations, Doctoral Programs, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Yumei Zou; Sathiamoorthy Kannan; Gurnam Kaur Sidhu – SAGE Open, 2024
Task design has been viewed to be essential in the context of language assessment. This study investigated whether increasing task complexity affects learners' writing performance. It employs three writing tasks with different levels of complexity based on Robinson's Componential Framework. A cohort of 278 participants was selected using a simple…
Descriptors: Difficulty Level, College Students, Foreign Countries, Writing Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Becker, Benjamin; van Rijn, Peter; Molenaar, Dylan; Debeer, Dries – Assessment & Evaluation in Higher Education, 2022
A common approach to increase test security in higher educational high-stakes testing is the use of different test forms with identical items but different item orders. The effects of such varied item orders are relatively well studied, but findings have generally been mixed. When multiple test forms with different item orders are used, we argue…
Descriptors: Information Security, High Stakes Tests, Computer Security, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Lions, Séverin; Dartnell, Pablo; Toledo, Gabriela; Godoy, María Inés; Córdova, Nora; Jiménez, Daniela; Lemarié, Julie – Educational and Psychological Measurement, 2023
Even though the impact of the position of response options on answers to multiple-choice items has been investigated for decades, it remains debated. Research on this topic is inconclusive, perhaps because too few studies have obtained experimental data from large-sized samples in a real-world context and have manipulated the position of both…
Descriptors: Multiple Choice Tests, Test Items, Item Analysis, Responses
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Omarov, Nazarbek Bakytbekovich; Mohammed, Aisha; Alghurabi, Ammar Muhi Khleel; Alallo, Hajir Mahmood Ibrahim; Ali, Yusra Mohammed; Hassan, Aalaa Yaseen; Demeuova, Lyazat; Viktorovna, Shvedova Irina; Nazym, Bekenova; Al Khateeb, Nashaat Sultan Afif – International Journal of Language Testing, 2023
The Multiple-choice (MC) item format is commonly used in educational assessments due to its economy and effectiveness across a variety of content domains. However, numerous studies have examined the quality of MC items in high-stakes and higher-education assessments and found many flawed items, especially in terms of distractors. These faulty…
Descriptors: Test Items, Multiple Choice Tests, Item Response Theory, English (Second Language)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lyness, Scott A.; Peterson, Kent; Yates, Kenneth – Education Sciences, 2021
The Performance Assessment for California Teachers (PACT) is a high stakes summative assessment that was designed to measure pre-service teacher readiness. We examined the inter-rater reliability (IRR) of trained PACT evaluators who rated 19 candidates. As measured by Cohen's weighted kappa, the overall IRR estimate was 0.17 (poor strength of…
Descriptors: High Stakes Tests, Performance Based Assessment, Teacher Effectiveness, Academic Language
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Bazvand, Ali Darabi; Kheirzadeh, Shiela; Ahmadi, Alireza – International Journal of Assessment Tools in Education, 2019
The findings of previous research into the compatibility of stakeholders' perceptions with statistical estimations of item difficulty are not seemingly consistent. Furthermore, most research shows that teachers' estimation of item difficulty is not reliable since they tend to overestimate the difficulty of easy items and underestimate the…
Descriptors: Foreign Countries, High Stakes Tests, Test Items, Difficulty Level
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Hohensinn, Christine; Baghaei, Purya – Psicologica: International Journal of Methodology and Experimental Psychology, 2017
In large scale multiple-choice (MC) tests alternate forms of a test may be developed to prevent cheating by changing the order of items or by changing the position of the response options. The assumption is that since the content of the test forms are the same the order of items or the positions of the response options do not have any effect on…
Descriptors: Multiple Choice Tests, Test Format, Test Items, Difficulty Level
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Yu, Xiaoli – International Journal of Language Testing, 2021
This study examined the development of text complexity for the past 25 years of reading comprehension passages in the National Matriculation English Test (NMET) in China. Text complexity of 206 reading passages at lexical, syntactic, and discourse levels has been measured longitudinally and compared across the years. The natural language…
Descriptors: Reading Comprehension, Reading Tests, Difficulty Level, Natural Language Processing
Peer reviewed Peer reviewed
Direct linkDirect link
Saif, Shahrzad; Ma, Jia; May, Lyn; Cheng, Liying – Assessment in Education: Principles, Policy & Practice, 2021
Effective test preparation for high-stakes English language tests is crucial for candidates whose futures depend on attaining a particular score. An increasing number of studies have investigated the role of test preparation; however, these studies have been exclusively conducted in individual contexts and countries around the world. This study…
Descriptors: Difficulty Level, Test Preparation, High Stakes Tests, English (Second Language)
Peer reviewed Peer reviewed
Direct linkDirect link
Ramsay-Jordan, Natasha – International Journal of Educational Reform, 2020
The most highlighted provision and consequence of the reauthorization of the 1965 Elementary and Secondary Education Act (ESEA), the No Child Left Behind (NCLB) Act of 2001, is obsessive practices of assessing students across the United States (U.S.). Despite newly named policies, including Every Student Succeed Act (ESSA) of 2015, which governs…
Descriptors: Preservice Teachers, Culturally Relevant Education, Standardized Tests, Teacher Education
Peer reviewed Peer reviewed
Direct linkDirect link
Jin, Yan; Yan, Ming – Language Assessment Quarterly, 2017
One major threat to validity in high-stakes testing is construct-irrelevant variance. In this study we explored whether the transition from a paper-and-pencil to a computer-based test mode in a high-stakes test in China, the College English Test, has brought about variance irrelevant to the construct being assessed in this test. Analyses of the…
Descriptors: Writing Tests, Computer Assisted Testing, Computer Literacy, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Baghaei, Purya; Ravand, Hamdollah – SAGE Open, 2019
In many reading comprehension tests, different test formats are employed. Two commonly used test formats to measure reading comprehension are sustained passages followed by some questions and cloze items. Individual differences in handling test format peculiarities could constitute a source of score variance. In this study, a bifactor Rasch model…
Descriptors: Cloze Procedure, Test Bias, Individual Differences, Difficulty Level
Previous Page | Next Page »
Pages: 1  |  2