Publication Date
In 2025 | 1 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 9 |
Since 2006 (last 20 years) | 14 |
Descriptor
Comparative Analysis | 40 |
Language Proficiency | 40 |
Test Reliability | 40 |
Language Tests | 33 |
English (Second Language) | 24 |
Test Validity | 23 |
Second Language Learning | 17 |
Foreign Countries | 16 |
Scores | 9 |
Second Language Instruction | 9 |
Test Construction | 9 |
More ▼ |
Source
Author
Publication Type
Education Level
Audience
Laws, Policies, & Programs
Assessments and Surveys
ACTFL Oral Proficiency… | 2 |
English Proficiency Test | 1 |
Michigan Test of English… | 1 |
What Works Clearinghouse Rating
Farshad Effatpanah; Purya Baghaei; Mona Tabatabaee-Yazdi; Esmat Babaii – Language Testing, 2025
This study aimed to propose a new method for scoring C-Tests as measures of general language proficiency. In this approach, the unit of analysis is sentences rather than gaps or passages. That is, the gaps correctly reformulated in each sentence were aggregated as sentence score, and then each sentence was entered into the analysis as a polytomous…
Descriptors: Item Response Theory, Language Tests, Test Items, Test Construction
Ozdemir, Burhanettin; Gelbal, Selahattin – Education and Information Technologies, 2022
The computerized adaptive tests (CAT) apply an adaptive process in which the items are tailored to individuals' ability scores. The multidimensional CAT (MCAT) designs differ in terms of different item selection, ability estimation, and termination methods being used. This study aims at investigating the performance of the MCAT designs used to…
Descriptors: Scores, Computer Assisted Testing, Test Items, Language Proficiency
Xu, Jing; Jones, Edmund; Laxton, Victoria; Galaczi, Evelina – Assessment in Education: Principles, Policy & Practice, 2021
Recent advances in machine learning have made automated scoring of learner speech widespread, and yet validation research that provides support for applying automated scoring technology to assessment is still in its infancy. Both the educational measurement and language assessment communities have called for greater transparency in describing…
Descriptors: Second Language Learning, Second Language Instruction, English (Second Language), Computer Software
Karami, Hossein; Kouhpaee Nejad, Mohammadhossein; Nourzadeh, Saeed; Ahmadi Shirazi, Masoumeh – International Journal of Bilingual Education and Bilingualism, 2020
This study was set to cross-validate a bilingual Persian-English version of the Vocabulary Size Test (VST) against the monolingual English version and compare Iranian EFL learners' performance on the two versions. Various bilingual versions of the VST have been developed based on the assumption that bilingual versions are not affected by the…
Descriptors: Bilingualism, Indo European Languages, English (Second Language), Second Language Learning
Isbell, Dan; Winke, Paula – Language Testing, 2019
The American Council on the Teaching of Foreign Languages (ACTFL) oral proficiency interview -- computer (OPIc) testing system represents an ambitious effort in language assessment: Assessing oral proficiency in over a dozen languages, on the same scale, from virtually anywhere at any time. Especially for users in contexts where multiple foreign…
Descriptors: Oral Language, Language Tests, Language Proficiency, Second Language Learning
Haberman, Shelby J.; Liu, Yang; Lee, Yi-Hsuan – ETS Research Report Series, 2019
Distractor analyses are routinely conducted in educational assessments with multiple-choice items. In this research report, we focus on three item response models for distractors: (a) the traditional nominal response (NR) model, (b) a combination of a two-parameter logistic model for item scores and a NR model for selections of incorrect…
Descriptors: Multiple Choice Tests, Scores, Test Reliability, High Stakes Tests
Longabach, Tanya; Peyton, Vicki – Language Testing, 2018
K-12 English language proficiency tests that assess multiple content domains (e.g., listening, speaking, reading, writing) often have subsections based on these content domains; scores assigned to these subsections are commonly known as subscores. Testing programs face increasing customer demands for the reporting of subscores in addition to the…
Descriptors: Comparative Analysis, Test Reliability, Second Language Learning, Language Proficiency
Thompson, Gregory L.; Cox, Troy L.; Knapp, Nieves – Foreign Language Annals, 2016
While studies have been done to rate the validity and reliability of the Oral Proficiency Interview (OPI) and Oral Proficiency Interview-Computer (OPIc) independently, a limited amount of research has analyzed the interexam reliability of these tests, and studies have yet to be conducted comparing the results of Spanish language learners who take…
Descriptors: Comparative Analysis, Oral Language, Language Proficiency, Spanish
Retnawati, Heri – Turkish Online Journal of Educational Technology - TOJET, 2015
This study aimed to compare the accuracy of the test scores as results of Test of English Proficiency (TOEP) based on paper and pencil test (PPT) versus computer-based test (CBT). Using the participants' responses to the PPT documented from 2008-2010 and data of CBT TOEP documented in 2013-2014 on the sets of 1A, 2A, and 3A for the Listening and…
Descriptors: Scores, Accuracy, Computer Assisted Testing, English (Second Language)
Negari, Giti Mousapour; Azizi, Aliye; Arani, Davood Khedmatkar – International Journal of Instruction, 2018
The present study attempted to investigate the effects of audio input enhancement on EFL learners' retention of intensifiers. To this end, two research questions were formulated. In order to address these research questions, this study attempted to reject two null hypotheses. Pretest-posttest control group quasi-experimental design was employed to…
Descriptors: English (Second Language), Second Language Learning, Second Language Instruction, Quasiexperimental Design
Kucuk, Funda; Walters, JoDee – ELT Journal, 2009
This article reports on a study of the validity and reliability of tests administered in an EFL university setting. The study addresses the question of how well face validity reflects more objective measures of the quality of a test, such as predictive validity and reliability. According to some researchers, face validity, defined as the surface…
Descriptors: Language Tests, Test Validity, Achievement Tests, English (Second Language)

Stansfield, Charles W.; Kenyon, Dorry Mann – System, 1992
Reviews research that sheds light on the comparability of Oral Proficiency Interview and the Simulated Oral Proficiency Interview. Suggestions are provided for further research. (16 references) (VWL)
Descriptors: Comparative Analysis, Interviews, Language Proficiency, Language Tests
Stansfield, Charles W. – 1990
The simulated oral proficiency interview (SOPI) is a semi-direct speaking test that models the format of the oral proficiency interview (OPI). The OPI is a method of assessing general speaking proficiency in a second language. The SOPI is a tape-recorded test consisting of six parts: simple personal background questions posed in a simulated…
Descriptors: Comparative Analysis, Interviews, Language Proficiency, Language Tests

Pang, Lee Yick – International Review of Applied Linguistics in Language Teaching, 1984
Examines and contests the claim that all language tests are in reality testing the same underlying ability which is very similar to the Spearman g-factor for intelligence. Conclusions indicate that the argument for the existence of a g-factor in language tests is not tenable on statistical grounds. (SL)
Descriptors: Comparative Analysis, Intelligence, Language Proficiency, Language Tests

Oller, John W., Jr.; Inal, Nevin – TESOL Quarterly, 1971
Descriptors: Cloze Procedure, Comparative Analysis, Educational Experiments, English (Second Language)