NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers1
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of…1
What Works Clearinghouse Rating
Showing all 13 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Polat, Murat – International Online Journal of Education and Teaching, 2022
Foreign language testing is a multi-dimensional phenomenon and obtaining objective and error-free scores on learners' language skills is often problematic. While assessing foreign language performance on high-stakes tests, using different testing approaches including Classical Test Theory (CTT), Generalizability Theory (GT) and/or Item Response…
Descriptors: Second Language Learning, Second Language Instruction, Item Response Theory, Language Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ayanwale, Musa Adekunle; Adeleke, Joshua Oluwatoyin; Mamadelo, Titilayo Iyabode – Journal of the International Society for Teacher Education, 2019
A scoring framework that does not reflect true performance of an examinee would ultimately result in an abnormal score. This study assessed invariance person estimates of 2017 Nigerian National Examinations Council Basic Education Certificate Examination Mathematics Multiple Choice using classical test theory (CTT) and item response theory (IRT)…
Descriptors: Test Theory, Item Response Theory, Scoring, National Competency Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Selvi, Hüseyin; Özdemir Alici, Devrim – International Journal of Assessment Tools in Education, 2018
In this study, it is aimed to investigate the impact of different missing data handling methods on the detection of Differential Item Functioning methods (Mantel Haenszel and Standardization methods based on Classical Test Theory and Likelihood Ratio Test method based on Item Response Theory). In this regard, on the data acquired from 1046…
Descriptors: Test Bias, Test Theory, Item Response Theory, Multiple Choice Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Chin, Huan; Chew, Cheng Meng; Lim, Hooi Lian; Thien, Lei Mee – International Journal of Science and Mathematics Education, 2022
Cognitive Diagnostic Assessment (CDA) is an alternative assessment which can give a clear picture of pupils' learning process and cognitive structures to education stakeholders so that appropriate instructional strategies can be designed to tailored pupils' needs. Coincide with this function, the Ordered Multiple-Choice (OMC) items were…
Descriptors: Mathematics Instruction, Mathematics Tests, Multiple Choice Tests, Diagnostic Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Ramsay, James O.; Wiberg, Marie – Journal of Educational and Behavioral Statistics, 2017
This article promotes the use of modern test theory in testing situations where sum scores for binary responses are now used. It directly compares the efficiencies and biases of classical and modern test analyses and finds an improvement in the root mean squared error of ability estimates of about 5% for two designed multiple-choice tests and…
Descriptors: Scoring, Test Theory, Computation, Maximum Likelihood Statistics
Ellis, David P. – ProQuest LLC, 2011
The current version of the International Language Testing Association (ILTA) Guidelines for Practice requires language testers to pretest items before including them on an exam, or when pretesting is not possible, to conduct post-hoc item analysis to ensure any malfunctioning items are excluded from scoring. However, the guidelines are devoid of…
Descriptors: Item Response Theory, High Stakes Tests, College Entrance Examinations, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Chang, Chun-Yen; Yeh, Ting-Kuang; Barufaldi, James P. – International Journal of Science Education, 2010
This study explored the phenomenon of testing effect during science concept assessments, including the mechanism behind it and its impact upon a learner's conceptual understanding. The participants consisted of 208 high school students, in either the 11th or 12th grade. Three types of tests (traditional multiple-choice test, correct concept test,…
Descriptors: Student Attitudes, Multiple Choice Tests, Climate, Misconceptions
Hutchinson, T. P. – 1984
One means of learning about the processes operating in a multiple choice test is to include some test items, called nonsense items, which have no correct answer. This paper compares two versions of a mathematical model of test performance to interpret test data that includes both genuine and nonsense items. One formula is based on the usual…
Descriptors: Foreign Countries, Guessing (Tests), Mathematical Models, Multiple Choice Tests
Peer reviewed Peer reviewed
Jaradat, Derar; Sawaged, Sari – Journal of Educational Measurement, 1986
The impact of the Subset Selection Technique (SST) for multiple-choice items on certain properties of a test was compared with that of two other methods, the Number Right and the Correction for Guessing Formula. Results indicated that SST outperformed the other two, producing higher reliability and validity without favoring high risk takers.…
Descriptors: Foreign Countries, Grade 9, Guessing (Tests), Measurement Techniques
Powell, J. C. – 1980
A multi-faceted model for the selection of answers for multiple-choice tests was developed from the findings of a series of exploratory studies. This model implies that answer selection should be curvilinear. A series of models were tested for fit using the chi square procedure. Data were collected from 359 elementary school students ages 9-12.…
Descriptors: Elementary Education, Foreign Countries, Goodness of Fit, Guessing (Tests)
Peer reviewed Peer reviewed
Budescu, David V.; Nevo, Baruch – Journal of Educational Measurement, 1985
The proportionality model assumes that total testing time is proportional to the number of test items and the number of options per multiple choice test item. This assumption was examined, using test items having from two to five options. The model was not supported. (Author/GDC)
Descriptors: College Entrance Examinations, Foreign Countries, Higher Education, Item Analysis
Peer reviewed Peer reviewed
Barnett-Foster, Debora; Nagy, Philip – Alberta Journal of Educational Research, 1995
Analysis of response strategies employed by 261 undergraduate chemistry students when answering multiple-choice and stem-equivalent constructed-response questions revealed no significant differences in types of solution strategies or types of errors across test format. However, analysis of student oral reports revealed a higher frequency of…
Descriptors: Chemistry, Constructed Response, Educational Research, Educational Testing
van Weeren, J., Ed. – 1983
Presented in this symposium reader are nine papers, four of which deal with the theory and impact of the Rasch model on language testing and five of which discuss final examinations in secondary schools in both general and specific terms. The papers are: "Introduction to Rasch Measurement: Some Implications for Language Testing" (J. J.…
Descriptors: Adolescents, Comparative Analysis, Comparative Education, Difficulty Level