Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 4 |
Descriptor
Difficulty Level | 43 |
Higher Education | 43 |
Test Reliability | 43 |
Test Items | 30 |
Multiple Choice Tests | 19 |
Item Analysis | 16 |
Test Construction | 16 |
Test Validity | 16 |
Test Format | 10 |
College Students | 8 |
Comparative Analysis | 7 |
More ▼ |
Source
Author
Publication Type
Education Level
Higher Education | 4 |
Postsecondary Education | 4 |
Elementary Secondary Education | 1 |
Audience
Practitioners | 2 |
Researchers | 2 |
Teachers | 1 |
Location
Australia | 2 |
Florida | 2 |
Turkey | 2 |
Asia | 1 |
Brazil | 1 |
Connecticut | 1 |
Denmark | 1 |
Egypt | 1 |
Estonia | 1 |
Georgia | 1 |
Germany | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Embedded Figures Test | 1 |
Graduate Management Admission… | 1 |
Students Evaluation of… | 1 |
What Works Clearinghouse Rating
COVID-19 Lockdown Effects on Student Grades of a University Engineering Course: A Psychometric Study
Santos, Hernan – IEEE Transactions on Education, 2022
Contribution: This article is centered on effects that the COVID-19 lockdown has produced on the student performance in specific engineering course. The study treats to evaluate if the changes in the teaching and student assessment have been suited. Background: Most of higher education courses have had to adapt to this situation and made quick…
Descriptors: COVID-19, Pandemics, Outcomes of Education, Educational Change
Omarov, Nazarbek Bakytbekovich; Mohammed, Aisha; Alghurabi, Ammar Muhi Khleel; Alallo, Hajir Mahmood Ibrahim; Ali, Yusra Mohammed; Hassan, Aalaa Yaseen; Demeuova, Lyazat; Viktorovna, Shvedova Irina; Nazym, Bekenova; Al Khateeb, Nashaat Sultan Afif – International Journal of Language Testing, 2023
The Multiple-choice (MC) item format is commonly used in educational assessments due to its economy and effectiveness across a variety of content domains. However, numerous studies have examined the quality of MC items in high-stakes and higher-education assessments and found many flawed items, especially in terms of distractors. These faulty…
Descriptors: Test Items, Multiple Choice Tests, Item Response Theory, English (Second Language)
Uzuner Yurt, Serap; Aktas, Elif – Educational Research and Reviews, 2016
In this study, the effects of the use of peer tutoring in Effective and Good Speech Course on students' success, perception of speech self-efficacy and speaking skills were examined. The study, designed as a mixed pattern in which quantitative and qualitative research approaches were combined, was carried out together with 57 students in 2014 to…
Descriptors: Peer Teaching, Tutoring, Higher Education, College Students

Tollefson, Nona – Educational and Psychological Measurement, 1987
This study compared the item difficulty, item discrimination, and test reliability of three forms of multiple-choice items: (1) one correct answer; (2) "none of the above" as a foil; and (3) "none of the above" as the correct answer. Twelve items in the three formats were administered in a college statistics examination. (BS)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests

Weiten, Wayne – Journal of Experimental Education, 1982
A comparison of double as opposed to single multiple-choice questions yielded significant differences in regard to item difficulty, item discrimination, and internal reliability, but not concurrent validity. (Author/PN)
Descriptors: Difficulty Level, Educational Testing, Higher Education, Multiple Choice Tests

Willoughby, T. Lee – Educational and Psychological Measurement, 1980
The reliability and validity of a priori estimates of item characteristics are assessed. Results suggest that judges can make a modest contribution to estimation prior to actual administration. (Author/GK)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Medical School Faculty

Newbould, Charles A. – Assessment and Evaluation in Higher Education, 1981
Test data were used to compare the grading of two forms of double mathematics: pure and applied math, and regular and advanced math. Results confirm expectations that in the former system, the grading is comparable, and in the latter, it is not necessarily comparable. Implications for student admission are discussed. (MSE)
Descriptors: Admission Criteria, College Admission, Difficulty Level, Foreign Countries

Straton, Ralph G.; Catts, Ralph M. – Educational and Psychological Measurement, 1980
Multiple-choice tests composed entirely of two-, three-, or four-choice items were investigated. Results indicated that number of alternatives per item was inversely related to item difficulty, but directly related to item discrimination. Reliability and standard error of measurement of three-choice item tests was equivalent or superior.…
Descriptors: Difficulty Level, Error of Measurement, Foreign Countries, Higher Education
Sinnott, Loraine T. – 1982
A standard method for exploring item bias is the intergroup comparison of item difficulties. This paper describes a refinement and generalization of this technique. In contrast to prior approaches, the proposed method deletes outlying items from the formulation of a criterion for identifying items as deviant. It also extends the mathematical…
Descriptors: College Entrance Examinations, Difficulty Level, Higher Education, Item Analysis
Weiten, Wayne – 1979
Two different formats for multiple-choice test items were compared in an experimental test given in a college class in introductory psychology. In one format, a question or incomplete statement was followed by four answers or completions, only one of which was correct. In the other format, the double multiple-choice version, the same questions…
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests

Huck, Schuyler W. – Journal of Educational Measurement, 1978
Providing examinees with advanced knowledge of the difficulty of an item led to an increase in test performance with no loss of reliability. This finding was consistent across several test formats. ( Author/JKS)
Descriptors: Difficulty Level, Feedback, Higher Education, Item Analysis

Weiten, Wayne – Journal of Experimental Education, 1984
The effects of violating four item construction principles were examined to assess the validity of the principles and the importance of students' test wiseness. While flawed items were significantly less difficult than sound items, differences in item discrimination, test reliability, and concurrent validity were not observed. (Author/BW)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests

Pray, W. Stephen; Popovich, Nicholas G. – American Journal of Pharmaceutical Education, 1985
Test development included designing, screening, and field testing of test items; compilation into an examination administered to a target group; and norm development for score comparison with a national sample. (MSE)
Descriptors: Difficulty Level, Doctoral Programs, Higher Education, Item Analysis

Hsu, Tse-Chi; And Others – Journal of Experimental Education, 1984
The indices of item difficulty and discrimination, the coefficients of effective length, and the average item information for both single- and multiple-answer items using six different scoring formulas were computed and compared. These formulas vary in terms of the assignment of partial credit and the correction for guessing. (Author/BW)
Descriptors: College Entrance Examinations, Comparative Analysis, Difficulty Level, Guessing (Tests)
Educational Testing Service, Princeton, NJ. – 1986
The final project report on development of an advanced Russian language listening and reading proficiency test is presented. It summarizes activities in the second year of the project, including dissemination of summer 1985 test validation results to participating higher education institutions, item analyses, completion of the final test edition,…
Descriptors: Advanced Courses, Difficulty Level, Higher Education, Language Proficiency