Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 4 |
Descriptor
Difficulty Level | 54 |
Higher Education | 54 |
Item Analysis | 54 |
Test Items | 50 |
Test Construction | 26 |
Multiple Choice Tests | 24 |
Test Reliability | 16 |
Test Validity | 15 |
College Entrance Examinations | 10 |
Test Format | 10 |
Achievement Tests | 7 |
More ▼ |
Source
Author
Publication Type
Reports - Research | 42 |
Journal Articles | 21 |
Speeches/Meeting Papers | 19 |
Reports - Evaluative | 7 |
Reports - Descriptive | 3 |
Information Analyses | 1 |
Tests/Questionnaires | 1 |
Education Level
Higher Education | 4 |
Postsecondary Education | 3 |
Elementary Secondary Education | 1 |
Audience
Researchers | 8 |
Practitioners | 2 |
Teachers | 1 |
Location
Florida | 1 |
Georgia | 1 |
Germany | 1 |
Iraq | 1 |
New Zealand | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 2 |
SAT (College Admission Test) | 2 |
California Achievement Tests | 1 |
Graduate Management Admission… | 1 |
What Works Clearinghouse Rating
Ferrara, Steve; Steedle, Jeffrey T.; Frantz, Roger S. – Applied Measurement in Education, 2022
Item difficulty modeling studies involve (a) hypothesizing item features, or item response demands, that are likely to predict item difficulty with some degree of accuracy; and (b) entering the features as independent variables into a regression equation or other statistical model to predict difficulty. In this review, we report findings from 13…
Descriptors: Reading Comprehension, Reading Tests, Test Items, Item Response Theory
Omarov, Nazarbek Bakytbekovich; Mohammed, Aisha; Alghurabi, Ammar Muhi Khleel; Alallo, Hajir Mahmood Ibrahim; Ali, Yusra Mohammed; Hassan, Aalaa Yaseen; Demeuova, Lyazat; Viktorovna, Shvedova Irina; Nazym, Bekenova; Al Khateeb, Nashaat Sultan Afif – International Journal of Language Testing, 2023
The Multiple-choice (MC) item format is commonly used in educational assessments due to its economy and effectiveness across a variety of content domains. However, numerous studies have examined the quality of MC items in high-stakes and higher-education assessments and found many flawed items, especially in terms of distractors. These faulty…
Descriptors: Test Items, Multiple Choice Tests, Item Response Theory, English (Second Language)
Kirschner, Sophie; Borowski, Andreas; Fischer, Hans E.; Gess-Newsome, Julie; von Aufschnaiter, Claudia – International Journal of Science Education, 2016
Teachers' professional knowledge is assumed to be a key variable for effective teaching. As teacher education has the goal to enhance professional knowledge of current and future teachers, this knowledge should be described and assessed. Nevertheless, only a limited number of studies quantitatively measures physics teachers' professional…
Descriptors: Evaluation Methods, Tests, Test Format, Science Instruction
Ariel, Robert; Dunlosky, John; Bailey, Heather – Journal of Experimental Psychology: General, 2009
Theories of self-regulated study assume that learners monitor item difficulty when making decisions about which items to select for study. To complement such theories, the authors propose an agenda-based regulation (ABR) model in which learners' study decisions are guided by an agenda that learners develop to prioritize items for study, given…
Descriptors: Test Items, Time Management, Item Analysis, Rewards

Tollefson, Nona – Educational and Psychological Measurement, 1987
This study compared the item difficulty, item discrimination, and test reliability of three forms of multiple-choice items: (1) one correct answer; (2) "none of the above" as a foil; and (3) "none of the above" as the correct answer. Twelve items in the three formats were administered in a college statistics examination. (BS)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests

Barker, Douglas; Ebel, Robert L. – Contemporary Educational Psychology, 1982
Two forms of an undergraduate examination were constructed. Tests varied with respect to item truth value (true, false) and method of phrasing (positive, negative). Negatively stated items were more difficult but not more discriminating than positively stated items. False items were not more difficult but were more discriminating than true items.…
Descriptors: Difficulty Level, Higher Education, Item Analysis, Response Style (Tests)

Green, Kathy – Educational and Psychological Measurement, 1985
Five sets of paired comparison judgments were made concerning test item difficulty, in order to identify the most probable source of intrasensitivity in the data. The paired comparisons method was useful in providing information about sensitivity to stimulus differences, but less useful for assessing dimensionality of judgment criteria.…
Descriptors: Adults, Difficulty Level, Evaluative Thinking, Higher Education

Willoughby, T. Lee – Educational and Psychological Measurement, 1980
The reliability and validity of a priori estimates of item characteristics are assessed. Results suggest that judges can make a modest contribution to estimation prior to actual administration. (Author/GK)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Medical School Faculty

Chalifour, Clark L.; Powers, Donald E. – Journal of Educational Measurement, 1989
Content characteristics of 1,400 Graduate Record Examination (GRE) analytical reasoning items were coded for item difficulty and discrimination. The results provide content characteristics for consideration in extending specifications for analytical reasoning items and a better understanding of the construct validity of these items. (TJH)
Descriptors: College Entrance Examinations, Construct Validity, Content Analysis, Difficulty Level
Prestwood, J. Stephen; Weiss, David J. – 1977
The accuracy with which testees perceived the difficulty of ability-test items was investigated by administering two 41-item conventional tests on verbal ability. High- and low-ability groups responded to test items by choosing the correct alternative and then rating each item's difficulty relative to their levels of ability. Least-squares…
Descriptors: Ability, Difficulty Level, Higher Education, Item Analysis

Green, Kathy E. – Educational and Psychological Measurement, 1983
This study was concerned with the reliability and validity of subjective judgments about five characteristics of multiple-choice test items from an introductory college-level astronomy test: (1) item difficulty, (2) language complexity, (3) content importance or relevance, (4) response set convergence, and (5) process complexity. (Author)
Descriptors: Achievement Tests, Astronomy, Difficulty Level, Evaluative Thinking
Tollefson, Nona; Chen, Ju Shan – 1986
This study compared item difficulty and item discrimination indices for parallel multiple-choice items in three content areas: measurement concepts, statistical terminology, and synonyms. The statistics and measurement items were administered in classes where graduate students taking the test were studying the content. Vocabulary items represented…
Descriptors: Difficulty Level, Graduate Students, Higher Education, Item Analysis
Sinnott, Loraine T. – 1982
A standard method for exploring item bias is the intergroup comparison of item difficulties. This paper describes a refinement and generalization of this technique. In contrast to prior approaches, the proposed method deletes outlying items from the formulation of a criterion for identifying items as deviant. It also extends the mathematical…
Descriptors: College Entrance Examinations, Difficulty Level, Higher Education, Item Analysis
Weiten, Wayne – 1979
Two different formats for multiple-choice test items were compared in an experimental test given in a college class in introductory psychology. In one format, a question or incomplete statement was followed by four answers or completions, only one of which was correct. In the other format, the double multiple-choice version, the same questions…
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests

Huck, Schuyler W. – Journal of Educational Measurement, 1978
Providing examinees with advanced knowledge of the difficulty of an item led to an increase in test performance with no loss of reliability. This finding was consistent across several test formats. ( Author/JKS)
Descriptors: Difficulty Level, Feedback, Higher Education, Item Analysis