Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedBurkhart, Barry R.; And Others – Journal of Clinical Psychology, 1980
Obvious items were good predictors of all criteria; neutral items overlapped considerably with obvious items; and subtle items generally did not contribute uniquely to the prediction of any of the concurrent measures of depression. (Author)
Descriptors: Depression (Psychology), Item Analysis, Personality Measures, Predictive Validity
Peer reviewedBerk, Ronald A. – Journal of Experimental Education, 1980
A sampling methodology is proposed for determining lengths of tests designed to assess the comprehension of written discourse. It is based on Bormuth's transformational analysis, within a domain-referenced framework. Guidelines are provided for computing sample size and selecting sentences to which the transformational rules can be applied.…
Descriptors: Reading Comprehension, Reading Tests, Sampling, Test Construction
Peer reviewedPlake, Barbara S.; Hoover, H. D. – Journal of Experimental Education, 1979
A follow-up technique is needed to identify items contributing to items-by-groups interaction when using an ANOVA procedure to examine a test for biased items. The method described includes distribution theory for assessing level of significance and is sensitive to items at all difficulty levels. (Author/GSK)
Descriptors: Analysis of Variance, Goodness of Fit, Item Analysis, Statistical Bias
Peer reviewedGower, D. M.; Daniels, D. J. – Studies in Educational Evaluation, 1980
The facility index of a test item is the percentage of testees choosing the correct response. A common curriculum model presents process and content as two dimensions of the difficulty of an educational task. Wheeler suggested a third dimension: precision. This three-dimensional model is applied to chemistry education. (BW)
Descriptors: Chemistry, Curriculum Design, Difficulty Level, Foreign Countries
Peer reviewedSirotnik, Kenneth A.; And Others – Studies in Educational Evaluation, 1980
Methodological issues are illuminated by reference to analyses of environment data. Analyses include clustering techniques applied at the item level instead of the conventional scale level. Clusters derived using different units of analysis (namely, between and pooled within) are compared and contrasted. (RL)
Descriptors: Attitude Measures, Classroom Environment, Cluster Analysis, Factor Analysis
Peer reviewedHarris, Chester W.; Pearlman, Andrea Pastorok – Journal of Educational Statistics, 1978
A theory and a procedure are presented for estimating a domain parameter and item parameters for test items in a homogeneous domain, such that the combined domain and item parameters account for observed proportions right for each item in a test. (CTM)
Descriptors: Achievement Tests, Difficulty Level, Item Analysis, Mathematical Models
Peer reviewedWadkins, J. R. Jefferson – American Mathematical Monthly, 1978
Some background information is given about the GRE. A detailed account of its construction, its recent history, and some of the thinking that has gone into it is related. (MP)
Descriptors: College Mathematics, Graduate Study, Higher Education, Standardized Tests
Peer reviewedZacharias, Jerrold R. – National Elementary Principal, 1979
Outlines a method for constructing minimum competency tests that are legitimate, answerable, and useful. At the heart of such tests is an item bank that is open to all to view. (Author/IRT)
Descriptors: Academic Standards, Elementary Secondary Education, Item Banks, Minimum Competency Testing
Peer reviewedQuereshi, M. Y.; Fisher, Thomas L. – Educational and Psychological Measurement, 1977
Logical estimates of item difficulty made by judges were compared to empirical estimates derived from a test administration. Results indicated substantial correspondence between logical and empirical estimates, and substantial variation among judges. Further, the more elaborate the system used by judges to make estimates, the more accurate the…
Descriptors: Court Judges, Difficulty Level, Evaluation Methods, Item Analysis
Peer reviewedMartinez, Michael E.; Katz, Irvin R. – Educational Assessment, 1996
Item level differences between a type of constructed response item (figural response) and comparable multiple choice items in the domain of architecture were studied. Data from 120 architects and architecture students show that item level differences in difficulty correspond to differences in cognitive processing requirements and that relations…
Descriptors: Architects, Architecture, Cognitive Processes, Constructed Response
Peer reviewedMeel, David E. – Educational Assessment, 1997
Four issues associated with calculator-available assessments are discussed: (1) benefits and problems of calculator use; (2) components of various types of calculator-available assessments; (3) construction of calculator-available assessments; and (4) the need for additional examination of items included on calculator-available assessments. (SLD)
Descriptors: Calculators, Elementary Secondary Education, Higher Education, Mathematics Tests
Peer reviewedFitzpatrick, Anne R.; And Others – Journal of Educational Measurement, 1996
One-parameter (1PPC) and two-parameter partial credit (2PPC) models were compared using real and simulated data with constructed response items present. Results suggest that the more flexible three-parameter logistic-2PPC model combination produces better model fit than the combination of the one-parameter logistic and the 1PPC models. (SLD)
Descriptors: Comparative Analysis, Constructed Response, Goodness of Fit, Performance Based Assessment
Peer reviewedO Cinneide, Barra – Journal of European Industrial Training, 1997
In a theory/lecture course with limited use of case method, it is possible to get students to choose case study questions on examinations if topics are interesting, attention directing, and well understood. Case study test questions give students opportunities for critical analysis and self-expression. (SK)
Descriptors: Academic Achievement, Business Education, Case Studies, Graduate Students
Peer reviewedHoran, Patrick M.; DiStefano, Christine; Motl, Robert W. – Structural Equation Modeling, 2003
Examined the analysis and interpretation of wording effects associated with positively and negatively worded survey items using data from a national sample of secondary school students (National Education Longitudinal Study 1988-1994). Results for 14,374 students support the interpretation of wording effects in terms of personality traits with…
Descriptors: Personality Traits, Secondary Education, Secondary School Students, Self Esteem
Peer reviewedMeijer, Rob R. – Journal of Educational Measurement, 2002
Used empirical data from a certification test to study methods from statistical process control that have been proposed to classify an item score pattern as fitting or misfitting the underlying item response theory model in computerized adaptive testing. Results for 1,392 examinees show that different types of misfit can be distinguished. (SLD)
Descriptors: Certification, Classification, Goodness of Fit, Item Response Theory


