Publication Date
| In 2026 | 0 |
| Since 2025 | 197 |
| Since 2022 (last 5 years) | 1067 |
| Since 2017 (last 10 years) | 2577 |
| Since 2007 (last 20 years) | 4938 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedHarasym, P. H.; And Others – Evaluation and the Health Professions, 1992
Findings from a study with approximately 200 first-year University of Calgary (Canada) nursing students provide evidence that the use of negation (e.g., not, except) should be limited in stems of multiple-choice test items and that a single-response negatively worded item should be converted to a multiple-response positively worded item. (SLD)
Descriptors: College Students, Foreign Countries, Higher Education, Multiple Choice Tests
Peer reviewedMcMurray, Mary Anne; And Others – Journal of Research in Science Teaching, 1991
Reports a study investigating the utility of 52 items, selected from a readily available item pool developed for instructional purposes, when the items are used to measure critical thinking abilities of biology students. The items had reasonably good internal consistency reliability and good concurrent validity. (PR)
Descriptors: Biology, College Science, Critical Thinking, Educational Research
Peer reviewedHoijtink, Herbert; Molenaar, Ivo W. – Psychometrika, 1992
The PARallELogram Analysis (PARELLA) model is a probabilistic parallelogram model that can be used for the measurement of latent attitudes or latent preferences. A method is presented for testing for differential item functioning (DIF) for the PARELLA model using the approach of D. Thissen and others (1988). (SLD)
Descriptors: Attitude Measures, Computer Simulation, Equations (Mathematics), Estimation (Mathematics)
Peer reviewedHaladyna, Thomas M. – Educational Measurement: Issues and Practice, 1992
Context-dependent item sets, containing a subset of test items related to a passage or stimulus, are discussed. A brief review of methods for developing item sets reveals their potential for measuring high-level thinking. Theories and technologies for scoring item sets remain largely experimental. Research needs are discussed. (SLD)
Descriptors: Cognitive Tests, Educational Technology, Licensing Examinations (Professions), Problem Solving
Peer reviewedBeaton, Albert E.; Allen, Nancy L. – Journal of Educational Statistics, 1992
The National Assessment of Educational Progress (NAEP) makes possible comparison of groups of students and provides information about what these groups know and can do. The scale anchoring techniques described in this chapter address the latter purpose. The direct method and the smoothing method of scale anchoring are discussed. (SLD)
Descriptors: Comparative Testing, Educational Assessment, Elementary Secondary Education, Knowledge Level
Peer reviewedLiou, Michelle; Chang, Chih-Hsin – Psychometrika, 1992
An extension is proposed for the network algorithm introduced by C.R. Mehta and N.R. Patel to construct exact tail probabilities for testing the general hypothesis that item responses are distributed according to the Rasch model. A simulation study indicates the efficiency of the algorithm. (SLD)
Descriptors: Algorithms, Computer Simulation, Difficulty Level, Equations (Mathematics)
Peer reviewedCliff, Norman; Donoghue, John R. – Psychometrika, 1992
A test theory using only ordinal assumptions is presented, based on the idea that the test items are a sample from a universe of items. The sum across items of the ordinal relations for a pair of persons on the universe items is analogous to a true score. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Item Response Theory, Item Sampling
Peer reviewedLuecht, Richard M.; Hirsch, Thomas M. – Applied Psychological Measurement, 1992
Derivations of several item selection algorithms for use in fitting test items to target information functions (IFs) are described. These algorithms, which use an average growth approximation of target IFs, were tested by generating six test forms and were found to provide reliable fit. (SLD)
Descriptors: Algorithms, Computer Assisted Testing, Equations (Mathematics), Goodness of Fit
Peer reviewedKelderman, Henk; Rijkes, Carl P. M. – Psychometrika, 1994
A loglinear item response theory (IRT) model is proposed that relates polytomously scored item responses to a multidimensional latent space. The analyst may specify a response function for each response, and each item may have a different number of response categories. Conditional maximum likelihood estimates are derived. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Goodness of Fit, Item Response Theory
Peer reviewedSchwarz, Shirley P.; And Others – Journal of Educational Measurement, 1991
Interviews were conducted with 104 students in masters' level classes to determine their reasons for changing test answers. Subjects previously had been instructed in answer-changing strategies. Most changes were for thought out reasons; few were because of clerical errors. Reconsideration of test items is probably underestimated in…
Descriptors: Achievement Gains, Graduate Students, Guessing (Tests), Higher Education
Stoneall, Linda – Training and Development, 1991
Describes questioning methods trainers can use to uncover training needs (interviews, surveys, test questions, program evaluations). Illustrates the use of questions at the beginning, middle, and end of training sessions. (SK)
Descriptors: Adult Education, Discussion (Teaching Technique), Evaluation Methods, Interviews
Peer reviewedBoekkooi-Timminga, Ellen – Applied Psychological Measurement, 1990
A new test construction model based on the Rasch model is proposed. This model, the cluster-based method, considers groups of interchangeable items rather than individual items and uses integer programing. Results for six test construction problems indicate that the method produces accurate results in small amounts of time. (SLD)
Descriptors: Cluster Analysis, Computer Assisted Testing, Equations (Mathematics), Item Banks
Peer reviewedFrary, Robert B. – Applied Measurement in Education, 1991
The use of the "none-of-the-above" option (NOTA) in 20 college-level multiple-choice tests was evaluated for classes with 100 or more students. Eight academic disciplines were represented, and 295 NOTA and 724 regular test items were used. It appears that the NOTA can be compatible with good classroom measurement. (TJH)
Descriptors: College Students, Comparative Testing, Difficulty Level, Discriminant Analysis
Peer reviewedRost, Jurgen – Applied Psychological Measurement, 1990
Combining Rasch and latent class models is presented as a way to overcome deficiencies and retain the positive features of both. An estimation algorithm is outlined, providing conditional maximum likelihood estimates of item parameters for each class. The model is illustrated with simulated data and real data (n=869 adults). (SLD)
Descriptors: Adults, Algorithms, Computer Simulation, Equations (Mathematics)
Peer reviewedHills, John R. – Educational Measurement: Issues and Practice, 1993
A scenario and accompanying questions and answers are posed to help educators examine possible problems in interpreting a student's test score profile. Profiles developed and used soundly are very helpful, but possible pitfalls in test interpretation must be recognized. (SLD)
Descriptors: Academic Achievement, Educational Assessment, Elementary Secondary Education, Performance


