Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
van Krimpen-Stoop, Edith M. L. A.; Meijer, Rob R. – 2000
Item scores that do not fit an assumed item response theory model may cause the latent trait value to be estimated inaccurately. For computerized adaptive tests (CAT) with dichotomous items, several person-fit statistics for detecting nonfitting item score patterns have been proposed. Both for paper-and-pencil (P&P) test and CATs, detection of…
Descriptors: Adaptive Testing, Computer Assisted Testing, Goodness of Fit, Item Response Theory
Zhu, Daming; Fan, Meichu – 1999
The convention for selecting starting points (that is, initial items) on a computerized adaptive test (CAT) is to choose as starting points items of medium difficulty for all examinees. Selecting a starting point based on prior information about an individual's ability was first suggested many years ago, but has been believed unimportant provided…
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Difficulty Level
Haladyna, Thomas M. – 1999
This book explains writing effective multiple-choice test items and studying responses to items to evaluate and improve them, two topics that are very important in the development of many cognitive tests. The chapters are: (1) "Providing a Context for Multiple-Choice Testing"; (2) "Constructed-Response and Multiple-Choice Item Formats"; (3)…
Descriptors: Constructed Response, Multiple Choice Tests, Test Construction, Test Format
Wright, Benjamin D.; Stone, Mark H. – 1979
This handbook explains how to do Rasch measurement. The emphasis is on practice, but theoretical explanations are also provided. The Forward contains an introduction to the topic of Rasch measurement. Chapters 2, 4, 5, and 7 use a small problem to illustrate the application of Rasch measurement in detail, and methodological issues are considered…
Descriptors: Item Response Theory, Mathematical Models, Measurement Techniques, Psychometrics
Colorado State Dept. of Education, Denver. – 2000
This document contains samples and explanations from Colorado's state reading and writing comprehension test for Grade 4, hereby released to the public (parents, teachers, and any other interested parties). It presents one reading comprehension item, the scoring guide for that item, and 6 different student responses, with the scores they received.…
Descriptors: Grade 4, Intermediate Grades, Reading Comprehension, Reading Tests
Ito, Kyoko; Sykes, Robert C. – 2000
This study investigated the practice of weighting a type of test item, such as constructed response, more than other types of items, such as selected response, to compute student scores for a mixed-item type of test. The study used data from statewide writing field tests in grades 3, 5, and 8 and considered two contexts, that in which a single…
Descriptors: Constructed Response, Elementary Education, Essay Tests, Test Construction
Peer reviewedNorton, Robert – Journal of Marriage and the Family, 1983
Examines the operationalization of marital quality indices used as dependent variables and the functioning and construction of marital quality variables. Presents both semantic and empirical criteria to judge the development of a marital quality index and a Quality Marriage Index (QMI) based on the introduced criteria. (Author/RC)
Descriptors: Factor Analysis, Marriage, Measures (Individuals), Spouses
Peer reviewedHoste, Roland – Journal of Biological Education, 1982
Factor analysis was used to investigate constructs underlying performance in a biology examination supporting an hypothesis suggesting that the constructs were related to knowledge (rather than other levels of Bloom's taxonomy) of topics. Indicates that the method provides a means of identifying what examination items test. (Author/JN)
Descriptors: Biology, Science Education, Science Tests, Secondary Education
Peer reviewedLoo, Robert – Perceptual and Motor Skills, 1982
The structure of the Group Embedded-Figures Test was examined. Data from 173 females and 93 males were subjected to Johnson's complete-linkage cluster analysis and principal components analysis. The results showed the importance of simple form type, especially Form E, the cube, and of item position. (Author/RD)
Descriptors: Cluster Analysis, Factor Analysis, Factor Structure, Foreign Countries
Peer reviewedMershon, Donald H. – Teaching of Psychology, 1982
Describes a method for increasing efficiency of examination production which uses file cards to store and organize test items. The process of reproducing tests directly from master copies made with file cards is discussed. (AM)
Descriptors: Efficiency, Higher Education, Item Banks, Job Simplification
Peer reviewedWillson, Victor L. – Educational and Psychological Measurement, 1982
The Serlin-Kaiser procedure is used to complete a principal components solution for scoring weights for all options of a given item. Coefficient alpha is maximized for a given multiple choice test. (Author/GK)
Descriptors: Analysis of Covariance, Factor Analysis, Multiple Choice Tests, Scoring Formulas
Peer reviewedFurnham, Adrian; Henderson, Monika – Journal of Clinical Psychology, 1982
Content analyzed four personality inventories, Eysenck Personality Questionnaire, Minnesota Multiphasic Personality Inventory, California Personality Inventory, and Edwards Personality Preference Schedule to determine whether each item in the four inventories contained situational, temporal, or comparative information. Wide differences appeared…
Descriptors: Behavior Patterns, Comparative Analysis, Data Analysis, Item Analysis
Peer reviewedTerwilliger, James S.; Lele, Kaustubh – Journal of Educational Measurement, 1979
Different indices for the internal consistency, reproducibility, or homogeneity of a test are based upon highly similar conceptual frameworks. Illustrations are presented to demonstrate how the maximum and minimum values of KR20 are influenced by test difficulty and the shape of the distribution of test scores. (Author/CTM)
Descriptors: Difficulty Level, Item Analysis, Mathematical Formulas, Statistical Analysis
Peer reviewedHsu, Louis M. – Educational and Psychological Measurement, 1980
Relative difficulty of Separate (Form S) and Grouped (Form G) True-False tests may be expected to be dependent on the ability levels of examinees. At some levels Form S should be less difficult, at others equally difficult, and at still others, more difficult, than Form G. (Author/RL)
Descriptors: Academic Ability, Cluster Grouping, Difficulty Level, Knowledge Level
Peer reviewedMoss, S. C.; Hogg, J. – Research Quarterly for Exercise and Sport, 1981
The variety of hand grips of 12 children, most of whom were moderately or severely retarded, were classified in order to begin an analysis of hand function. Test reliability was not as great when items were presented to the children as compared to when children were observed or rated by videotape. (FG)
Descriptors: Mental Retardation, Motor Development, Object Manipulation, Preschool Children


