Publication Date
| In 2026 | 0 |
| Since 2025 | 215 |
| Since 2022 (last 5 years) | 1084 |
| Since 2017 (last 10 years) | 2594 |
| Since 2007 (last 20 years) | 4955 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedWright, Benjamin D.; Douglas, Graham A. – Applied Psychological Measurement, 1977
A procedure for obtaining Rasch model estimates of item difficulty and of ability is detailed. The procedure approximates the optimal but difficult to obtain "unconditional" estimates. (JKS)
Descriptors: Item Analysis, Latent Trait Theory, Mathematical Models, Measurement
Peer reviewedClaudy, John G. – Applied Psychological Measurement, 1978
Option weighting is an alternative to increasing test length as a means of improving the reliability of a test. The effects on test reliability of option weighting procedures were compared in two empirical studies using four independent sets of items. Biserial weights were found to be superior. (Author/CTM)
Descriptors: Higher Education, Item Analysis, Scoring Formulas, Test Items
Smith, Charles W. – Illinois Schools Journal, 1987
Provides recommendations to aid teachers (elementary and secondary) in planning tests, constructing item types (essay, short answer, true-false, or other alternate response, matching, multiple-choice), administering tests, scoring tests, analyzing tests, and using test results. (KH)
Descriptors: Elementary Secondary Education, Teacher Made Tests, Test Construction, Test Format
Peer reviewedLong, H. B. – International Journal of Lifelong Education, 1987
Data were collected from 117 college students to examine the validity of the Self-Directed Learning Readiness Scale. Two assumptions were generated from the results: (1) most items on the scale are correlated with the subject's total score and (2) most items are not correlated with the subject's age. (CH)
Descriptors: Adult Education, Learning Readiness, Personal Autonomy, Postsecondary Education
Peer reviewedSecolsky, Charles – Journal of Educational Measurement, 1987
For measuring the face validity of a test, Nevo suggested that test takers and nonprofessional users rate items on a five point scale. This article questions the ability of those raters and the credibility of the aggregated judgment as evidence of the validity of the test. (JAZ)
Descriptors: Content Validity, Measurement Techniques, Rating Scales, Test Items
Peer reviewedWoodruff, David – Journal of Educational Statistics, 1986
The purpose of the present paper is to derive linear equating methods for the common item nonequivalent populations design from explicitly stated congeneric type test score models. The equating methods developed are compared with previously developed methods and applied to five professionally constructed examinations administered to approximately…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Scores
Peer reviewedEmbretson (Whitely), Susan – Psychometrika, 1984
The purpose of this paper is to propose a general multicomponent latent trait model (LTM) for response processes. It combines the linear logistic LTM with the multicomponent LTM. Joint maximum likelihood estimators are presented for parameters of the general multicomponent LTM and an application to cognitive test items is described. (Author/BW)
Descriptors: Cognitive Tests, Estimation (Mathematics), Latent Trait Theory, Mathematical Models
Peer reviewedHoyt, Kenneth B. – Journal of Counseling & Development, 1986
The microcomputer version of the Ohio Vocational Interest Survey (OVIS II) differs from the machine-scored version in its ability to incorporate data from the OVIS II:Career Planner in its printed report. It differs from the hand-scored version in its ability to include data from the OVIS II:Work Characteristic Analysis in its printed report.…
Descriptors: Comparative Analysis, Computer Assisted Testing, Microcomputers, Test Format
Peer reviewedCross, David – British Journal of Language Teaching, 1984
Suggests that the English-as-a-foreign-language field would benefit by using the same types of tests as those of "Graded Objectives in Modern Languages," which are used in British schools to test foreign language learning. Presents test items which are typical of the British tests. (SED)
Descriptors: Achievement Tests, English (Second Language), Language Tests, Second Language Learning
Peer reviewedOosterhof, Albert C.; And Others – Educational and Psychological Measurement, 1984
A supplementary treatment is proposed which helps identify sources of bias affecting groups of test items. This treatment is illustrated with the tranformed item-difficulty method as applied to an evaluation of a test used to help select applicants to be admitted to an aviation training program. (Author/BW)
Descriptors: Adults, Aptitude Tests, Difficulty Level, Graphs
Mizokawa, Donald T.; Hamlin, Michael D. – Educational Technology, 1984
Suggestions for software design in computer managed testing (CMT) cover instructions to testees, their physical format, provision of practice items, and time limit information; test item presentation, physical format, discussion of task demands, review capabilities, and rate of presentation; pedagogically helpful utilities; typefonts; vocabulary;…
Descriptors: Computer Assisted Testing, Decision Making, Guidelines, Test Construction
Peer reviewedSharpley, Christopher F.; Rogers, H. Jane – Journal of Clinical Psychology, 1985
Compared items from psychologically naive vs. psychologically sophisticated item-writers vs. a standardized test (N=552). Results showed that nonpsychologists with no formal definition of the construct they were to measure were able to write items that were as valid as those elicited from psychologists. (BH)
Descriptors: Anxiety, Foreign Countries, Lay People, Measurement Techniques
Peer reviewedVan Der Flier, Henk; And Others – Journal of Educational Measurement, 1984
Two strategies for assessing item bias are discussed: methods comparing item difficulties unconditional on ability and methods comparing probabilities of response conditional on ability. Results suggest that the iterative logit method is an improvement on the noniterative one and is efficient in detecting biased and unbiased items. (Author/DWH)
Descriptors: Algorithms, Evaluation Methods, Item Analysis, Scores
Peer reviewedConger, Anthony J. – Educational and Psychological Measurement, 1983
A paradoxical phenomenon of decreases in reliability as the number of elements averaged over increases is shown to be possible in multifacet reliability procedures (intraclass correlations or generalizability coefficients). Conditions governing this phenomenon are presented along with implications and cautions. (Author)
Descriptors: Generalizability Theory, Test Construction, Test Items, Test Length
Howe, Roger; Scheaffer, Richard; Lindquist, Mary; Philip, Frank; Halbrook, Arthur – US Department of Education, 2004
This document contains the framework and a set of recommendations for the 2005 NAEP mathematics assessment. It includes descriptions of the mathematical content of the test, the types of test questions, and recommendations for administration of the test. In broad terms, this framework attempts to answer the question: What mathematics should be…
Descriptors: National Competency Tests, Student Evaluation, Mathematics Achievement, Test Items


