Publication Date
| In 2026 | 0 |
| Since 2025 | 5 |
| Since 2022 (last 5 years) | 45 |
| Since 2017 (last 10 years) | 91 |
| Since 2007 (last 20 years) | 144 |
Descriptor
| Test Format | 418 |
| Test Reliability | 418 |
| Test Validity | 243 |
| Test Construction | 135 |
| Test Items | 119 |
| Higher Education | 88 |
| Multiple Choice Tests | 68 |
| Foreign Countries | 67 |
| Testing | 65 |
| Test Interpretation | 61 |
| Comparative Analysis | 57 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 33 |
| Teachers | 23 |
| Administrators | 18 |
| Researchers | 12 |
| Community | 1 |
| Counselors | 1 |
| Policymakers | 1 |
| Students | 1 |
| Support Staff | 1 |
Location
| New York | 9 |
| Turkey | 8 |
| California | 7 |
| Canada | 6 |
| Japan | 6 |
| Germany | 4 |
| United Kingdom | 4 |
| Georgia | 3 |
| Israel | 3 |
| France | 2 |
| Indonesia | 2 |
| More ▼ | |
Laws, Policies, & Programs
| Individuals with Disabilities… | 1 |
| Job Training Partnership Act… | 1 |
| No Child Left Behind Act 2001 | 1 |
| Pell Grant Program | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedKon, Jane Heckley; Martin-Kniep, Giselle O. – Social Education, 1992
Describes a case study to determine whether performance tests are a feasible alternative to multiple-choice tests. Examines the difficulties of administering and scoring performance assessments. Explains that the study employed three performance tests and one multiple-choice test. Concludes that performance test administration and scoring was no…
Descriptors: Educational Objectives, Educational Research, Educational Testing, Geography Instruction
Peer reviewedRafaeli, Sheizaf; Tractinsky, Noam – Computers in Human Behavior, 1991
Discussion of time-related measures in computerized ability tests focuses on a study of college students that used two intelligence test item types to develop a multitrait, multimethod assessment of response time measures. Convergent and discriminant validation are discussed, correlations between response time and accuracy are examined, and…
Descriptors: Computer Assisted Testing, Correlation, Higher Education, Intelligence Tests
Peer reviewedRogers, W. Todd; Harley, Dwight – Educational and Psychological Measurement, 1999
Examined item-level and test-level characteristics for items in a high-stakes school-leaving mathematics examination. Results from 158 students show that the influence of testwiseness is lessened when three-option items are used. Tests of three-option items are at least equivalent to four-option item tests in terms of internal-consistency score…
Descriptors: Comparative Analysis, High School Students, High Schools, High Stakes Tests
Buser, Karen – 1996
Most seasoned test developers recognize the importance of thoughtful decision making when constructing a test. Unfortunately, many classroom achievement tests are created by novice test developed who have not received sufficient instruction in item writing (G. Gulliksen, 1986; R. J. Stiggins, 1991). The result is often a test that is poorly…
Descriptors: Achievement Tests, Decision Making, Educational Planning, Evaluation Methods
Metropolitan Atlanta Consortium of Consultants and Lead Speech-Language Pathologists, GA. – 1990
This guide presents ratings of assessment instruments for use by speech-language pathologists with preschool students. Tests are reviewed in alphabetical order on forms filled out by practicing speech-language pathologists, including data on speech components covered by each test, age range, factors of norms where norms are used, reliability,…
Descriptors: Diagnostic Tests, Examiners, Preschool Education, Preschool Tests
Sewall, Timothy J. – 1986
This paper addresses the issue of whether four of the learning styles instruments currently available are of sufficient psychometric quality to warrant their continued use either for research or educational purposes. Four instruments, which purport to measure learning styles, were selected for review. Criteria for selection were based in part on…
Descriptors: Adult Education, Adult Learning, Cognitive Style, Personality Measures
Oosterhof, Albert C.; Coats, Pamela K. – 1981
Instructors who develop classroom examinations that require students to provide a numerical response to a mathematical problem are often very concerned about the appropriateness of the multiple-choice format. The present study augments previous research relevant to this concern by comparing the difficulty and reliability of multiple-choice and…
Descriptors: Comparative Analysis, Difficulty Level, Grading, Higher Education
Eignor, Daniel R.; Hambleton, Ronald K. – 1979
The purpose of the investigation was to obtain some relationships among (1) test lengths, (2) shape of domain-score distributions, (3) advancement scores, and (4) several criterion-referenced test score reliability and validity indices. The study was conducted using computer simulation methods. The values of variables under study were set to be…
Descriptors: Comparative Analysis, Computer Assisted Testing, Criterion Referenced Tests, Cutting Scores
Peer reviewedKibby, Michael W. – Journal of Reading, 1986
Reviews the 3-R's Test that measures achievement in three academic content areas: reading, language, and mathematics. Cautions that it lacks data supporting its validity and scaling. (HOD)
Descriptors: Academic Achievement, Achievement Tests, Cognitive Development, Elementary Secondary Education
Hopper, Margaret F. – 2001
This paper provides an overview of the types of testing accommodations used for students with disabilities and presents arguments for and against their use. It begins by discussing student participation in educational assessments and federal requirements concerning the participation of students with disabilities. The types of accommodations are…
Descriptors: Academic Accommodations (Disabilities), Academic Standards, Disabilities, Educational Assessment
Peer reviewedAnd Others; Mann, Irene T. – Applied Psychological Measurement, 1979
Several methodological problems (particularly the assumed bipolarity of scales, instructions regarding use of the midpoint, and concept-scale interaction) which may contribute to a lack of precision in the semantic differential technique were investigated. Results generally supported the use of the semantic differential. (Author/JKS)
Descriptors: Analysis of Variance, Computer Assisted Testing, Higher Education, Rating Scales
Peer reviewedHarasym, P. H.; And Others – Evaluation and the Health Professions, 1980
Coded, as opposed to free response items, in a multiple choice physiology test had a cueing effect which raised students' scores, especially for lower achievers. Reliability of coded items was also lower. Item format and scoring method had an effect on test results. (GDC)
Descriptors: Achievement Tests, Comparative Testing, Cues, Higher Education
Peer reviewedPomplun, Mark; Omar, Md Hafidz – Educational and Psychological Measurement, 1997
Four threats to validity of an alternative objective test item format, the multiple-mark format, were studied with data from a state-mandated assessment with about 30,000 students at each of three grade levels. Reliability and validity coefficients show that the format has promise as an objective format that can be aligned with new curriculum…
Descriptors: Curriculum Development, Elementary School Students, Elementary Secondary Education, Objective Tests
Lazarus, Belinda; Killu, Kim – Diagnostique, 1999
This article describes the second edition of the Attention Deficit Disorders Evaluation Scale (ADDES II), an individually administered behavior rating scale developed to assist in the identification and service of children with attention deficit hyperactivity disorders. Its purpose, administration, interpretation of scores, standardization,…
Descriptors: Attention Deficit Disorders, Behavior Rating Scales, Children, Disability Identification
Pollock, Judith M.; And Others – 1997
This report describes an experiment in constructed response testing undertaken in conjunction with the National Education Longitudinal Study of 1988 (NELS:88). Constructed response questions are those that require students to produce their own response rather than selecting the correct answer from several options. Participants in this experiment…
Descriptors: Constructed Response, Costs, High School Students, High Schools


