Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedTorrance, Harry – British Journal of Sociology of Education, 1981
Focuses on mental tests, test developers, and the social and historical context in which mental tests were developed in the United States and England. Suggests that the tests were produced in an intellectual and social climate heavily influenced by eugenicist theory and that more attention should be paid to research which casts doubts on the…
Descriptors: Comparative Analysis, Comparative Education, Elementary Secondary Education, Measurement Objectives
Peer reviewedRindler, Susan Ellerin – Educational and Psychological Measurement, 1980
A short verbal aptitude test was administered under varying time limits with answer sheets specially designed to allow items that had been skipped to be identified. It appeared advantageous for the more able (based on grade point averages) but disadvantageous for the less able to skip items. (Author/RL)
Descriptors: Aptitude Tests, Difficulty Level, Higher Education, Response Style (Tests)
Peer reviewedWilloughby, T. Lee – Educational and Psychological Measurement, 1980
The reliability and validity of a priori estimates of item characteristics are assessed. Results suggest that judges can make a modest contribution to estimation prior to actual administration. (Author/GK)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Medical School Faculty
Peer reviewedKabel, Robert L. – Chemical Engineering Education, 1979
This heat transfer problem requires students to analyze the feasibility of moving icebergs from the Antarctic to Saudi Arabia to provide fresh water. (BB)
Descriptors: Chemistry, College Science, Computation, Engineering Education
Peer reviewedMartin, Joseph J. – Chemical Engineering Education, 1979
This problem requires the student to calculate the volume of air breathed by a bicyclist every hour in Denver, Colorado, and New York City. (BB)
Descriptors: Chemistry, College Science, Computation, Energy
Peer reviewedBlack, Thomas R. – Journal of Research in Science Teaching, 1980
Explores variables influencing the cognitive emphasis of teachers' examinations. Examination questions from Nigerian secondary school science teachers were analyzed according to Bloom's Taxonomy. The influence of the following variables on levels of questions was investigated: teachers' educational backgrounds, subjects taught, grade level taught,…
Descriptors: Difficulty Level, Science Education, Science Tests, Secondary Education
Peer reviewedWilcox, Rand R. – Applied Psychological Measurement, 1980
This paper discusses how certain recent technical advances might be extended to examine proficiency tests which are conceptualized as representing a variety of skills with one or more items per skill. In contrast to previous analyses, errors in the item level are included. (Author/BW)
Descriptors: Mastery Tests, Minimum Competencies, Minimum Competency Testing, Sampling
Peer reviewedHuynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models
Peer reviewedJaeger, Richard M. – Journal of Educational Measurement, 1981
Five indices are discussed that should logically discriminate between situations in which: (1) the linear equating method (LEM) adequately adjusts for difference between score distributions of two approximately parallel test forms; or (2) a method more complex than the linear equating method is needed. (RL)
Descriptors: College Entrance Examinations, Comparative Analysis, Difficulty Level, Equated Scores
Peer reviewedHoste, R. – British Journal of Educational Psychology, 1981
In this paper, a proposal is made by which a content validity coefficient can be calculated. An example of the use of the coefficient is given, demonstrating that different question combinations in a CSE biology examination in which a choice of questions was given gave different levels of content validity. (Author)
Descriptors: Achievement Tests, Biology, Content Analysis, Item Sampling
Peer reviewedWarren, Gordon – Journal of Research in Science Teaching, 1979
The paper documents experiments designed to compare essay and multiple-choice tests as means of testing science learning. Results indicate it is easier to score high on multiple-choice tests; some students substitute quantity on essay tests; and essay tests reveal weaknesses hidden by multiple-choice tests. (RE)
Descriptors: Educational Research, Educational Theories, Evaluation Methods, Instructional Materials
Peer reviewedKnifong, J. Dan – Journal for Research in Mathematics Education, 1980
The computational difficulty of the word problem sections of eight standardized achievement tests was analyzed with respect to the variety of computational procedures and the number of digits per whole number computation. Analysis reveals considerable variation among the current tests in terms of computational procedure and difficulty. (Author/MK)
Descriptors: Computation, Difficulty Level, Educational Research, Elementary Education
Peer reviewedHolliday, William G.; Partridge, Louise A. – Journal of Research in Science Teaching, 1979
Investigates two evaluative hypotheses related to test item sequence and the performance of the students who take the tests. (SA)
Descriptors: Educational Research, Elementary Education, Evaluation, Measurement
Peer reviewedJohanson, George A. – Evaluation Practice, 1997
A discussion of differential item functioning (DIF) in the context of attitude assessment is followed by examples involving the detection of DIF on an attitudes-toward-science scale completed by 1,550 elementary school students and the finding of no DIF in a workshop evaluation completed by 1,682 adults. (SLD)
Descriptors: Adults, Attitude Measures, Attitudes, Elementary Education
Peer reviewedZenisky, April L.; Sireci, Stephen G. – Applied Measurement in Education, 2002
Reviews and illustrates some of the current technological developments in computer-based testing, focusing on novel item formats and automated scoring methodologies. The review shows a number of innovations being researched and implemented. (SLD)
Descriptors: Educational Innovation, Educational Technology, Elementary Secondary Education, Large Scale Assessment


