Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 2 |
Descriptor
Item Analysis | 20 |
Statistical Analysis | 20 |
Testing Problems | 20 |
Test Validity | 9 |
Test Construction | 8 |
Test Bias | 6 |
Test Items | 6 |
Achievement Tests | 5 |
Latent Trait Theory | 5 |
Adaptive Testing | 3 |
Culture Fair Tests | 3 |
More ▼ |
Source
Applied Measurement in… | 1 |
Assessment in Education:… | 1 |
Journal of Economic Education | 1 |
Psychometrika | 1 |
Author
Publication Type
Reports - Research | 16 |
Speeches/Meeting Papers | 5 |
Journal Articles | 3 |
Reports - Descriptive | 1 |
Reports - Evaluative | 1 |
Education Level
Secondary Education | 1 |
Audience
Researchers | 3 |
Location
Netherlands | 1 |
Laws, Policies, & Programs
Emergency School Aid Act 1972 | 1 |
Assessments and Surveys
California Achievement Tests | 1 |
Metropolitan Readiness Tests | 1 |
What Works Clearinghouse Rating
van Rijn, P. W.; Beguin, A. A.; Verstralen, H. H. F. M. – Assessment in Education: Principles, Policy & Practice, 2012
While measurement precision is relatively easy to establish for single tests and assessments, it is much more difficult to determine for decision making with multiple tests on different subjects. This latter is the situation in the system of final examinations for secondary education in the Netherlands and is used as an example in this paper. This…
Descriptors: Secondary Education, Tests, Foreign Countries, Decision Making
Wollack, James A. – Applied Measurement in Education, 2006
Many of the currently available statistical indexes to detect answer copying lack sufficient power at small [alpha] levels or when the amount of copying is relatively small. Furthermore, there is no one index that is uniformly best. Depending on the type or amount of copying, certain indexes are better than others. The purpose of this article was…
Descriptors: Statistical Analysis, Item Analysis, Test Length, Sample Size
Sympson, James B. – 1976
Latent trait test score theory is discussed primarily in terms of Birnbaum's three-parameter logistic model, and with some reference to the Rasch model. Equations and graphic illustrations are given for item characteristic curves and item information curves. An example is given for a hypothetical 20-item adaptive test, showing cumulative results…
Descriptors: Adaptive Testing, Bayesian Statistics, Item Analysis, Latent Trait Theory
Frederiksen, Norman – 1976
A number of different ways of ascertaining whether or not a test measures the same thing in different cultures are examined. Methods range from some that are obvious and simple to those requiring statistical and psychological sophistication. Simpler methods include such things as having candidates "think aloud" and interviewing them about how they…
Descriptors: Analysis of Covariance, Culture Fair Tests, Factor Analysis, Item Analysis
Doolittle, Allen E. – 1985
Differential item performance (DIP) is discussed as a concept that does not necessarily imply item bias or unfairness to subgroups of examinees. With curriculum-based achievement tests, DIP is presented as a valid reflection of group differences in requisite skills and instruction. Using data from a national testing of the ACT Assessment, this…
Descriptors: Achievement Tests, High Schools, Item Analysis, Mathematics Achievement
Miller, M. David; Burstein, Leigh – 1981
Two studies are presented in this report. The first is titled "Empirical Studies of Multilevel Approaches to Test Development and Interpretation: Measuring Between-Group Differences in Instruction." Because of a belief that schooling does affect student achievement, researchers have questioned the empirical and measurement techniques…
Descriptors: Error Patterns, Evaluation Methods, Item Analysis, Models
Green, Donald Ross – 1976
During the past few years the problem of bias in testing has become an increasingly important issue. In most research, bias refers to the fair use of tests and has thus been defined in terms of an outside criterion measure of the performance being predicted by the test. Recently however, there has been growing interest in assessing bias when such…
Descriptors: Achievement Tests, Item Analysis, Mathematical Models, Minority Groups
Kuntz, Patricia – 1982
The quality of mathematics multiple choice items and their susceptibility to test wiseness were examined. Test wiseness was defined as "a subject's capacity to utilize the characteristics and formats of the test and/or test taking situation to receive a high score." The study used results of the Graduate Record Examinations Aptitude Test (GRE) and…
Descriptors: Cues, Item Analysis, Multiple Choice Tests, Psychometrics
Scheuneman, Janice – 1975
In order to screen out items which may be biased against some ethnic group prior to the final selection of items in test construction, a statistical technique for assessing item bias was developed. Based on a theoretical formulation of R. B. Darlington, the method compares the performance of individuals who belong to different ethnic groups, but…
Descriptors: Achievement Tests, Content Analysis, Cultural Influences, Ethnic Groups
Waller, Michael I. – 1974
In latent trait models the standard procedure for handling the problem caused by guessing on multiple choice tests is to estimate a parameter which is intended to measure the "guessingness" inherent in an item. Birnbaum's three parameter model, which handles guessing in this manner, ignores individual differences in guessing tendency. This paper…
Descriptors: Goodness of Fit, Guessing (Tests), Individual Differences, Item Analysis

Lord, Frederic M. – Psychometrika, 1974
Omitted items cannot properly be treated as wrong when estimating ability and item parameters. A convenient method for utilizing the information provided by omissions is presented. Theoretical and empirical justifications are presented for the estimates obtained by the new method. (Author)
Descriptors: Academic Ability, Guessing (Tests), Item Analysis, Latent Trait Theory

Gohmann, Stephan F.; Spector, Lee C. – Journal of Economic Education, 1989
Compares the effect of content ordering and scrambled ordering on examinations in courses, such as economics, that require quantitative skills. Empirical results suggest that students do no better if they are given a content-ordered rather than a scrambled examination as student performance is not adversely affected by scrambled ordered…
Descriptors: Cheating, Economics Education, Educational Research, Grading
Rodgers, Ron – 1974
The construct of motivation toward school is vaguely defined. The Junior Index of Motivation (JIM Scale) is one of few instruments claiming validity in measuring motivation toward school among junior and semior high students. This study discusses the shortcomings of the JIM scale, and compares item and total score characteristics and correlates…
Descriptors: Adolescents, Correlation, Grade Point Average, Item Analysis
Broussard, Rolland L. – 1985
The cultural bias of the Adult Performance Level Assessment, Form AA-l (APLA) was examined. The potential influence of cultural differences on scores of a major ethnic group, Acadians or Cajuns, was investigated. Assessment items most prone to produce differences in scores were isolated and administered to selected groups. No significant…
Descriptors: Adult Basic Education, Adult Literacy, Culture Fair Tests, Ethnic Groups
Legg, Sue M. – 1982
A case study of the Florida Teacher Certification Examination (FTCE) program was described to assist others launching the development of large scale item banks. FTCE has four subtests: Mathematics, Reading, Writing, and Professional Education. Rasch calibrated item banks have been developed for all subtests except Writing. The methods used to…
Descriptors: Cutting Scores, Difficulty Level, Field Tests, Item Analysis
Previous Page | Next Page ยป
Pages: 1 | 2