NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 14 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Wainer, Howard; Bradlow, Eric; Wang, Xiaohui – Journal of Educational and Behavioral Statistics, 2010
Confucius pointed out that the first step toward wisdom is calling things by the right name. The term "Differential Item Functioning" (DIF) did not arise fully formed from the miasma of psychometrics, it evolved from a variety of less accurate terms. Among its forebears was "item bias" but that term has a pejorative connotation…
Descriptors: Test Bias, Difficulty Level, Test Items, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Dorans, Neil J. – Harvard Educational Review, 2010
In his 2003 article in the "Harvard Educational Review" (HER), Freedle claimed that the SAT was both culturally and statistically biased and proposed a solution to ameliorate this bias. The author argued (Dorans, 2004a) that these claims were based on serious computational errors. In particular, he focused on how Freedle's table 2 was…
Descriptors: College Entrance Examinations, Test Bias, Test Items, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Santelices, Maria Veronica; Wilson, Mark – Harvard Educational Review, 2010
In their paper "Unfair Treatment? The Case of Freedle, the SAT, and the Standardization Approach to Differential Item Functioning" (Santelices & Wilson, 2010), the authors studied claims of differential effects of the SAT on Latinos and African Americans through the methodology of differential item functioning (DIF). Previous…
Descriptors: College Entrance Examinations, Test Bias, Test Items, Difficulty Level
Dorans, Neil J. – Educational Testing Service, 2010
Santelices and Wilson (2010) claimed to have addressed technical criticisms of Freedle (2003) presented in Dorans (2004a) and elsewhere. Santelices and Wilson's abstract claimed that their study confirmed that SAT[R] verbal items do function differently for African American and White subgroups. In this commentary, I demonstrate that the…
Descriptors: College Entrance Examinations, Verbal Tests, Test Bias, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Freedle, Roy O. – Harvard Educational Review, 2010
In this commentary, the author discusses two recent replications (Santelices & Wilson, 2010; Scherbaum & Goldstein, 2008) of some of his earlier work on SAT items using the differential item functioning (DIF) statistic wherein he contrasted the test performance of African American examinees with White examinees (Freedle, 2003). In this…
Descriptors: College Entrance Examinations, Test Bias, Test Items, Difficulty Level
Wainer, Howard – 1982
This paper is the transcript of a talk given to those who use test information but who have little technical background in test theory. The concepts of modern test theory are compared with traditional test theory, as well as a probable future test theory. The explanations given are couched within an extended metaphor that allows a full description…
Descriptors: Difficulty Level, Latent Trait Theory, Metaphors, Test Items
Thorndike, Robert L. – 1983
In educational testing, one is concerned to get as much information as possible about a given examinee from each minute of testing time. Maximum information is obtained when the difficulty of each test exercise matches the estimated ability level of the examinee. The goal of adaptive testing is to accomplish this. Adaptive patterns are reviewed…
Descriptors: Adaptive Testing, Computer Assisted Testing, Difficulty Level, Latent Trait Theory
Clariana, Roy B. – 1990
Research has shown that multiple-choice questions formed by transforming or paraphrasing a reading passage provide a measure of student comprehension. It is argued that similar transformation and paraphrasing of lesson questions is an appropriate way to form parallel multiple-choice items to be used as a posttest measure of student comprehension.…
Descriptors: Comprehension, Computer Assisted Testing, Difficulty Level, Measurement Techniques
Ebel, Robert L. – 1981
An alternate-choice test item is a simple declarative sentence, one portion of which is given with two different wordings. For example, "Foundations like Ford and Carnegie tend to be (1) eager (2) hesitant to support innovative solutions to educational problems." The examinee's task is to choose the alternative that makes the sentence…
Descriptors: Comparative Testing, Difficulty Level, Guessing (Tests), Multiple Choice Tests
Peer reviewed Peer reviewed
Smetherham, D. – British Journal of Educational Studies, 1979
The item bank (Rasch model) has serious limitations for the national assessment of subject performance: it requires agreement about the scope and organization of each school subject and about the difficulty level of each test item; and items tend to focus on factual knowledge, ignoring other desirable learning outcomes. (SJL)
Descriptors: Behavioral Objectives, Classification, Difficulty Level, Educational Assessment
Peer reviewed Peer reviewed
Linn, Robert L.; Drasgow, Fritz – Educational Measurement: Issues and Practice, 1987
This article discusses the application of the Golden Rule procedure to items of the Scholastic Aptitude Test. Using item response theory, the analyses indicate that the Golden Rule procedures are ineffective in detecting biased items and may undermine the reliability and validity of tests. (Author/JAZ)
Descriptors: College Entrance Examinations, Difficulty Level, Item Analysis, Latent Trait Theory
Peer reviewed Peer reviewed
Eckstein, Shulamith Graus – Developmental Review, 2000
Extends dynamic model of cognitive growth proposed by van Geert in three directions: (1) added a term to consider exposure to material to be learned; (2) developed method to apply model to cross-sectional studies; and (3) developed procedure to scale cognitive abilities tests with items of varying difficulty. Tests model with 2- to 15-year-olds'…
Descriptors: Adolescents, Children, Cognitive Ability, Cognitive Development
Thomas, Gregory P. – 1986
This paper argues that no single measurement strategy serves all purposes and that applying methods and techniques which allow a variety of data elements to be retrieved and juxtaposed may be an investment in the future. Item response theory, Rasch model, and latent trait theory are all approaches to a single conceptual topic. An abbreviated look…
Descriptors: Achievement Tests, Adaptive Testing, Criterion Referenced Tests, Data Collection
Bayless, David L.; Nix, Charles W. – 1979
The merits and hazards of minimum competency testing for the individual student or for student groups are discussed. Types of groups which lend themselves to group application and some important factors in determining the parameters of a group are discussed. Ten critical issues related to minimum competency testing are identified: (1) scope of…
Descriptors: Academic Standards, Cutting Scores, Difficulty Level, Elementary Secondary Education