NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 6,316 to 6,330 of 9,530 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Dempster, Edith R.; Reddy, Vijay – Science Education, 2007
This study investigated the relationship between readability of 73 text-only multiple-choice questions from Trends in International Mathematics and Science Study (TIMSS) 2003 and performance of two groups of South African learners: those with limited English-language proficiency (learners attending African schools) and those with better…
Descriptors: Instructional Effectiveness, Foreign Countries, Disadvantaged Youth, Sentences
Chyn, Susan; And Others – 1995
The current study, carried out jointly by Test Development and Statistical Analysis staff at Educational Testing Service investigated the feasibility of the Automated Item Selection (AIS) procedure for the Test of English as a Foreign Language (TOEFL). Item-response theory (IRT)-based statistical specifications were developed. Two TOEFL test forms…
Descriptors: English (Second Language), Item Banks, Item Response Theory, Language Tests
Mislevy, Robert J.; Wilson, Mark – 1992
Standard item response theory (IRT) models posit latent variables to account for regularities in students' performance on test items. They can accommodate learning only if the expected changes in performance are smooth, and, in an appropriate metric, uniform over items. Wilson's "Saltus" model extends the ideas of IRT to development that…
Descriptors: Bayesian Statistics, Change, Development, Item Response Theory
Stocking, Martha L. – 1994
As adaptive testing moves toward operational implementation in large scale testing programs, where it is important that adaptive tests be as parallel as possible to existing linear tests, a number of practical issues arise. This paper concerns three such issues. First, optimum item pool size is difficult to determine in advance of pool…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Standards
Pomplun, Mark; And Others – 1992
This study evaluated the use of bivariate matching as a solution to the problem of studying differential item functioning (DIF) with formula scored tests. Using Scholastic Aptitude Test verbal data with large samples, both male/female and black/white group comparisons were investigated. Mantel-Haenszel (MH) delta-(D) DIF values and DIF category…
Descriptors: Blacks, Criteria, Females, Item Bias
van den Bergh, Huub; And Others – 1995
The term differential item functioning (DIF) refers to whether or not the same psychological constructs are measured across different groups. If an item does not measure the same skills or subskills in different populations, it is said to function differentially or to display item bias. A multilevel approach to DIF is proposed. In such a model,…
Descriptors: Cluster Analysis, Estimation (Mathematics), Identification, Item Bias
Hicks, Marilyn M. – 1988
Several exploratory analyses of the fifths data generated by Test of English as a Foreign Language (TOEFL) item analyses were developed in order to evaluate the effects of options on the discriminability of difficult items and to identify difficult items with low, unreliable biserials that had been rejected by test developers, but for which…
Descriptors: Difficulty Level, Estimation (Mathematics), Identification, Item Analysis
Yang, Wen-Ling; Houang, Richard T. – 1996
The influence of anchor length on the accuracy of test equating was studied using Tucker's linear method and two Item-Response-Theory (IRT) based methods, focusing on whether equating accuracy improved with more anchor items, whether the anchor effect depended on the equating method used, and the adequacy of the inclusion of the guessing parameter…
Descriptors: Equated Scores, Estimation (Mathematics), Guessing (Tests), Item Response Theory
Frey, Sharon L. – 1996
The Mantel-Haenszel procedure (N. Mantel and W. Haenszel, 1959) and its extension to constructed response items, the Generalized Mantel Haenszel (A. Agresti, 1990), compare performance of subgroups across different score groups to determine differential item functioning (DIF). At each level of comparison, or score group, the subgroups are…
Descriptors: Ability, Comparative Analysis, Constructed Response, Ethnic Groups
Johanson, George A.; Doston, Glenn – 1994
Analyses of questionnaire data from a program evaluation indicate that the two dichotomous items "Would you recommend this to a friend?" and "Would you choose to do this again?" are not as interchangeable as might be expected from the survey literature. As part of the evaluation of a university program, a survey of graduates…
Descriptors: College Graduates, Data Analysis, Graduate Surveys, Higher Education
Plake, Barbara S.; Impara, James C. – 1996
This study investigated the intrajudge consistency of Angoff-based item performance estimates. The examination used was a certification examination in an emergency medicine specialty. Ten expert panelists rated the same 24 items twice during an operational standard setting study. Results indicate that the panelists were highly consistent, in terms…
Descriptors: Cutting Scores, Interrater Reliability, Licensing Examinations (Professions), Performance Based Assessment
McKinley, Robert – 1989
A confirmatory approach to assessing test structure using multidimensional item response theory (MIRT) was developed and evaluated. The approach involved adding to the exponent of the MIRT model an item structure matrix that allows the user to specify the ability dimensions measured by an item. Various combinations of item structures were fit to…
Descriptors: Ability, Chi Square, Goodness of Fit, Item Response Theory
Roberts, James S.; Laughlin, James E. – 1996
Binary or graded disagree-agree responses to attitude items are often collected for the purpose of attitude measurement. Although such data are sometimes analyzed with cumulative measurement models, recent investigations suggest that unfolding models are more appropriate (J. S. Roberts, 1995; W. H. Van Schuur and H. A. L. Kiers, 1994). Advances in…
Descriptors: Attitude Measures, Estimation (Mathematics), Item Response Theory, Mathematical Models
Wainer, Howard; And Others – 1991
It is sometimes sensible to think of the fundamental unit of test construction as being larger than an individual item. This unit, dubbed the testlet, must pass muster in the same way that items do. One criterion of a good item is the absence of differential item functioning (DIF). The item must function in the same way as all important…
Descriptors: Definitions, Identification, Item Bias, Item Response Theory
Crehan, Kevin D.; And Others – 1993
Among the measurement techniques receiving greater attention is the context-dependent item set or testlet. The context-dependent item set consists of a scenario and related test questions. This item format is generally believed to be able to tap higher level thinking. Unfortunately, this item form leads to inter-item dependence within item sets…
Descriptors: Comparative Analysis, Item Response Theory, Measurement Techniques, Reading Tests
Pages: 1  |  ...  |  418  |  419  |  420  |  421  |  422  |  423  |  424  |  425  |  426  |  ...  |  636