NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 6,091 to 6,105 of 9,552 results Save | Export
Taube, Kurt T.; Newman, Larry S. – 1996
A method of estimating Rasch-model difficulty calibrations from judges' ratings of item difficulty is described. The ability of judges to estimate item difficulty was assessed by correlating estimated and empirical calibrations on each of four examinations offered by the American Association of State Social Work Boards. Thirteen members of the…
Descriptors: Correlation, Cutting Scores, Difficulty Level, Estimation (Mathematics)
Johanson, George A.; Johanson, Susan N. – 1996
Differential item functioning (DIF), or item bias, occurs when individuals in a focal group respond differently to a test item than do individuals in a reference group even when comparisons are restricted to individuals with similar overall skill levels on the trait in question. It is common in constructing a questionnaire or survey to recommend…
Descriptors: Achievement Tests, Data Analysis, Evaluation Methods, Item Analysis
Alberta Dept. of Education, Edmonton. Language Services Branch. – 1995
The French as a Second Language model tests for advanced levels 7, 8, and 9 were designed to evaluate students' language performance, as outlined in the program of studies for Alberta, Canada, in listening and reading comprehension and oral and written production, communication skills, culture, language and general language knowledge. The tests…
Descriptors: Advanced Courses, Foreign Countries, French, Language Tests
Yang, Wen-Ling – 1997
Using an anchor-item design of test equating, the effects of three equating methods (Tucker linear and two three-parameter item-response-theory-based (3PL-IRT) methods), and the content representativeness of anchor items on the accuracy of equating were examined; and an innovative way of evaluating equating accuracy appropriate for the particular…
Descriptors: Equated Scores, Item Response Theory, Raw Scores, Test Construction
PDF pending restoration PDF pending restoration
Thompson, Bruce; And Others – 1997
This study was conducted to investigate the construct validity of scores on the Personal Preferences Self-Description Questionnaire (PPSDQ), a measure of Jungian types. Confirmatory factor analysis methods were used to investigate the structures underlying PPSDQ responses of 641 university students. The model fit statistics were generally…
Descriptors: College Students, Construct Validity, Goodness of Fit, Higher Education
Mathieu, Cindy K. – 1997
This paper presents six steps in test construction generally recommended by measurement textbook authors. The focus is primarily on paper-and-pencil achievement tests as used by class instructions, although the discussion touches on the construction of other types of assessment. The six steps are: (1) determine the test purpose; (2) determine the…
Descriptors: Achievement Tests, Difficulty Level, Measurement Techniques, Selection
Go, Imelda C.; Woodruff, David J. – 1996
In previous works, D. J. Woodruff derived expressions for three different conditional test score variances: (1) the conditional standard error of prediction (CSEP); (2) the conditional standard error of measurement in prediction (CSEMP); and (3) the conditional standard error of estimation (CSEE). He also presented step-up formulas that require…
Descriptors: College Entrance Examinations, Error of Measurement, Estimation (Mathematics), High School Students
Taylor, Catherine S. – 1996
This study investigated the impact of task directions on the mathematical performance of high school students from six classes. Students analyzed data regarding school dropout by answering six short-answer questions and writing a letter discussing the trends and their predictions about school dropout. Tasks were scored using two methods: (1) trait…
Descriptors: Dropouts, High School Students, High Schools, Mathematics Tests
Rudner, Lawrence M.; And Others – 1995
Fit statistics provide a direct measure of assessment accuracy by analyzing the fit of measurement models to an individual's (or group's) response pattern. Students that lose interest during the assessment, for example, will miss exercises that are within their abilities. Such students will respond correctly to some more difficult items and…
Descriptors: Difficulty Level, Educational Assessment, Goodness of Fit, Measurement Techniques
Bergstrom, Betty A.; Stahl, John A. – 1992
This paper reports a method for assessing the adequacy of existing item banks for computer adaptive testing. The method takes into account content specifications, test length, and stopping rules, and can be used to determine if an existing item bank is adequate to administer a computer adaptive test efficiently across differing levels of examinee…
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Evaluation Methods
Tang, Huixing – 1994
A method is presented for the simultaneous analysis of differential item functioning (DIF) in multi-factor situations. The method is unique in that it combines item response theory (IRT) and analysis of variance (ANOVA), takes a simultaneous approach to multifactor DIF analysis, and is capable of capturing interaction and controlling for possible…
Descriptors: Ability, Analysis of Variance, Difficulty Level, Error of Measurement
Kwak, Nohoon; Davenport, Ernest C., Jr.; Davison, Mark L. – 1998
The purposes of this study were to introduce the iterative purification procedure and to compare this with the two-step purification procedure, to compare false positive error rates and the power of five observed score approaches and to identify factors affecting power and false positive rates in each method. This study used 2,400 data sets that…
Descriptors: Ability, Comparative Analysis, Error of Measurement, Estimation (Mathematics)
Parshall, Cynthia G.; Kromrey, Jeffrey D.; Chason, Walter M.; Yi, Qing – 1997
Accuracy of item parameter estimates is a critical concern for any application of item response theory (IRT). However, the necessary sample sizes are often difficult to obtain in practice, particularly for the more complex models. A promising avenue of research concerns modified item response models. This study both replicates and improves on an…
Descriptors: Ability, Error of Measurement, Estimation (Mathematics), Item Response Theory
Meijer, Rob R.; van Krimpen-Stoop, Edith M. L. A. – 1998
Several person-fit statistics have been proposed to detect item score patterns that do not fit an item response theory model. To classify response patterns as not fitting a model, a distribution of a person-fit statistic is needed. The null distributions of several fit statistics have been investigated using conventionally administered tests, but…
Descriptors: Ability, Adaptive Testing, Foreign Countries, Item Response Theory
Schiel, Jeff – 1998
This report examines both the substantiveness and statistical significance of differences between mean American College Testing (ACT) Program assessment scale scores. The first part of the report describes the development and results of a method for interpreting the substantiveness of mean differences in terms of test items correctly answered. For…
Descriptors: College Entrance Examinations, High School Students, High Schools, Scores
Pages: 1  |  ...  |  403  |  404  |  405  |  406  |  407  |  408  |  409  |  410  |  411  |  ...  |  637