NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 6,076 to 6,090 of 9,554 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Huang, Yi-Min; Trevisan, Mike; Storfer, Andrew – International Journal for the Scholarship of Teaching and Learning, 2007
Despite the prevalence of multiple choice items in educational testing, there is a dearth of empirical evidence for multiple choice item writing rules. The purpose of this study was to expand the base of empirical evidence by examining the use of the "all-of-the-above" option in a multiple choice examination in order to assess how…
Descriptors: Multiple Choice Tests, Educational Testing, Ability Grouping, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Timms, Michael; Schneider, Steven; Lee, Cindy; Rolfhus, Eric – Regional Educational Laboratory Southwest (NJ1), 2007
This policy research document is intended for Louisiana policymakers to use when examining possible changes to the state assessment's alignment with the National Assessment of Educational Progress (NAEP). The 2009 NAEP test is not yet in existence, so the purpose of this report is to give policymakers a head start in determining where they might,…
Descriptors: Federal Legislation, Test Items, Testing, Science Tests
Lecointe, Darius A. – 1995
The purpose of this Item Response Theory study was to investigate how the expected reduction in item information, due to the collapsing of response categories in performance assessment data, was affected by varying testing conditions: item difficulty, item discrimination, inter-rater reliability, and direction of collapsing. The investigation used…
Descriptors: Classification, Computer Simulation, Difficulty Level, Interrater Reliability
PDF pending restoration PDF pending restoration
Kirisci, Levent; Hsu, Tse-Chi – 1995
The main goal of this study was to assess how sensitive unidimensional parameter estimates derived from BILOG were when the unidimensionality assumption was violated and the underlying ability distribution was not multivariate normal. A multidimensional three-parameter logistic distribution that was a straightforward generalization of the…
Descriptors: Ability, Comparative Analysis, Correlation, Difficulty Level
Dorans, Neil J.; Schmitt, Alicia P. – 1991
Differential item functioning (DIF) assessment attempts to identify items or item types for which subpopulations of examinees exhibit performance differentials that are not consistent with the performance differentials typically seen for those subpopulations on collections of items that purport to measure a common construct. DIF assessment…
Descriptors: Computer Assisted Testing, Constructed Response, Educational Assessment, Item Bias
Boldt, R. F. – 1992
The Test of Spoken English (TSE) is an internationally administered instrument for assessing nonnative speakers' proficiency in speaking English. The research foundation of the TSE examination described in its manual refers to two sources of variation other than the achievement being measured: interrater reliability and internal consistency.…
Descriptors: Adults, Analysis of Variance, Interrater Reliability, Language Proficiency
DeMauro, Gerald E. – 1990
Three papers describe the three stages of developing the National Teacher Examination (NTE) School Psychologist Specialty Area Test. The first stage is described in the paper entitled "Knowledge Areas Important to School Psychology." A survey of the membership of the National Association of School Psychologists helped determine knowledge…
Descriptors: Certification, Elementary Secondary Education, Job Analysis, Job Skills
Pommerich, Mary; And Others – 1995
The Mantel-Haenszel (MH) statistic for identifying differential item functioning (DIF) commonly conditions on the observed test score as a surrogate for conditioning on latent ability. When the comparison group distributions are not completely overlapping (i.e., are incongruent), the observed score represents different levels of latent ability…
Descriptors: Ability, Comparative Analysis, Difficulty Level, Item Bias
Sireci, Stephen G. – 1995
The purpose of this paper is to clarify the seemingly discrepant views of test theorists and test developers about terminology related to the evaluation of test content. The origin and evolution of the concept of content validity are traced, and the concept is reformulated in a way that emphasizes the notion that content domain definition,…
Descriptors: Construct Validity, Content Validity, Definitions, Item Analysis
Messick, Samuel – 1992
Authentic and direct assessments of performances and products are conceptualized in terms of multiple distinctions having implications for validation. These include contrasts between performances and products, between assessment of performance per se and performance assessment of competence or other constructs, between structured and unstructured…
Descriptors: Cognitive Processes, Competence, Educational Assessment, Evaluation Methods
Wainer, Howard; And Others – 1993
The relationship between the multiple-choice and free-response sections of the Computer Science and Chemistry tests of the College Board's Advanced Placement program was studied. Confirmatory factor analysis showed that the free-response sections measure the same underlying proficiency as the multiple-choice sections for the most part. However,…
Descriptors: Advanced Placement, Chemistry, Computer Science, High School Students
Bode, Rita K. – 1995
This study describes the creation of measures of teachers' use of ability grouping in instruction using Rasch analysis. The dimensionality of the proposed construct was also investigated. Results of the Rasch analysis are compared to the results using composites to illustrate how the description of a construct can vary depending on the method used…
Descriptors: Ability Grouping, Classification, Educational Practices, Item Response Theory
Stuck, Ivan – 1995
By focusing on "appropriateness" and "adequacy" of inference and action, unified validity may be misused in rejecting valid test outcomes. The notion of levels of validity is challenged, the necessity of assumption is argued, and experience is proposed as the basis of validity. "Consequential validity" is interpreted as an optional predictive…
Descriptors: Evaluation Methods, Measurement Techniques, Measures (Individuals), Predictive Validity
van der Linden, Wim J. – 1995
Dichotomous item response theory (IRT) models can be viewed as families of stochastically ordered distributions of responses to test items. This paper explores several properties of such distributions. The focus is on the conditions under which stochastic order in families of conditional distributions is transferred to their inverse distributions,…
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Foreign Countries
Schnipke, Deborah L. – 1996
When running out of time on a multiple-choice test, some examinees are likely to respond rapidly to the remaining unanswered items in an attempt to get some items right by chance. Because these responses will tend to be incorrect, the presence of "rapid-guessing behavior" could cause these items to appear to be more difficult than they…
Descriptors: Difficulty Level, Estimation (Mathematics), Guessing (Tests), Item Response Theory
Pages: 1  |  ...  |  402  |  403  |  404  |  405  |  406  |  407  |  408  |  409  |  410  |  ...  |  637