NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 5,041 to 5,055 of 9,552 results Save | Export
Lee, Guemin; Kolen, Michael J.; Frisbie, David A.; Ankenmann, Robert D. – 1998
Item response models can be applied in many test equating situations by making strong statistical assumptions. Thus, studying the robustness of the models to violations of the assumptions and investigating model-data fit are essential in all item response theory (IRT) equating applications (M. Kolen and R. Brennan, 1995). Previous studies dealing…
Descriptors: Equated Scores, Item Response Theory, Robustness (Statistics), Tables (Data)
Cohen, Allan S.; Kim, Seock-Ho; Wollack, James A. – 1998
This paper provides a review of procedures for detection of differential item functioning (DIF) for item response theory (IRT) and observed score methods for the graded response model. In addition, data from a test anxiety scale were analyzed to examine the congruence among these procedures. Data from Nasser, Takahashi, and Benson (1997) were…
Descriptors: Identification, Item Bias, Item Response Theory, Scores
Spray, Judith A.; And Others – 1990
Test data generated according to two different multidimensional item response theory (IRT) models were compared at both the item response level and the test score level to determine whether measurable differences between the models could be detected when the data sets were constrained to be equivalent in terms of item "p"-values. The…
Descriptors: Ability, Comparative Analysis, Item Response Theory, Mathematical Models
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zhang, Jinming; Chang, Hua-Hua – ETS Research Report Series, 2005
This paper compares the use of multiple pools versus a single pool with respect to test security against large-scale item sharing among some examinees in a computer-based test, under the assumption that a randomized item selection method is used. It characterizes the conditions under which employing multiple pools is better than using a single…
Descriptors: Comparative Analysis, Test Items, Item Banks, Computer Assisted Testing
Baker, Eva; Polin, Linda – 1978
The validity studies planned for the Test Design activities deal primarily with the appropriateness of items generated for a domain. Previous exploratory work in the field related to overall test content appropriateness ratings has not been satisfactory. Studies which are solely based on correlational data suffer from confounding with…
Descriptors: Questionnaires, Rating Scales, Test Construction, Test Format
Peer reviewed Peer reviewed
Berk, Ronald A. – Educational and Psychological Measurement, 1978
Three formulae developed to correct item-total correlations for spuriousness were evaluated. Relationships among corrected, uncorrected, and item-remainder correlations were determined by computing sets of mean, minimum, and maximum deviation coefficients and Spearman rank correlations for nine test lengths. (Author/JKS)
Descriptors: Correlation, Intermediate Grades, Item Analysis, Test Construction
Peer reviewed Peer reviewed
Alexander, John J., Ed. – Journal of Chemical Education, 1978
Two exam questions are presented. One suitable for advanced undergraduate or beginning graduate courses in organic chemistry, is on equivalent expressions for the description of several pericyclic reactions. The second, for general chemistry students, asks for an estimation of the rate of decay of a million-year-old Uranium-238 sample. (BB)
Descriptors: Chemistry, Evaluation, Higher Education, Problem Sets
Peer reviewed Peer reviewed
Huck, Schuyler W. – Educational and Psychological Measurement, 1978
A modification of Hoyt's analysis of variance model for test analysis was proposed by Lu. A difficulty that may be encountered in using Lu's modification is examined, and a solution is proposed. (JKS)
Descriptors: Analysis of Variance, Difficulty Level, Item Analysis, Test Items
Peer reviewed Peer reviewed
Vegelius, Jan – Educational and Psychological Measurement, 1977
Generalizations of the G index as a measure of similarity between persons beyond the dichotomous situation are discussed. An attempt is made to present a generalization that does not require dichotomization of the items for cases where the number of response alternatives may differ. (Author/JKS)
Descriptors: Correlation, Item Analysis, Measurement Techniques, Multidimensional Scaling
Peer reviewed Peer reviewed
Dorans, Neil J.; Kulick, Edward – Journal of Educational Measurement, 1986
The standardization method for assessing unexpected differential item performance or differential item functioning is introduced. Findings of five studies are summarized, in which the statistical method of standardization is used to look for unexpected differences in item performance across different subpopulations of the Scholastic Aptitude Test.…
Descriptors: Groups, Item Analysis, Sociometric Techniques, Standardized Tests
Peer reviewed Peer reviewed
Wilcox, Rand R. – Educational and Psychological Measurement, 1983
This article provides unbiased estimates of the proportion of items in an item domain that an examinee would answer correctly if every item were attempted, when a closed sequential testing procedure is used. (Author)
Descriptors: Estimation (Mathematics), Psychometrics, Scores, Sequential Approach
Peer reviewed Peer reviewed
Devito, Anthony J.; Kubis, Joseph F. – Educational and Psychological Measurement, 1983
Alternate forms of the state anxiety (A-State) and trait anxiety (A-Trait) scales of the State-Trait Anxiety Inventory (STAI) were constructed by dividing the 20 items of each scale into two briefer forms having 10 items each. The alternate forms and item statistics are presented. (Author/BW)
Descriptors: Anxiety, Higher Education, Item Analysis, Personality Measures
Dimitrov, Dimiter M. – 2003
This paper provides analytic evaluations of expected (marginal) true-score measures for binary items given their item response theory (IRT) calibration. Under the assumption of normal trait distributions, marginalized true scores, error variance, true score variance, and reliability for norm-referenced and criterion-references interpretations are…
Descriptors: Item Response Theory, Reliability, Test Construction, Test Items
Bunch, Michael B. – 2002
This module explains test review as it is now performed in most large-scale testing programs. It addresses the fundamental aspects of item review principally for fairness, sensitivity, and bias, and to a lesser degree for content and construct validity. The module is designed for a survey course and thus has a broad, rather than deep, focus. The…
Descriptors: Large Scale Assessment, Review (Reexamination), Test Construction, Test Items
Veldkamp, Bernard P.; van der Linden, Wim J.; Ariel, Adelaide – 2002
This paper presents an approach to item pool design that has the potential to improve on the quality of current item pools in educational and psychological testing and thus to increase both measurement precision and validity. The approach consists of the application of mathematical programming techniques to calculate optimal blueprints for item…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Test Construction
Pages: 1  |  ...  |  333  |  334  |  335  |  336  |  337  |  338  |  339  |  340  |  341  |  ...  |  637