Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 8 |
Descriptor
Group Testing | 31 |
Item Analysis | 31 |
Test Construction | 9 |
Achievement Tests | 7 |
Statistical Analysis | 7 |
Test Items | 7 |
Test Reliability | 6 |
Test Validity | 6 |
Comparative Analysis | 5 |
Evaluation Methods | 5 |
Intelligence Tests | 5 |
More ▼ |
Source
Author
Guhn, Martin | 2 |
Lewy, Arieh | 2 |
Angoff, William H. | 1 |
Bart, William. | 1 |
Birney, Damian P. | 1 |
Bolt, Sara E. | 1 |
Bowers, John J. | 1 |
Brennan, Robert L. | 1 |
Camilli, Gregory | 1 |
Decker, Dawn M. | 1 |
Denson, Teri A. | 1 |
More ▼ |
Publication Type
Reports - Research | 14 |
Journal Articles | 8 |
Speeches/Meeting Papers | 4 |
Reports - Evaluative | 3 |
Guides - General | 1 |
Guides - Non-Classroom | 1 |
Reports - Descriptive | 1 |
Education Level
Elementary Secondary Education | 3 |
Early Childhood Education | 2 |
Grade 4 | 1 |
Audience
Location
Canada | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Phillips, Gary W. – Applied Measurement in Education, 2015
This article proposes that sampling design effects have potentially huge unrecognized impacts on the results reported by large-scale district and state assessments in the United States. When design effects are unrecognized and unaccounted for they lead to underestimating the sampling error in item and test statistics. Underestimating the sampling…
Descriptors: State Programs, Sampling, Research Design, Error of Measurement
Li, Ying; Jiao, Hong; Lissitz, Robert W. – Journal of Applied Testing Technology, 2012
This study investigated the application of multidimensional item response theory (IRT) models to validate test structure and dimensionality. Multiple content areas or domains within a single subject often exist in large-scale achievement tests. Such areas or domains may cause multidimensionality or local item dependence, which both violate the…
Descriptors: Achievement Tests, Science Tests, Item Response Theory, Measures (Individuals)
Camilli, Gregory – Educational Research and Evaluation, 2013
In the attempt to identify or prevent unfair tests, both quantitative analyses and logical evaluation are often used. For the most part, fairness evaluation is a pragmatic attempt at determining whether procedural or substantive due process has been accorded to either a group of test takers or an individual. In both the individual and comparative…
Descriptors: Alternative Assessment, Test Bias, Test Content, Test Format
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias
Decker, Dawn M.; Bolt, Sara E. – Assessment for Effective Intervention, 2008
The intent of large-scale assessment systems is to promote student achievement toward specific standards by holding schools accountable for the performance of all students. However, it is difficult to know whether large-scale assessment systems are having this intended effect as they are currently implemented. In this article, the authors examine…
Descriptors: Intervention, Academic Achievement, Accountability, Achievement Rating

Menne, John W.; Tolsma, Robert J. – Journal of Educational Measurement, 1971
Descriptors: Discriminant Analysis, Group Testing, Item Analysis, Psychometrics

Garfinkel, Robin; Thorndike, Robert L. – Child Development, 1976
This study was conducted to determine how items of the Stanford-Binet Intelligence Scale, Form L-M, had performed in the 1930's standardization sample in comparison with the 1972 standardization sample. (SB)
Descriptors: Comparative Analysis, Comparative Testing, Group Testing, Intelligence Tests
Guhn, Martin; Janus, Magdalena; Hertzman, Clyde – Early Education and Development, 2007
This invited special issue of "Early Education and Development" presents research related to the Early Development Instrument (EDI; Janus & Offord, 2007), a community tool to assess children's school readiness at a population level. In this editorial introduction, we first sketch out recent trends in school readiness research that call for a…
Descriptors: School Readiness, Item Analysis, Group Testing, Journal Articles
Lewy, Arieh – 1970
Statistical procedures employed in item analysis are based on the performance of the individual on a given test. If, however, one desires to assess the efficiency of some treatment, the researcher is interested in the performance of groups of students who received the treatment rather than in the performance of any one individual. In such cases…
Descriptors: Achievement Tests, Correlation, Group Norms, Group Testing
Guhn, Martin; Gadermann, Anne; Zumbo, Bruno D. – Early Education and Development, 2007
The present study investigates whether the Early Development Instrument (Offord & Janus, 1999) measures school readiness similarly across different groups of children. We employ ordinal logistic regression to investigate differential item functioning, a method of examining measurement bias. For 40,000 children, our analysis compares groups…
Descriptors: School Readiness, Kindergarten, Child Development, Program Validation
Montague, Margariete A. – 1972
This study investigated the feasibility of concurrently and randomly sampling examinees and items in order to estimate group achievement. Seven 32-item tests reflecting a 640-item universe of simple open sentences were used such that item selection (random, systematic) and assignment (random, systematic) of items (four, eight, sixteen) to forms…
Descriptors: Analysis of Variance, Elementary School Students, Group Testing, Item Analysis
Brennan, Robert L. – 1981
This handbook treats a restricted set of statistical procedures for addressing some of the most prevalent technical issues that arise in domain-referenced testing. The procedures discussed here were chosen because they do not necessitate extensive computations. The five major sections of the paper cover: (1) item analysis procedures for using data…
Descriptors: Classification, Criterion Referenced Tests, Cutting Scores, Group Testing
Sternberg, Robert J.; Grigorenko, Elena L.; Birney, Damian P.; Fredine, Nancy; Jarvin, Linda; Jeltova, Ida – National Research Center on the Gifted and Talented, 2007
The goal of this research project was to investigate the use of dynamic assessment to increase equity, fairness, and accuracy in the testing of abilities and achievement. Dynamic tests have been found to reveal developing expertise in underrepresented minorities around the world that is not revealed by conventional static tests. The…
Descriptors: Control Groups, Grade 4, Hispanic Americans, Asian Americans
Miller, M. David – 1981
Techniques for maximizing discrimination between groups in norm referenced measurement to reflect sensitivity to group differences inherent in the evaluation of multilevel educational systems are discussed. Data from the Beginning Teacher Evaluation Study (BTES) were used to examine relationships between reading and mathematics achievement and…
Descriptors: Achievement Tests, Correlation, Group Behavior, Group Testing

Bart, William.; And Others – Educational and Psychological Measurement, 1986
An alternative way of studying group differences is proposed based on ordering analysis using item hierarchies as a basis of comparison between two groups. Subjects were sets of twins in elementary school. Results showed that blacks and whites and males and females had similar item hierarchies for complex items. (Author/LMO)
Descriptors: Elementary Education, Group Testing, Intelligence Tests, Item Analysis