Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 8 |
Descriptor
Comparative Testing | 28 |
Item Analysis | 28 |
Test Items | 28 |
Test Format | 12 |
Multiple Choice Tests | 9 |
Higher Education | 8 |
Test Validity | 8 |
Test Construction | 7 |
Comparative Analysis | 6 |
Foreign Countries | 6 |
Achievement Tests | 5 |
More ▼ |
Source
Author
Clarke, S. C. T. | 2 |
Hughes, Carolyn | 2 |
Little, Todd D. | 2 |
Palmer, Susan B. | 2 |
Seo, Hyojeong | 2 |
Shogren, Karrie A. | 2 |
Thompson, James R. | 2 |
Wehmeyer, Michael L. | 2 |
Albanese, Mark A. | 1 |
Ankenman, Robert D. | 1 |
Bejar, Isaac I. | 1 |
More ▼ |
Publication Type
Reports - Research | 22 |
Journal Articles | 14 |
Speeches/Meeting Papers | 6 |
Reports - Evaluative | 5 |
Reports - Descriptive | 1 |
Education Level
Elementary Secondary Education | 3 |
Elementary Education | 2 |
Higher Education | 2 |
Grade 4 | 1 |
Grade 8 | 1 |
Audience
Researchers | 2 |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Seo, Hyojeong; Shogren, Karrie A.; Wehmeyer, Michael L.; Hughes, Carolyn; Thompson, James R.; Little, Todd D.; Palmer, Susan B. – Career Development and Transition for Exceptional Individuals, 2016
This study examined similarities and differences in measurement properties and score comparability of the "Supports Intensity Scale-Adult Version" (16-64 years) and the "Supports Intensity Scale-Children's Version" (5-16 years). Data were collected from 142 adolescents with intellectual disability with both versions of the…
Descriptors: Adolescents, Intellectual Disability, Special Needs Students, Transitional Programs
Seo, Hyojeong; Shogren, Karrie A.; Wehmeyer, Michael L.; Hughes, Carolyn; Thompson, James R.; Little, Todd D.; Palmer, Susan B. – Grantee Submission, 2016
This study examined similarities and differences in measurement properties and score comparability of the "Supports Intensity Scale-Adult Version" (16-64 years) and the "Supports Intensity Scale-Children's Version" (5-16 years). Data were collected from 142 adolescents with intellectual disability with both versions of the…
Descriptors: Adolescents, Intellectual Disability, Special Needs Students, Transitional Programs
Laprise, Shari L. – College Teaching, 2012
Successful exam composition can be a difficult task. Exams should not only assess student comprehension, but be learning tools in and of themselves. In a biotechnology course delivered to nonmajors at a business college, objective multiple-choice test questions often require students to choose the exception or "not true" choice. Anecdotal student…
Descriptors: Feedback (Response), Test Items, Multiple Choice Tests, Biotechnology
Taherbhai, Husein; Seo, Daeryong; Bowman, Trinell – British Educational Research Journal, 2012
Literature in the United States provides many examples of no difference in student achievement when measured against the mode of test administration i.e., paper-pencil and online versions of the test. However, most of these researches centre on "regular" students who do not require differential teaching methods or different evaluation…
Descriptors: Learning Disabilities, Statistical Analysis, Teaching Methods, Test Format
Sparfeldt, Jorn R.; Kimmel, Rumena; Lowenkamp, Lena; Steingraber, Antje; Rost, Detlef H. – Educational Assessment, 2012
Multiple-choice (MC) reading comprehension test items comprise three components: text passage, questions about the text, and MC answers. The construct validity of this format has been repeatedly criticized. In three between-subjects experiments, fourth graders (N[subscript 1] = 230, N[subscript 2] = 340, N[subscript 3] = 194) worked on three…
Descriptors: Test Items, Reading Comprehension, Construct Validity, Grade 4
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias
Schulz, Wolfram; Fraillon, Julian – Educational Research and Evaluation, 2011
When comparing data derived from tests or questionnaires in cross-national studies, researchers commonly assume measurement invariance in their underlying scaling models. However, different cultural contexts, languages, and curricula can have powerful effects on how students respond in different countries. This article illustrates how the…
Descriptors: Citizenship Education, International Studies, Item Response Theory, International Education
Ferdous, Abdullah A.; Plake, Barbara S. – Educational and Psychological Measurement, 2007
In an Angoff standard setting procedure, judges estimate the probability that a hypothetical randomly selected minimally competent candidate will answer correctly each item in the test. In many cases, these item performance estimates are made twice, with information shared with the panelists between estimates. Especially for long tests, this…
Descriptors: Test Items, Probability, Item Analysis, Standard Setting (Scoring)

Ilai, Doron; Willerman, Lee – Intelligence, 1989
Items showing sex differences on the revised Wechsler Adult Intelligence Scale (WAIS-R) were studied. In a sample of 206 young adults (110 males and 96 females), 15 items demonstrated significant sex differences, but there was no relationship of item-specific gender content to sex differences in item performance. (SLD)
Descriptors: Comparative Testing, Females, Intelligence Tests, Item Analysis

Crehan, Kevin D.; And Others – Educational and Psychological Measurement, 1993
Studies with 220 college students found that multiple-choice test items with 3 items are more difficult than those with 4 items, and items with the none-of-these option are more difficult than those without this option. Neither format manipulation affected item discrimination. Implications for test construction are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Difficulty Level, Distractors (Tests)
Chen, Shu-Ying; Ankenman, Robert D. – Journal of Educational Measurement, 2004
The purpose of this study was to compare the effects of four item selection rules--(1) Fisher information (F), (2) Fisher information with a posterior distribution (FP), (3) Kullback-Leibler information with a posterior distribution (KP), and (4) completely randomized item selection (RN)--with respect to the precision of trait estimation and the…
Descriptors: Test Length, Adaptive Testing, Computer Assisted Testing, Test Selection
Byrne, Barbara M.; And Others – 1991
Extending the earlier work of B. M. Byrne and P. Baron (1990), the factorial invariance of the 21-item Beck Depression Inventory (BDI) was tested using 351 non-clinical adolescent males and 334 non-clinical adolescent females. All subjects were in grades 9 through 12 and attended the same secondary school in a large metropolitan area in central…
Descriptors: Adolescents, Affective Measures, Analysis of Covariance, Comparative Testing
Bethscheider, Janine K. – 1992
Standard and experimental forms of the Johnson O'Connor Research Foundations Analytical Reasoning test were administered to 1,496 clients of the Foundation (persons seeking information about aptitude for educational and career decisions). The objectives were to develop a new form of the test and to better understand what makes some items more…
Descriptors: Adults, Aptitude Tests, Career Choice, Comparative Testing
Chissom, Brad; Chukabarah, Prince C. O. – 1985
The comparative effects of various sequences of test items were examined for over 900 graduate students enrolled in an educational research course at The University of Alabama, Tuscaloosa. experiment, which was conducted a total of four times using four separate tests, presented three different arrangements of 50 multiple-choice items: (1)…
Descriptors: Analysis of Variance, Comparative Testing, Difficulty Level, Graduate Students
Pine, Steven M.; Weiss, David J. – 1978
This report examines how selection fairness is influenced by the characteristics of a selection instrument in terms of its distribution of item difficulties, level of item discrimination, degree of item bias, and testing strategy. Computer simulation was used in the administration of either a conventional or Bayesian adaptive ability test to a…
Descriptors: Adaptive Testing, Bayesian Statistics, Comparative Testing, Computer Assisted Testing
Previous Page | Next Page ยป
Pages: 1 | 2