Publication Date
| In 2026 | 0 |
| Since 2025 | 200 |
| Since 2022 (last 5 years) | 1070 |
| Since 2017 (last 10 years) | 2580 |
| Since 2007 (last 20 years) | 4941 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – ETS Research Report Series, 2008
This study examined variations of a nonequivalent groups equating design used with constructed-response (CR) tests to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, the study investigated the use of anchor CR item rescoring in the context of classical…
Descriptors: Equated Scores, Comparative Analysis, Test Format, Responses
Oosterhof, Albert; Rohani, Faranak; Sanfilippo, Carol; Stillwell, Peggy; Hawkins, Karen – Online Submission, 2008
In assessment, the ability to construct test items that measure a targeted skill is fundamental to validity and alignment. The ability to do the reverse is also important: determining what skill an existing test item measures. This paper presents a model for classifying test items that builds on procedures developed by others, including Bloom…
Descriptors: Test Items, Classification, Models, Cognitive Ability
Ullstadius, Eva; Carlstedt, Berit; Gustafsson, Jan-Eric – International Journal of Testing, 2008
The influence of general and verbal ability on each of 72 verbal analogy test items were investigated with new factor analytical techniques. The analogy items together with the Computerized Swedish Enlistment Battery (CAT-SEB) were given randomly to two samples of 18-year-old male conscripts (n = 8566 and n = 5289). Thirty-two of the 72 items had…
Descriptors: Test Items, Verbal Ability, Factor Analysis, Swedish
Verheggen, M. M.; Muijtjens, A. M. M.; Os, J. Van; Schuwirth, L. W. T. – Advances in Health Sciences Education, 2008
Background: To establish credible, defensible and acceptable passing scores for written tests is a challenge for health profession educators. Angoff procedures are often used to establish pass/fail decisions for written and performance tests. In an Angoff procedure judges' expertise and professional skills are assumed to influence their ratings of…
Descriptors: Health Occupations, Performance Tests, Scoring, Item Response Theory
Wang, Xiaohui; Bradlow, Eric T.; Wainer, Howard; Muller, Eric S. – Journal of Educational and Behavioral Statistics, 2008
In the course of screening a form of a medical licensing exam for items that function differentially (DIF) between men and women, the authors used the traditional Mantel-Haenszel (MH) statistic for initial screening and a Bayesian method for deeper analysis. For very easy items, the MH statistic unexpectedly often found DIF where there was none.…
Descriptors: Bayesian Statistics, Licensing Examinations (Professions), Medicine, Test Items
Penfield, Randall D. – Applied Psychological Measurement, 2008
The examination of measurement invariance in polytomous items is complicated by the possibility that the magnitude and sign of lack of invariance may vary across the steps underlying the set of polytomous response options, a concept referred to as differential step functioning (DSF). This article describes three classes of nonparametric DSF effect…
Descriptors: Simulation, Nonparametric Statistics, Item Response Theory, Computation
Camilli, Gregory; Prowker, Adam; Dossey, John A.; Lindquist, Mary M.; Chiu, Ting-Wei; Vargas, Sadako; de la Torre, Jimmy – Journal of Educational Measurement, 2008
A new method for analyzing differential item functioning is proposed to investigate the relative strengths and weaknesses of multiple groups of examinees. Accordingly, the notion of a conditional measure of difference between two groups (Reference and Focal) is generalized to a conditional variance. The objective of this article is to present and…
Descriptors: Test Bias, National Competency Tests, Grade 4, Difficulty Level
Frank, Brian W.; Kanim, Stephen E.; Gomez, Luanna S. – Physical Review Special Topics - Physics Education Research, 2008
We describe the results of an experiment conducted to test predictions about student responses to questions about motion based on an explicit model of student thinking in terms of the cuing of a variety of different physical intuitions or conceptual resources. This particular model allows us to account for observed variations in patterns of…
Descriptors: Prediction, Student Reaction, College Students, Test Items
Korobko, Oksana B.; Glas, Cees A. W.; Bosker, Roel J.; Luyten, Johan W. – Journal of Educational Measurement, 2008
Methods are presented for comparing grades obtained in a situation where students can choose between different subjects. It must be expected that the comparison between the grades is complicated by the interaction between the students' pattern and level of proficiency on one hand, and the choice of the subjects on the other hand. Three methods…
Descriptors: Item Response Theory, Test Items, Comparative Analysis, Grades (Scholastic)
Dodson, Chad S.; Darragh, James; Williams, Allison – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2008
When expectations and stereotypes are activated at retrieval, they spontaneously create distorted and illusory recollections that are consistent with these expectations. Participants studied doctor (physician)-related and lawyer-related statements that were presented by 2 different people. When informed, on a subsequent source memory test, (i.e.,…
Descriptors: Test Items, Stereotypes, Familiarity, Memory
Morrow, James R., Jr.; Ede, Alison – Research Quarterly for Exercise and Sport, 2009
Statewide physical fitness testing is gaining popularity in the United States because of increased childhood obesity levels, the relations between physical fitness and academic performance, and the hypothesized relations between adult characteristics and childhood physical activity, physical fitness, and health behaviors. Large-scale physical…
Descriptors: Obesity, Physical Activities, Testing, Physical Fitness
Davison, Mark L.; Kim, Se-Kang; Close, Catherine – Multivariate Behavioral Research, 2009
A profile is a vector of scores for one examinee. The mean score in the vector can be interpreted as a measure of overall profile height, the variance can be interpreted as a measure of within person variation, and the ipsatized vector of score deviations about the mean can be said to describe the pattern in the score profile. A within person…
Descriptors: Vocational Interests, Interest Inventories, Profiles, Scores
Zhao, Wen; Coniam, David – Journal of Vocational Education and Training, 2009
Within the general context of the attention being given to vocational education in China and imminent major reform in the English language vocational education curriculum, this paper describes the construction of a set of calibrated tests of reading skills, grammar and vocabulary knowledge to inform curriculum planners about levels of ability in…
Descriptors: Foreign Countries, Vocational Education, English (Second Language), Secondary School Curriculum
Parshall, Cynthia G.; Harmes, J. Christine – Journal of Applied Testing Technology, 2009
Many exam programs have begun to include innovative item types in their operational assessments. While innovative item types appear to have great promise for expanding measurement, there can also be genuine challenges to their successful implementation. In this paper we present a set of four activities that can be beneficially incorporated into…
Descriptors: Test Items, Test Construction, Measurement, Educational Assessment
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment

Peer reviewed
Direct link
