NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers3
What Works Clearinghouse Rating
Showing 1 to 15 of 68 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Walker, Michael E. – Educational Measurement: Issues and Practice, 2022
Test equating requires collecting data to link the scores from different forms of a test. Problems arise when equating samples are not equivalent and the test forms to be linked share no common items by which to measure or adjust for the group nonequivalence. Using data from five operational test forms, we created five pairs of research forms for…
Descriptors: Ability, Tests, Equated Scores, Testing Problems
Zita Lysaght; Michael O'Leary; Angela Mazzone; Conor Scully – Sage Research Methods Cases, 2022
Since 2018, colleagues from two research centers at Dublin City University have been collaborating to develop a measurement scale to assess individuals' ability to identify workplace bullying. Having agreed on an operational definition of the construct, an item pool of 26 workplace bullying scenarios, that is, short descriptions of…
Descriptors: Foreign Countries, Test Construction, Test Validity, Test Reliability
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sánchez Sánchez, Ernesto; García Rios, Víctor N.; Silvestre Castro, Eleazar; Licea, Guadalupe Carrasco – North American Chapter of the International Group for the Psychology of Mathematics Education, 2020
In this paper, we address the following questions: What misconceptions do high school students exhibit in their first encounter with significance test problems through a repeated sampling approach? Which theory or framework could explain the presence and features of such patterns? With brief prior instruction on the use of Fathom software to…
Descriptors: High School Students, Misconceptions, Statistical Significance, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Diao, Hongyu; Keller, Lisa – Applied Measurement in Education, 2020
Examinees who attempt the same test multiple times are often referred to as "repeaters." Previous studies suggested that repeaters should be excluded from the total sample before equating because repeater groups are distinguishable from non-repeater groups. In addition, repeaters might memorize anchor items, causing item drift under a…
Descriptors: Licensing Examinations (Professions), College Entrance Examinations, Repetition, Testing Problems
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Egbo, Anthonia Chinonyelum – World Journal of Education, 2015
This study investigated the Counselling strategies for curbing "Examination Malpractices" in Secondary Schools in Enugu State Nigeria. The researcher used three research questions. The Design used was a descriptive survey design. Sample consisted of 335 respondents comprising principals (N = 19), PTA secretaries (N = 19), teachers (N =…
Descriptors: Counseling Techniques, Questionnaires, Foreign Countries, Surveys
Peer reviewed Peer reviewed
Direct linkDirect link
Phillips, Gary W. – Applied Measurement in Education, 2015
This article proposes that sampling design effects have potentially huge unrecognized impacts on the results reported by large-scale district and state assessments in the United States. When design effects are unrecognized and unaccounted for they lead to underestimating the sampling error in item and test statistics. Underestimating the sampling…
Descriptors: State Programs, Sampling, Research Design, Error of Measurement
Haberman, Shelby J. – Educational Testing Service, 2010
Sampling errors limit the accuracy with which forms can be linked. Limitations on accuracy are especially important in testing programs in which a very large number of forms are employed. Standard inequalities in mathematical statistics may be used to establish lower bounds on the achievable inking accuracy. To illustrate results, a variety of…
Descriptors: Testing Programs, Equated Scores, Sampling, Accuracy
Sukin, Tia M. – ProQuest LLC, 2010
The presence of outlying anchor items is an issue faced by many testing agencies. The decision to retain or remove an item is a difficult one, especially when the content representation of the anchor set becomes questionable by item removal decisions. Additionally, the reason for the aberrancy is not always clear, and if the performance of the…
Descriptors: Simulation, Science Achievement, Sampling, Data Analysis
Wang, Shudong; Jiao, Hong; Jin, Ying; Thum, Yeow Meng – Online Submission, 2010
The vertical scales of large-scale achievement tests created by using item response theory (IRT) models are mostly based on cluster (or correlated) educational data in which students usually are clustered in certain groups or settings (classrooms or schools). While such application directly violated assumption of independent sample of person in…
Descriptors: Scaling, Achievement Tests, Data Analysis, Item Response Theory
Hart, Ray; Casserly, Michael; Uzzell, Renata; Palacios, Moses; Corcoran, Amanda; Spurgeon, Liz – Council of the Great City Schools, 2015
There has been little data collected on how much testing actually goes on in America's schools and how the results are used. So in the Spring of 2014, the Council staff developed and launched a survey of assessment practices. This report presents the findings from that survey and subsequent Council analysis and review of the data. It also offers…
Descriptors: Urban Schools, Student Evaluation, Testing Programs, Testing
Shoemaker, David M. – 1972
Investigated empirically through post mortem item-examinee sampling were the relative merits of two alternative procedures for allocating items to subtests in multiple matrix sampling and the feasibility of using the jackknife in approximating standard errors of estimate. The results indicate clearly that a partially balanced incomplete block…
Descriptors: Error of Measurement, Item Sampling, Matrices, Sampling
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Yuguo; Small, Dylan – Psychometrika, 2005
Rasch proposed an exact conditional inference approach to testing his model but never implemented it because it involves the calculation of a complicated probability. This paper furthers Rasch's approach by (1) providing an efficient Monte Carlo methodology for accurately approximating the required probability and (2) illustrating the usefulness…
Descriptors: Testing Problems, Probability, Methods, Testing
National Education Association, Washington, DC. – 1975
The National Education Association's Task Force on Testing has stated its opinion that standardized tests are overused. The task force suggests that the application of sampling techniques and a variety of alternatives to current testing practices would accomplish the same purposes. Representatives of the testing industry have indicated that the…
Descriptors: Accountability, Alternative Assessment, Cost Effectiveness, Educational Testing
Thrash, Susan K.; Porter, Andrew C. – 1974
The purpose of this paper is to prove that one currently recommended method of obtaining the reliability of an instrument defined on a population of aggregate units is invalid. This method randomly splits the aggregate into two halves, correlates the two half unit scores by a Pearson product moment correlation coefficient, and corrects the…
Descriptors: Comparative Analysis, Correlation, Measurement Techniques, Sampling
Peer reviewed Peer reviewed
Philp, H.; Kelly, M. – British Journal of Educational Psychology, 1974
Article examined the validity of some of Piaget's and Bruner's findings in non-western cultures, the relationship between cognitive models that they developed, and the implications for curriculum, teaching and learning of any data newly discovered. (Author/RK)
Descriptors: Cognitive Development, Educational Psychology, Elementary School Students, Measurement Instruments
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5