NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 3,571 to 3,585 of 9,552 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Padilla, Jose Luis; Hidalgo, M. Dolores; Benitez, Isabel; Gomez-Benito, Juana – Psicologica: International Journal of Methodology and Experimental Psychology, 2012
The analysis of differential item functioning (DIF) examines whether item responses differ according to characteristics such as language and ethnicity, when people with matching ability levels respond differently to the items. This analysis can be performed by calculating various statistics, one of the most important being the Mantel-Haenszel,…
Descriptors: Foreign Countries, Test Bias, Computer Software, Computer Software Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Perrett, Jamis J. – Journal of Statistics Education, 2012
This article demonstrates how textbooks differ in their description of the term "experimental unit". Advanced Placement Statistics teachers and students are often limited in their statistical knowledge by the information presented in their classroom textbook. Definitions and descriptions differ among textbooks as well as among different…
Descriptors: Statistics, Advanced Placement Programs, Textbooks, Mathematics Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Schroeder, Jacob; Murphy, Kristen L.; Holme, Thomas A. – Journal of Chemical Education, 2012
General chemistry tests from the Examinations Institute of the Division of Chemical Education of the American Chemical Society have been analyzed to identify factors that may influence how individual test items perform. In this paper, issues of item order (position within a set of items that comprise a test) and answer order (position of correct…
Descriptors: Chemistry, Test Items, Individual Testing, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Klauer, Karl Christoph; Kellen, David – Psychological Review, 2012
Rosner and Kochanski (2009) noticed an inconsistency in the mathematical statement of the Law of Categorical Judgment and derived "the valid equation, the Law of Categorical Judgment (Corrected)" (p. 125). The purpose of this comment is to point out that the law can be corrected in many different ways, leading to substantially different…
Descriptors: Test Items, Goodness of Fit, Mathematics Education, Models
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Buick, J. M. – European Journal of Physics Education, 2011
Aspects of assessment in physics are considered with the aim of designing assessments that will encourage a deep approach to student learning and will ultimately lead to higher levels of achievement. A range of physics questions are considered and categorized by the level of knowledge and understanding which is require for a successful answer.…
Descriptors: Physics, Taxonomy, Science Achievement, Knowledge Level
Peer reviewed Peer reviewed
Direct linkDirect link
van Hartingsveldt, Margo J.; de Groot, Imelda J. M.; Aarts, Pauline B. M.; Nijhuis-van der Sanden, Maria W. G. – Developmental Medicine & Child Neurology, 2011
Aim: To establish if there are psychometrically sound standardized tests or test items to assess handwriting readiness in 5- and 6-year-old children on the levels of occupations activities/tasks and performance. Method: Electronic databases were searched to identify measurement instruments. Tests were included in a systematic review if: (1)…
Descriptors: Writing Readiness, Test Items, Handwriting, Standardized Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Davis-Becker, Susan L.; Buckendahl, Chad W.; Gerrow, Jack – International Journal of Testing, 2011
Throughout the world, cut scores are an important aspect of a high-stakes testing program because they are a key operational component of the interpretation of test scores. One method for setting standards that is prevalent in educational testing programs--the Bookmark method--is intended to be a less cognitively complex alternative to methods…
Descriptors: Standard Setting (Scoring), Cutting Scores, Educational Testing, Licensing Examinations (Professions)
Peer reviewed Peer reviewed
Direct linkDirect link
Miller, Michael B.; Guerin, Scott A.; Wolford, George L. – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2011
The false memory effect produced by the Deese/Roediger & McDermott (DRM) paradigm is reportedly impervious to warnings to avoid false alarming to the critical lures (D. A. Gallo, H. L. Roediger III, & K. B. McDermott, 2001). This finding has been used as strong evidence against models that attribute the false alarms to a decision…
Descriptors: Models, Memory, Recognition (Psychology), Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Charlton, Shawn R.; Gossett, Bradley D.; Charlton, Veda A. – Psychological Record, 2011
Temporal discounting, the loss in perceived value associated with delayed outcomes, correlates with a number of personality measures, suggesting that an item-level analysis of trait measures might provide a more detailed understanding of discounting. The current report details two studies that investigate the utility of such an item-level…
Descriptors: Personality Measures, Test Items, Item Analysis, Delay of Gratification
Peer reviewed Peer reviewed
Direct linkDirect link
Fukuhara, Hirotaka; Kamata, Akihito – Applied Psychological Measurement, 2011
A differential item functioning (DIF) detection method for testlet-based data was proposed and evaluated in this study. The proposed DIF model is an extension of a bifactor multidimensional item response theory (MIRT) model for testlets. Unlike traditional item response theory (IRT) DIF models, the proposed model takes testlet effects into…
Descriptors: Item Response Theory, Test Bias, Test Items, Bayesian Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Oliveri, Maria E.; Ercikan, Kadriye – Applied Measurement in Education, 2011
In this study, we examine the degree of construct comparability and possible sources of incomparability of the English and French versions of the Programme for International Student Assessment (PISA) 2003 problem-solving measure administered in Canada. Several approaches were used to examine construct comparability at the test- (examination of…
Descriptors: Foreign Countries, English, French, Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Jones, Andrew T. – Applied Psychological Measurement, 2011
Practitioners often depend on item analysis to select items for exam forms and have a variety of options available to them. These include the point-biserial correlation, the agreement statistic, the B index, and the phi coefficient. Although research has demonstrated that these statistics can be useful for item selection, no research as of yet has…
Descriptors: Test Items, Item Analysis, Cutting Scores, Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Jinghua; Sinharay, Sandip; Holland, Paul; Feigenbaum, Miriam; Curley, Edward – Educational and Psychological Measurement, 2011
Two different types of anchors are investigated in this study: a mini-version anchor and an anchor that has a less spread of difficulty than the tests to be equated. The latter is referred to as a midi anchor. The impact of these two different types of anchors on observed score equating are evaluated and compared with respect to systematic error…
Descriptors: Equated Scores, Test Items, Difficulty Level, Statistical Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Svetina, Dubravka; Rutkowski, Leslie – Large-scale Assessments in Education, 2014
Background: When studying student performance across different countries or cultures, an important aspect for comparisons is that of score comparability. In other words, it is imperative that the latent variable (i.e., construct of interest) is understood and measured equivalently across all participating groups or countries, if our inferences…
Descriptors: Test Items, Item Response Theory, Item Analysis, Regression (Statistics)
Peer reviewed Peer reviewed
Direct linkDirect link
Kortemeyer, Gerd – Physical Review Special Topics - Physics Education Research, 2014
Item response theory (IRT) becomes an increasingly important tool when analyzing "big data" gathered from online educational venues. However, the mechanism was originally developed in traditional exam settings, and several of its assumptions are infringed upon when deployed in the online realm. For a large-enrollment physics course for…
Descriptors: Item Response Theory, Online Courses, Electronic Learning, Homework
Pages: 1  |  ...  |  235  |  236  |  237  |  238  |  239  |  240  |  241  |  242  |  243  |  ...  |  637