NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 2,881 to 2,895 of 9,552 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Yasuda, Jun-ichiro; Taniguchi, Masa-aki – Physical Review Special Topics - Physics Education Research, 2013
In this study, we evaluate the structural validity of Q.16 and Q.7 in the Force Concept Inventory (FCI). We address whether respondents who answer Q.16 and Q.7 correctly actually have an understanding of the concepts of physics tested in the questions. To examine respondents' levels of understanding, we use subquestions that test them on concepts…
Descriptors: Physics, Scientific Concepts, Science Tests, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Patton, Jeffrey M.; Cheng, Ying; Yuan, Ke-Hai; Diao, Qi – Applied Psychological Measurement, 2013
Variable-length computerized adaptive testing (VL-CAT) allows both items and test length to be "tailored" to examinees, thereby achieving the measurement goal (e.g., scoring precision or classification) with as few items as possible. Several popular test termination rules depend on the standard error of the ability estimate, which in turn depends…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Length, Ability
Peer reviewed Peer reviewed
Direct linkDirect link
San Martin, Ernesto; Rolin, Jean-Marie; Castro, Luis M. – Psychometrika, 2013
In this paper, we study the identification of a particular case of the 3PL model, namely when the discrimination parameters are all constant and equal to 1. We term this model, 1PL-G model. The identification analysis is performed under three different specifications. The first specification considers the abilities as unknown parameters. It is…
Descriptors: Item Response Theory, Models, Identification, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Gierl, Mark J.; Lai, Hollis; Li, Johnson – Educational Research and Evaluation, 2013
The purpose of this study is to evaluate the performance of CATSIB (Computer Adaptive Testing-Simultaneous Item Bias Test) for detecting differential item functioning (DIF) when items in the matching and studied subtest are administered adaptively in the context of a realistic multi-stage adaptive test (MST). MST was simulated using a 4-item…
Descriptors: Adaptive Testing, Test Bias, Computer Assisted Testing, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Bolt, Daniel M.; Wollack, James A.; Suh, Youngsuk – Psychometrika, 2012
Nested logit models have been presented as an alternative to multinomial logistic models for multiple-choice test items (Suh and Bolt in "Psychometrika" 75:454-473, 2010) and possess a mathematical structure that naturally lends itself to evaluating the incremental information provided by attending to distractor selection in scoring. One potential…
Descriptors: Test Items, Multiple Choice Tests, Models, Scoring
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Teker, Gulsen Tasdelen; Dogan, Nuri – Educational Sciences: Theory and Practice, 2015
Reliability and differential item functioning (DIF) analyses were conducted on testlets displaying local item dependence in this study. The data set employed in the research was obtained from the answers given by 1,500 students to the 20 items included in six testlets given in English Proficiency Exam by the School of Foreign Languages of a state…
Descriptors: Foreign Countries, Test Items, Test Bias, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Marie, S. Maria Josephine Arokia; Edannur, Sreekala – Journal of Educational Technology, 2015
This paper focused on the analysis of test items constructed in the paper of teaching Physical Science for B.Ed. class. It involved the analysis of difficulty level and discrimination power of each test item. Item analysis allows selecting or omitting items from the test, but more importantly item analysis is a tool to help the item writer improve…
Descriptors: Item Analysis, Relevance (Education), Standardized Tests, Achievement Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kamarulzaman, Mohammad Shah; Shaari, Ahmad Jelani – Malaysian Online Journal of Educational Technology, 2015
It has been long accepted that students are themselves great resources when it comes to developing questions and activity guidelines. The present study utilizes a strategic understanding of how students can be encouraged to perform better in preparation for exams, by allowing them to frame their own subject wise questions. The application of drill…
Descriptors: Test Items, Test Preparation, Test Wiseness, Academic Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Jackson, Margaret C.; Linden, David E. J.; Roberts, Mark V.; Kriegeskorte, Nikolaus; Haenschel, Corinna – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2015
A number of studies have shown that visual working memory (WM) is poorer for complex versus simple items, traditionally accounted for by higher information load placing greater demands on encoding and storage capacity limits. Other research suggests that it may not be complexity that determines WM performance per se, but rather increased…
Descriptors: Visual Perception, Short Term Memory, Test Items, Cognitive Processes
Peer reviewed Peer reviewed
Direct linkDirect link
Kam, Chester Chun Seng; Zhou, Mingming – Educational and Psychological Measurement, 2015
Previous research has found the effects of acquiescence to be generally consistent across item "aggregates" within a single survey (i.e., essential tau-equivalence), but it is unknown whether this phenomenon is consistent at the" individual item" level. This article evaluated the often assumed but inadequately tested…
Descriptors: Test Items, Surveys, Criteria, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Socha, Alan; DeMars, Christine E.; Zilberberg, Anna; Phan, Ha – International Journal of Testing, 2015
The Mantel-Haenszel (MH) procedure is commonly used to detect items that function differentially for groups of examinees from various demographic and linguistic backgrounds--for example, in international assessments. As in some other DIF methods, the total score is used to match examinees on ability. In thin matching, each of the total score…
Descriptors: Test Items, Educational Testing, Evaluation Methods, Ability Grouping
Peer reviewed Peer reviewed
Direct linkDirect link
Jones, Ian; Inglis, Matthew – Educational Studies in Mathematics, 2015
School mathematics examination papers are typically dominated by short, structured items that fail to assess sustained reasoning or problem solving. A contributory factor to this situation is the need for student work to be marked reliably by a large number of markers of varied experience and competence. We report a study that tested an…
Descriptors: Problem Solving, Mathematics Instruction, Mathematics Tests, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Tai, Sophie; Chen, Hao-Jan – Research-publishing.net, 2015
The communicative language teaching approach has dominated English teaching and learning since the 1970s. In Taiwan, standardized and highstakes English tests also put focus on the assessment of learners' communicative competence. While the test contents change, the modifications teachers made are superficial rather than substantial. A comparative…
Descriptors: Foreign Countries, Communicative Competence (Languages), Test Items, Item Analysis
Buri, John R.; Cromett, Cristina E.; Post, Maria C.; Landis, Anna Marie; Alliegro, Marissa C. – Online Submission, 2015
Rationale is presented for the derivation of a new measure of stressful life events for use with students [Negative Life Events Scale for Students (NLESS)]. Ten stressful life events questionnaires were reviewed, and the more than 600 items mentioned in these scales were culled based on the following criteria: (a) only long-term and unpleasant…
Descriptors: Experience, Social Indicators, Stress Variables, Affective Measures
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Jinming; Li, Jie – Journal of Educational Measurement, 2016
An IRT-based sequential procedure is developed to monitor items for enhancing test security. The procedure uses a series of statistical hypothesis tests to examine whether the statistical characteristics of each item under inspection have changed significantly during CAT administration. This procedure is compared with a previously developed…
Descriptors: Computer Assisted Testing, Test Items, Difficulty Level, Item Response Theory
Pages: 1  |  ...  |  189  |  190  |  191  |  192  |  193  |  194  |  195  |  196  |  197  |  ...  |  637