NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1,711 to 1,725 of 9,552 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Monroe, Scott – Journal of Educational and Behavioral Statistics, 2021
This research proposes a new statistic for testing latent variable distribution fit for unidimensional item response theory (IRT) models. If the typical assumption of normality is violated, then item parameter estimates will be biased, and dependent quantities such as IRT score estimates will be adversely affected. The proposed statistic compares…
Descriptors: Item Response Theory, Simulation, Scores, Comparative Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Supriyati, Yetti; Iriyadi, Deni; Falani, Ilham – Journal of Technology and Science Education, 2021
This study aims to develop a score equating application for computer-based school exams using parallel test kits with 25% anchor items. The items are arranged according to HOTS (High Order Thinking Skill) category, and use a scientific approach according to the physics lessons characteristics. Therefore, the questions were made using stimulus,…
Descriptors: Physics, Science Instruction, Teaching Methods, Equated Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Richardson, Connor J.; Smith, Trevor I.; Walter, Paul J. – Physical Review Physics Education Research, 2021
Ishimoto, Davenport, and Wittmann have previously reported analyses of data from student responses to the Force and Motion Conceptual Evaluation (FMCE), in which they used item response curves (IRCs) to make claims about American and Japanese students' relative likelihood to choose certain incorrect responses to some questions. We have used an…
Descriptors: Motion, Physics, Science Instruction, Concept Formation
Wang, Shichao; Li, Dongmei; Steedle, Jeffrey – ACT, Inc., 2021
Speeded tests set time limits so that few examinees can reach all items, and power tests allow most test-takers sufficient time to attempt all items. Educational achievement tests are sometimes described as "timed power tests" because the amount of time provided is intended to allow nearly all students to complete the test, yet this…
Descriptors: Timed Tests, Test Items, Achievement Tests, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Debeer, Dries; Ali, Usama S.; van Rijn, Peter W. – Journal of Educational Measurement, 2017
Test assembly is the process of selecting items from an item pool to form one or more new test forms. Often new test forms are constructed to be parallel with an existing (or an ideal) test. Within the context of item response theory, the test information function (TIF) or the test characteristic curve (TCC) are commonly used as statistical…
Descriptors: Test Format, Test Construction, Statistical Analysis, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Wrigley, Terry – FORUM: for promoting 3-19 comprehensive education, 2017
This text represents two extracts from a submission to the House of Commons Select Committee's investigation into primary school tests. The first part is a critique of the 2016 tests, particularly the Reading and Grammar tests for 11-year-olds and also the highly regulated "teacher assessment" of Writing. The second part is a set of…
Descriptors: Elementary Education, Tests, Student Evaluation, Reading Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Peterson, Christina Hamme; Peterson, N. Andrew; Powell, Kristen Gilmore – Measurement and Evaluation in Counseling and Development, 2017
Cognitive interviewing (CI) is a method to identify sources of confusion in assessment items and to assess validity evidence on the basis of content and response processes. We introduce readers to CI and describe a process for conducting such interviews and analyzing the results. Recommendations for best practice are provided.
Descriptors: Test Items, Test Construction, Interviews, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Clauser, Brian E.; Baldwin, Peter; Margolis, Melissa J.; Mee, Janet; Winward, Marcia – Journal of Educational Measurement, 2017
Validating performance standards is challenging and complex. Because of the difficulties associated with collecting evidence related to external criteria, validity arguments rely heavily on evidence related to internal criteria--especially evidence that expert judgments are internally consistent. Given its importance, it is somewhat surprising…
Descriptors: Evaluation Methods, Standard Setting, Cutting Scores, Expertise
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2017
An increasing concern of producers of educational assessments is fraudulent behavior during the assessment (van der Linden, 2009). Benefiting from item preknowledge (e.g., Eckerly, 2017; McLeod, Lewis, & Thissen, 2003) is one type of fraudulent behavior. This article suggests two new test statistics for detecting individuals who may have…
Descriptors: Test Items, Cheating, Testing Problems, Identification
Peer reviewed Peer reviewed
Direct linkDirect link
Sadeghi, Karim; Abolfazli Khonbi, Zainab – Language Testing in Asia, 2017
As perfectly summarised by Ida Lawrence, "Testing is growing by leaps and bounds across the world. There is a realization that a nation's well-being depends crucially on the educational achievement of its population. Valid tests are an essential tool to evaluate a nation's educational standing and to implement efficacious educational reforms.…
Descriptors: Test Items, Item Response Theory, Computer Assisted Testing, Adaptive Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Hilton, Charlotte Emma – International Journal of Social Research Methodology, 2017
The development of questionnaires, surveys and psychometric scales is an iterative research process that includes a number of carefully planned stages. Pretesting is a method of checking that questions work as intended and are understood by those individuals who are likely to respond to them. However, detailed reports of appropriate methods to…
Descriptors: Questionnaires, Pretesting, Interviews, Test Construction
Paneerselvam, Bavani – ProQuest LLC, 2017
Multiple-choice retrieval practice with additional lures reduces retention on a later test (Roediger & Marsh, 2005). However, the mechanism underlying the negative outcomes with additional lures is poorly understood. Given that the positive outcomes of retrieval practice are associated with enhanced relational and item-specific processing…
Descriptors: Multiple Choice Tests, Test Items, Item Analysis, Recall (Psychology)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zamora-Lobato, Teresa; García-Santillán, Arturo; Molchanova, Violetta S. – European Journal of Contemporary Education, 2019
The aim of study focused in identify if there are a set of variables that explain the level of anxiety towards mathematics in Telebachillerato students. For this, the test designed by Muñoz and Mato-Vázquez (2007) was used. The test comprises by 24 items with 5 dimensions. In order to get data, were surveyed 201 regular students enrolled in the…
Descriptors: Factor Analysis, Mathematics Anxiety, Secondary School Mathematics, High School Students
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Munoz, Albert; Mackay, Jonathon – Journal of University Teaching and Learning Practice, 2019
Online testing is a popular practice for tertiary educators, largely owing to efficiency in automation, scalability, and capability to add depth and breadth to subject offerings. As with all assessments, designs need to consider whether student cheating may be inadvertently made easier and more difficult to detect. Cheating can jeopardise the…
Descriptors: Cheating, Test Construction, Computer Assisted Testing, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Kroehne, Ulf; Buerger, Sarah; Hahnel, Carolin; Goldhammer, Frank – Educational Measurement: Issues and Practice, 2019
For many years, reading comprehension in the Programme for International Student Assessment (PISA) was measured via paper-based assessment (PBA). In the 2015 cycle, computer-based assessment (CBA) was introduced, raising the question of whether central equivalence criteria required for a valid interpretation of the results are fulfilled. As an…
Descriptors: Reading Comprehension, Computer Assisted Testing, Achievement Tests, Foreign Countries
Pages: 1  |  ...  |  111  |  112  |  113  |  114  |  115  |  116  |  117  |  118  |  119  |  ...  |  637