NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1,696 to 1,710 of 9,533 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Debeer, Dries; Ali, Usama S.; van Rijn, Peter W. – Journal of Educational Measurement, 2017
Test assembly is the process of selecting items from an item pool to form one or more new test forms. Often new test forms are constructed to be parallel with an existing (or an ideal) test. Within the context of item response theory, the test information function (TIF) or the test characteristic curve (TCC) are commonly used as statistical…
Descriptors: Test Format, Test Construction, Statistical Analysis, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Wrigley, Terry – FORUM: for promoting 3-19 comprehensive education, 2017
This text represents two extracts from a submission to the House of Commons Select Committee's investigation into primary school tests. The first part is a critique of the 2016 tests, particularly the Reading and Grammar tests for 11-year-olds and also the highly regulated "teacher assessment" of Writing. The second part is a set of…
Descriptors: Elementary Education, Tests, Student Evaluation, Reading Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Peterson, Christina Hamme; Peterson, N. Andrew; Powell, Kristen Gilmore – Measurement and Evaluation in Counseling and Development, 2017
Cognitive interviewing (CI) is a method to identify sources of confusion in assessment items and to assess validity evidence on the basis of content and response processes. We introduce readers to CI and describe a process for conducting such interviews and analyzing the results. Recommendations for best practice are provided.
Descriptors: Test Items, Test Construction, Interviews, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Clauser, Brian E.; Baldwin, Peter; Margolis, Melissa J.; Mee, Janet; Winward, Marcia – Journal of Educational Measurement, 2017
Validating performance standards is challenging and complex. Because of the difficulties associated with collecting evidence related to external criteria, validity arguments rely heavily on evidence related to internal criteria--especially evidence that expert judgments are internally consistent. Given its importance, it is somewhat surprising…
Descriptors: Evaluation Methods, Standard Setting, Cutting Scores, Expertise
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2017
An increasing concern of producers of educational assessments is fraudulent behavior during the assessment (van der Linden, 2009). Benefiting from item preknowledge (e.g., Eckerly, 2017; McLeod, Lewis, & Thissen, 2003) is one type of fraudulent behavior. This article suggests two new test statistics for detecting individuals who may have…
Descriptors: Test Items, Cheating, Testing Problems, Identification
Peer reviewed Peer reviewed
Direct linkDirect link
Sadeghi, Karim; Abolfazli Khonbi, Zainab – Language Testing in Asia, 2017
As perfectly summarised by Ida Lawrence, "Testing is growing by leaps and bounds across the world. There is a realization that a nation's well-being depends crucially on the educational achievement of its population. Valid tests are an essential tool to evaluate a nation's educational standing and to implement efficacious educational reforms.…
Descriptors: Test Items, Item Response Theory, Computer Assisted Testing, Adaptive Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Hilton, Charlotte Emma – International Journal of Social Research Methodology, 2017
The development of questionnaires, surveys and psychometric scales is an iterative research process that includes a number of carefully planned stages. Pretesting is a method of checking that questions work as intended and are understood by those individuals who are likely to respond to them. However, detailed reports of appropriate methods to…
Descriptors: Questionnaires, Pretesting, Interviews, Test Construction
Paneerselvam, Bavani – ProQuest LLC, 2017
Multiple-choice retrieval practice with additional lures reduces retention on a later test (Roediger & Marsh, 2005). However, the mechanism underlying the negative outcomes with additional lures is poorly understood. Given that the positive outcomes of retrieval practice are associated with enhanced relational and item-specific processing…
Descriptors: Multiple Choice Tests, Test Items, Item Analysis, Recall (Psychology)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zamora-Lobato, Teresa; García-Santillán, Arturo; Molchanova, Violetta S. – European Journal of Contemporary Education, 2019
The aim of study focused in identify if there are a set of variables that explain the level of anxiety towards mathematics in Telebachillerato students. For this, the test designed by Muñoz and Mato-Vázquez (2007) was used. The test comprises by 24 items with 5 dimensions. In order to get data, were surveyed 201 regular students enrolled in the…
Descriptors: Factor Analysis, Mathematics Anxiety, Secondary School Mathematics, High School Students
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Munoz, Albert; Mackay, Jonathon – Journal of University Teaching and Learning Practice, 2019
Online testing is a popular practice for tertiary educators, largely owing to efficiency in automation, scalability, and capability to add depth and breadth to subject offerings. As with all assessments, designs need to consider whether student cheating may be inadvertently made easier and more difficult to detect. Cheating can jeopardise the…
Descriptors: Cheating, Test Construction, Computer Assisted Testing, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Kroehne, Ulf; Buerger, Sarah; Hahnel, Carolin; Goldhammer, Frank – Educational Measurement: Issues and Practice, 2019
For many years, reading comprehension in the Programme for International Student Assessment (PISA) was measured via paper-based assessment (PBA). In the 2015 cycle, computer-based assessment (CBA) was introduced, raising the question of whether central equivalence criteria required for a valid interpretation of the results are fulfilled. As an…
Descriptors: Reading Comprehension, Computer Assisted Testing, Achievement Tests, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Reed, Deborah K.; Stevenson, Nathan; LeBeau, Brandon C. – Elementary School Journal, 2019
This study investigated the effects of imposing task- or process-oriented reading behaviors on reading comprehension assessment performance. Students in grades 5-8 (N = 275) were randomly assigned to hear multiple-choice items read aloud before or after reading a test passage and when they were and were not allowed access to the passage while…
Descriptors: Reading Comprehension, Reading Tests, Multiple Choice Tests, Reading Aloud to Others
Peer reviewed Peer reviewed
Direct linkDirect link
Ackermann, Nicole; Siegfried, Christin – Citizenship, Social and Economics Education, 2019
Studies indicate that male students outperform female students in economic literacy and that a specific item format (selected-response, constructed-response) favours either males or females. This study analyses the relationship between item format and gender in economic-civic competence using the WBK-T2 test ("revidierter Test zur…
Descriptors: Foreign Countries, Test Format, Economics, Knowledge Level
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Gu, Lixiong; Ling, Guangming; Qu, Yanxuan – ETS Research Report Series, 2019
Research has found that the "a"-stratified item selection strategy (STR) for computerized adaptive tests (CATs) may lead to insufficient use of high a items at later stages of the tests and thus to reduced measurement precision. A refined approach, unequal item selection across strata (USTR), effectively improves test precision over the…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Use, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Raymond, Mark R.; Stevens, Craig; Bucak, S. Deniz – Advances in Health Sciences Education, 2019
Research suggests that the three-option format is optimal for multiple choice questions (MCQs). This conclusion is supported by numerous studies showing that most distractors (i.e., incorrect answers) are selected by so few examinees that they are essentially nonfunctional. However, nearly all studies have defined a distractor as nonfunctional if…
Descriptors: Multiple Choice Tests, Credentials, Test Format, Test Items
Pages: 1  |  ...  |  110  |  111  |  112  |  113  |  114  |  115  |  116  |  117  |  118  |  ...  |  636