Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 32 |
Since 2006 (last 20 years) | 51 |
Descriptor
Comparative Analysis | 61 |
Statistical Analysis | 61 |
Test Format | 61 |
Foreign Countries | 24 |
Test Items | 24 |
Scores | 22 |
Multiple Choice Tests | 20 |
Computer Assisted Testing | 16 |
Correlation | 13 |
Item Response Theory | 12 |
College Students | 10 |
More ▼ |
Source
Author
Ali, Usama S. | 2 |
Livingston, Samuel A. | 2 |
Adair, Desmond | 1 |
Ahmadi, Alireza | 1 |
Aksakalli, Ayhan | 1 |
Alemi, Minoo | 1 |
Alpayar, Cagla | 1 |
Anakwe, Bridget | 1 |
Anderson, Daniel | 1 |
Ardi, Muhammad | 1 |
Backes, Ben | 1 |
More ▼ |
Publication Type
Reports - Research | 57 |
Journal Articles | 51 |
Speeches/Meeting Papers | 5 |
Tests/Questionnaires | 4 |
Reports - Evaluative | 2 |
Dissertations/Theses -… | 1 |
Numerical/Quantitative Data | 1 |
Reports - Descriptive | 1 |
Education Level
Audience
Location
Germany | 3 |
Iran | 3 |
Turkey | 3 |
Sweden | 2 |
Taiwan | 2 |
Chile (Santiago) | 1 |
China | 1 |
Denmark | 1 |
Ghana | 1 |
Indonesia | 1 |
Israel | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Soysal, Sumeyra; Yilmaz Kogar, Esin – International Journal of Assessment Tools in Education, 2021
In this study, whether item position effects lead to DIF in the condition where different test booklets are used was investigated. To do this the methods of Lord's chi-square and Raju's unsigned area with the 3PL model under with and without item purification were used. When the performance of the methods was compared, it was revealed that…
Descriptors: Item Response Theory, Test Bias, Test Items, Comparative Analysis
Debeer, Dries; Ali, Usama S.; van Rijn, Peter W. – Journal of Educational Measurement, 2017
Test assembly is the process of selecting items from an item pool to form one or more new test forms. Often new test forms are constructed to be parallel with an existing (or an ideal) test. Within the context of item response theory, the test information function (TIF) or the test characteristic curve (TCC) are commonly used as statistical…
Descriptors: Test Format, Test Construction, Statistical Analysis, Comparative Analysis
Neuert, Cornelia E. – Field Methods, 2017
Previous research has shown that check-all-that-apply (CATA) and forced-choice (FC) question formats do not produce comparable results. The cognitive processes underlying respondents' answers to both types of formats still require clarification. This study contributes to filling this gap by using eye-tracking data. Both formats are compared by…
Descriptors: Measurement Techniques, Test Format, Eye Movements, Cognitive Processes
Kieftenbeld, Vincent; Boyer, Michelle – Applied Measurement in Education, 2017
Automated scoring systems are typically evaluated by comparing the performance of a single automated rater item-by-item to human raters. This presents a challenge when the performance of multiple raters needs to be compared across multiple items. Rankings could depend on specifics of the ranking procedure; observed differences could be due to…
Descriptors: Automation, Scoring, Comparative Analysis, Test Items
Wang, Lu; Steedle, Jeffrey – ACT, Inc., 2020
In recent ACT mode comparability studies, students testing on laptop or desktop computers earned slightly higher scores on average than students who tested on paper, especially on the ACT® reading and English tests (Li et al., 2017). Equating procedures adjust for such "mode effects" to make ACT scores comparable regardless of testing…
Descriptors: Test Format, Reading Tests, Language Tests, English
Ford, Jeremy W.; Conoyer, Sarah J.; Lembke, Erica S.; Smith, R. Alex; Hosp, John L. – Assessment for Effective Intervention, 2018
In the present study, two types of curriculum-based measurement (CBM) tools in science, Vocabulary Matching (VM) and Statement Verification for Science (SV-S), a modified Sentence Verification Technique, were compared. Specifically, this study aimed to determine whether the format of information presented (i.e., SV-S vs. VM) produces differences…
Descriptors: Curriculum Based Assessment, Evaluation Methods, Measurement Techniques, Comparative Analysis
Tengberg, Michael – Language Testing, 2017
Reading comprehension tests are often assumed to measure the same, or at least similar, constructs. Yet, reading is not a single but a multidimensional form of processing, which means that variations in terms of reading material and item design may emphasize one aspect of the construct at the cost of another. The educational systems in Denmark,…
Descriptors: Foreign Countries, National Competency Tests, Reading Tests, Comparative Analysis
Hubbard, Joanna K.; Potts, Macy A.; Couch, Brian A. – CBE - Life Sciences Education, 2017
Assessments represent an important component of undergraduate courses because they affect how students interact with course content and gauge student achievement of course objectives. To make decisions on assessment design, instructors must understand the affordances and limitations of available question formats. Here, we use a crossover…
Descriptors: Test Format, Questioning Techniques, Undergraduate Students, Objective Tests
Aksakalli, Ayhan; Turgut, Umit; Salar, Riza – Journal of Education and Practice, 2016
The purpose of this study is to investigate whether students are more successful on abstract or illustrated test questions. To this end, the questions on an abstract test were changed into a visual format, and these tests were administered every three days to a total of 240 students at six middle schools located in the Erzurum city center and…
Descriptors: Comparative Analysis, Scores, Middle School Students, Grade 8
Backes, Ben; Cowan, James – National Center for Analysis of Longitudinal Data in Education Research (CALDER), 2018
Nearly two dozen states now administer online exams. These tests have real consequences: their results feed into accountability systems, which have been used for more than a decade to hold schools and districts accountable for their students' learning. We examine the rollout of computer-based testing in Massachusetts over 2 years to investigate…
Descriptors: Computer Assisted Testing, Academic Achievement, Standardized Tests, Achievement Tests
Jonick, Christine; Schneider, Jennifer; Boylan, Daniel – Accounting Education, 2017
The purpose of the research is to examine the effect of different response formats on student performance on introductory accounting exam questions. The study analyzes 1104 accounting students' responses to quantitative questions presented in two formats: multiple-choice and fill-in. Findings indicate that response format impacts student…
Descriptors: Introductory Courses, Accounting, Test Format, Multiple Choice Tests
Bendulo, Hermabeth O.; Tibus, Erlinda D.; Bande, Rhodora A.; Oyzon, Voltaire Q.; Milla, Norberto E.; Macalinao, Myrna L. – International Journal of Evaluation and Research in Education, 2017
Testing or evaluation in an educational context is primarily used to measure or evaluate and authenticate the academic readiness, learning advancement, acquisition of skills, or instructional needs of learners. This study tried to determine whether the varied combinations of arrangements of options and letter cases in a Multiple-Choice Test (MCT)…
Descriptors: Test Format, Multiple Choice Tests, Test Construction, Eye Movements
Beserra, Vagner; Nussbaum, Miguel; Grass, Antonio – Interactive Learning Environments, 2017
When using educational video games, particularly drill-and-practice video games, there are several ways of providing an answer to a quiz. The majority of paper-based options can be classified as being either multiple-choice or constructed-response. Therefore, in the process of creating an educational drill-and-practice video game, one fundamental…
Descriptors: Multiple Choice Tests, Drills (Practice), Educational Games, Video Games
Lee, Yuan-Hsuan – Interactive Learning Environments, 2018
Premised on Web 2.0 technology, the current study investigated the effect of facilitating critical thinking using the Collaborative Questioning, Reading, Answering, and Checking (C-QRAC) collaboration script on university students' science reading literacy in flipped learning conditions. Participants were 85 Taiwanese university students recruited…
Descriptors: Foreign Countries, College Students, Scripts, Critical Thinking
Cormier, Damien C.; Bulut, Okan; Singh, Deepak; Kennedy, Kathleen E.; Wang, Kun; Heudes, Alethea; Lekwa, Adam J. – Journal of Psychoeducational Assessment, 2018
The selection and interpretation of individually administered norm-referenced cognitive tests that are administered to culturally and linguistically diverse (CLD) students continue to be an important consideration within the psychoeducational assessment process. Understanding test directions during the assessment of cognitive abilities is…
Descriptors: Intelligence Tests, Cognitive Ability, High Stakes Tests, Children