Publication Date
In 2025 | 0 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 7 |
Descriptor
Comparative Testing | 46 |
Difficulty Level | 46 |
Test Items | 26 |
Higher Education | 21 |
Multiple Choice Tests | 15 |
Test Construction | 13 |
Computer Assisted Testing | 12 |
Test Format | 11 |
Test Reliability | 11 |
College Students | 9 |
Scores | 9 |
More ▼ |
Source
Author
Publication Type
Reports - Research | 46 |
Journal Articles | 19 |
Speeches/Meeting Papers | 17 |
Tests/Questionnaires | 4 |
Numerical/Quantitative Data | 1 |
Opinion Papers | 1 |
Reports - Evaluative | 1 |
Education Level
Higher Education | 3 |
Postsecondary Education | 3 |
Adult Education | 1 |
Elementary Education | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
Grade 7 | 1 |
Grade 8 | 1 |
Audience
Researchers | 2 |
Laws, Policies, & Programs
Elementary and Secondary… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Javier Del Olmo-Muñoz; Pascual D. Diago; David Arnau; David Arnau-Blasco; José Antonio González-Calero – ZDM: Mathematics Education, 2024
This research, following a sequential mixed-methods design, delves into metacognitive control in problem solving among 5- to 6-year-olds, using two floor-robot environments. In an initial qualitative phase, 82 pupils participated in tasks in which they directed a floor robot to one of two targets, with the closer target requiring more cognitive…
Descriptors: Elementary School Students, Metacognition, Robotics, Computer Simulation
Agus Santoso; Heri Retnawati; Timbul Pardede; Ibnu Rafi; Munaya Nikma Rosyada; Gulzhaina K. Kassymova; Xu Wenxin – Practical Assessment, Research & Evaluation, 2024
The test blueprint is important in test development, where it guides the test item writer in creating test items according to the desired objectives and specifications or characteristics (so-called a priori item characteristics), such as the level of item difficulty in the category and the distribution of items based on their difficulty level.…
Descriptors: Foreign Countries, Undergraduate Students, Business English, Test Construction
Lahner, Felicitas-Maria; Lörwald, Andrea Carolin; Bauer, Daniel; Nouns, Zineb Miriam; Krebs, René; Guttormsen, Sissel; Fischer, Martin R.; Huwendiek, Sören – Advances in Health Sciences Education, 2018
Multiple true-false (MTF) items are a widely used supplement to the commonly used single-best answer (Type A) multiple choice format. However, an optimal scoring algorithm for MTF items has not yet been established, as existing studies yielded conflicting results. Therefore, this study analyzes two questions: What is the optimal scoring algorithm…
Descriptors: Scoring Formulas, Scoring Rubrics, Objective Tests, Multiple Choice Tests
Li, Hongli; Suen, Hoi K. – International Multilingual Research Journal, 2015
This study examines how Chinese ESL learners recognize English words while responding to a multiple-choice reading test as compared to Romance-language-speaking ESL learners. Four adult Chinese ESL learners and three adult Romance-language-speaking ESL learners participated in a think-aloud study with the Michigan English Language Assessment…
Descriptors: Chinese, English (Second Language), English Language Learners, Romance Languages
Wall, Jeffrey D.; Knapp, Janice – Journal of Information Systems Education, 2014
Learning technical computing skills is increasingly important in our technology driven society. However, learning technical skills in information systems (IS) courses can be difficult. More than 20 percent of students in some technical courses may dropout or fail. Unfortunately, little is known about students' perceptions of the difficulty of…
Descriptors: Undergraduate Students, Information Systems, Grounded Theory, Statistical Analysis
Wang, Jianjun – School Science and Mathematics, 2011
As the largest international study ever taken in history, the Trend in Mathematics and Science Study (TIMSS) has been held as a benchmark to measure U.S. student performance in the global context. In-depth analyses of the TIMSS project are conducted in this study to examine key issues of the comparative investigation: (1) item flaws in mathematics…
Descriptors: Test Items, Figurative Language, Item Response Theory, Benchmarking
Ward, Chris; Yates, Dan; Song, Joon – American Journal of Business Education, 2009
This study examined the extent to which student engagement is associated with a traditional assessment of student knowledge. In this study, ETS Business Major Field Test (MFT) scores were compared to student's self-reported survey responses to specific questions on the National Survey of Student Engagement (NSSE). Areas of the NSSE survey such as…
Descriptors: Pilot Projects, Learner Engagement, Business, Business Skills
Frisbie, David A. – 1981
The relative difficulty ratio (RDR) is used as a method of representing test difficulty. The RDR is the ratio of a test mean to the ideal mean, the point midway between the perfect score and the mean chance score for the test. The RDR tranformation is a linear scale conversion method but not a linear equating method in the classical sense. The…
Descriptors: Comparative Testing, Difficulty Level, Evaluation Methods, Raw Scores
Ebel, Robert L. – 1981
An alternate-choice test item is a simple declarative sentence, one portion of which is given with two different wordings. For example, "Foundations like Ford and Carnegie tend to be (1) eager (2) hesitant to support innovative solutions to educational problems." The examinee's task is to choose the alternative that makes the sentence…
Descriptors: Comparative Testing, Difficulty Level, Guessing (Tests), Multiple Choice Tests

Rocklin, Thomas; O'Donnell, Angela M. – Journal of Educational Psychology, 1987
An experiment was conducted that contrasted a variant of computerized adaptive testing, self-adapted testing, with two traditional tests. Participants completed a self-report of text anxiety and were randomly assigned to take one of the three tests of verbal ability. Subjects generally chose more difficult items as the test progressed. (Author/LMO)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Difficulty Level

Wahlstrom, Merlin; And Others – Canadian Journal of Education, 1986
An important aspect of Ontario's participation in the Second International Study of Mathematics was a comparative analysis of students' mathematics achievement from 1968 to 1982. Achievement levels remained remarkably constant. The problem of declining achievement in the United States was not apparent in this analysis of Ontario students. (LMO)
Descriptors: Achievement Tests, Comparative Testing, Difficulty Level, Foreign Countries

Frary, Robert B. – Applied Measurement in Education, 1991
The use of the "none-of-the-above" option (NOTA) in 20 college-level multiple-choice tests was evaluated for classes with 100 or more students. Eight academic disciplines were represented, and 295 NOTA and 724 regular test items were used. It appears that the NOTA can be compatible with good classroom measurement. (TJH)
Descriptors: College Students, Comparative Testing, Difficulty Level, Discriminant Analysis

Crehan, Kevin D.; And Others – Educational and Psychological Measurement, 1993
Studies with 220 college students found that multiple-choice test items with 3 items are more difficult than those with 4 items, and items with the none-of-these option are more difficult than those without this option. Neither format manipulation affected item discrimination. Implications for test construction are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Difficulty Level, Distractors (Tests)

Anderson, Paul S.; Hyers, Albert D. – 1991
Three descriptive statistics (difficulty, discrimination, and reliability) of multiple-choice (MC) test items were compared to those of a new (1980s) format of machine-scored questions. The new method, answer-bank multi-digit testing (MDT), uses alphabetized lists of up to 1,000 alternatives and approximates the completion style of assessment…
Descriptors: College Students, Comparative Testing, Computer Assisted Testing, Correlation
Cizek, Gregory J. – 1991
A commonly accepted rule for developing equated examinations using the common-items non-equivalent groups (CINEG) design is that items common to the two examinations being equated should be identical. The CINEG design calls for two groups of examinees to respond to a set of common items that is included in two examinations. In practice, this rule…
Descriptors: Certification, Comparative Testing, Difficulty Level, Higher Education