Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 21 |
Descriptor
Difficulty Level | 44 |
Multiple Choice Tests | 44 |
Test Items | 34 |
Test Construction | 17 |
Test Format | 11 |
Reading Comprehension | 10 |
Item Response Theory | 9 |
Foreign Countries | 8 |
Item Analysis | 8 |
Reading Tests | 8 |
Goodness of Fit | 7 |
More ▼ |
Source
Author
Tindal, Gerald | 5 |
Alonzo, Julie | 4 |
Freedle, Roy | 2 |
Kolstad, Rosemarie K. | 2 |
Kostin, Irene | 2 |
Liu, Kimy | 2 |
Park, Bitnara Jasmine | 2 |
Abdullah, Saifuddin Kumar | 1 |
Al-Hamly, Mashael | 1 |
Berg, Kathryn | 1 |
Bielinski, John | 1 |
More ▼ |
Publication Type
Reports - Evaluative | 44 |
Journal Articles | 25 |
Speeches/Meeting Papers | 7 |
Numerical/Quantitative Data | 5 |
Information Analyses | 3 |
Reports - Research | 1 |
Education Level
Audience
Practitioners | 1 |
Teachers | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 2 |
National Assessment of… | 2 |
SAT (College Admission Test) | 2 |
Law School Admission Test | 1 |
Praxis Series | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Loudon, Catherine; Macias-Muñoz, Aide – Advances in Physiology Education, 2018
Different versions of multiple-choice exams were administered to an undergraduate class in human physiology as part of normal testing in the classroom. The goal was to evaluate whether the number of options (possible answers) per question influenced the effectiveness of this assessment. Three exams (each with three versions) were given to each of…
Descriptors: Multiple Choice Tests, Test Construction, Test Items, Science Tests
Travis T. Fuchs; Mike Arsenault – School Science Review, 2017
Students, as well as teachers, often learn what makes sense to them, even when it is wrong. These misconceptions are a problem. The authors sought a quick, quantitative way of identifying student misconceptions in secondary science. Using the University of Toronto's National Biology Competition test data, this article presents a method of quickly…
Descriptors: Science Education, Secondary School Science, Misconceptions, Scientific Concepts
Scully, Darina – Practical Assessment, Research & Evaluation, 2017
Across education, certification and licensure, there are repeated calls for the development of assessments that target "higher-order thinking," as opposed to mere recall of facts. A common assumption is that this necessitates the use of constructed response or essay-style test questions; however, empirical evidence suggests that this may…
Descriptors: Test Construction, Test Items, Multiple Choice Tests, Thinking Skills
Childs, Iraphne R. W.; Berg, Kathryn – Geographical Education, 2015
The Australian Geography Competition (AGC) was established in 1995 by the Royal Geographical Society of Queensland (RGSQ) and the Australian Geography Teachers' Association to promote the study of geography in Australian secondary schools and to reward student excellence in geographical studies. Initially focusing on students at the lower…
Descriptors: Foreign Countries, Geography, Geography Instruction, Rewards
Luebke, Stephen; Lorie, James – Journal of Applied Testing Technology, 2013
This article is a brief account of the use of Bloom's Taxonomy of Educational Objectives (Bloom, Engelhart, Furst, Hill, & Krathwohl, 1956) by staff of the Law School Admission Council in the 1990 development of redesigned specifications for the Reading Comprehension section of the Law School Admission Test. Summary item statistics for the…
Descriptors: Classification, Educational Objectives, Reading Comprehension, Law Schools
Hamzah, Mohd Sahandri Gani; Abdullah, Saifuddin Kumar – Online Submission, 2011
The evaluation of learning is a systematic process involving testing, measuring and evaluation. In the testing step, a teacher needs to choose the best instrument that can test the minds of students. Testing will produce scores or marks with many variations either in homogeneous or heterogeneous forms that will be used to categorize the scores…
Descriptors: Test Items, Item Analysis, Difficulty Level, Testing
Malau-Aduli, Bunmi S.; Zimitat, Craig – Assessment & Evaluation in Higher Education, 2012
The aim of this study was to assess the effect of the introduction of peer review processes on the quality of multiple-choice examinations in the first three years of an Australian medical course. The impact of the peer review process and overall quality assurance (QA) processes were evaluated by comparing the examination data generated in earlier…
Descriptors: Foreign Countries, Peer Evaluation, Multiple Choice Tests, Test Construction
Laprise, Shari L. – College Teaching, 2012
Successful exam composition can be a difficult task. Exams should not only assess student comprehension, but be learning tools in and of themselves. In a biotechnology course delivered to nonmajors at a business college, objective multiple-choice test questions often require students to choose the exception or "not true" choice. Anecdotal student…
Descriptors: Feedback (Response), Test Items, Multiple Choice Tests, Biotechnology
Irvin, P. Shawn; Alonzo, Julie; Lai, Cheng-Fei; Park, Bitnara Jasmine; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the seventh-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Reading Comprehension, Testing Programs, Statistical Analysis, Grade 7
Cawthon, Stephanie – American Annals of the Deaf, 2011
Linguistic complexity of test items is one test format element that has been studied in the context of struggling readers and their participation in paper-and-pencil tests. The present article presents findings from an exploratory study on the potential relationship between linguistic complexity and test performance for deaf readers. A total of 64…
Descriptors: Language Styles, Test Content, Syntax, Linguistics
Kibble, Jonathan D.; Johnson, Teresa – Advances in Physiology Education, 2011
The purpose of this study was to evaluate whether multiple-choice item difficulty could be predicted either by a subjective judgment by the question author or by applying a learning taxonomy to the items. Eight physiology faculty members teaching an upper-level undergraduate human physiology course consented to participate in the study. The…
Descriptors: Test Items, Hidden Curriculum, Reliability, Physiology
Severo, Milton; Tavares, Maria A. Ferreira – Anatomical Sciences Education, 2010
The nature of anatomy education has changed substantially in recent decades, though the traditional multiple-choice written examination remains the cornerstone of assessing students' knowledge. This study sought to measure the quality of a clinical anatomy multiple-choice final examination using item response theory (IRT) models. One hundred…
Descriptors: Evaluation Methods, Anatomy, Item Response Theory, Medical Education
Scafe, Marla G. – American Journal of Business Education, 2011
The purpose of this study was to evaluate the effectiveness of group testing as a pedagogical technique to enhance learning in a difficult subject such as statistics. Individual test scores were compared to their group test scores for the same, identical test. A t test was used to compare the scores for 157 randomly selected MBA students enrolled…
Descriptors: Group Testing, Individual Testing, Statistical Analysis, Comparative Analysis
Park, Bitnara Jasmine; Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2011
This technical report describes the process of development and piloting of reading comprehension measures that are appropriate for seventh-grade students as part of an online progress screening and monitoring assessment system, http://easycbm.com. Each measure consists of an original fictional story of approximately 1,600 to 1,900 words with 20…
Descriptors: Reading Comprehension, Reading Tests, Grade 7, Test Construction
Lingard, Jennifer; Minasian-Batmanian, Laura; Vella, Gilbert; Cathers, Ian; Gonzalez, Carlos – Assessment & Evaluation in Higher Education, 2009
Effective criterion referenced assessment requires grade descriptors to clarify to students what skills are required to gain higher grades. But do students and staff actually have the same perception of the grading system, and if so, do they perform better than those whose perceptions are less accurately aligned with those of staff? Since…
Descriptors: Feedback (Response), Prior Learning, Physics, Difficulty Level