NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 16 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Rao, Dhawaleswar; Saha, Sujan Kumar – IEEE Transactions on Learning Technologies, 2020
Automatic multiple choice question (MCQ) generation from a text is a popular research area. MCQs are widely accepted for large-scale assessment in various domains and applications. However, manual generation of MCQs is expensive and time-consuming. Therefore, researchers have been attracted toward automatic MCQ generation since the late 90's.…
Descriptors: Multiple Choice Tests, Test Construction, Automation, Computer Software
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Laliyo, Lukman Abdul Rauf; Hamdi, Syukrul; Pikoli, Masrid; Abdullah, Romario; Panigoro, Citra – European Journal of Educational Research, 2021
One of the issues that hinder the students' learning progress is the inability to construct an epistemological explanation of a scientific phenomenon. Four-tier multiple-choice (hereinafter, 4TMC) instrument and Partial-Credit Model were employed to elaborate on the diagnosis process of the aforementioned problem. This study was to develop and…
Descriptors: Learning Processes, Multiple Choice Tests, Models, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Hamdi, Syukrul; Kartowagiran, Badrun; Haryanto – International Journal of Instruction, 2018
The purpose of this study was to develop a Mathematics test instrument testlet model for a classroom assessment at elementary school. Testlet Model is a group of multiple choice question acquiring similar information with different grade of responses model. This research was conducted in East Lombok, Indonesia. The design used was research…
Descriptors: Test Items, Models, Elementary School Mathematics, Mathematics Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Ihme, Jan Marten; Senkbeil, Martin; Goldhammer, Frank; Gerick, Julia – European Educational Research Journal, 2017
The combination of different item formats is found quite often in large scale assessments, and analyses on the dimensionality often indicate multi-dimensionality of tests regarding the task format. In ICILS 2013, three different item types (information-based response tasks, simulation tasks, and authoring tasks) were used to measure computer and…
Descriptors: Foreign Countries, Computer Literacy, Information Literacy, International Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Hee-Sun; Liu, Ou Lydia; Linn, Marcia C. – Applied Measurement in Education, 2011
This study explores measurement of a construct called knowledge integration in science using multiple-choice and explanation items. We use construct and instructional validity evidence to examine the role multiple-choice and explanation items plays in measuring students' knowledge integration ability. For construct validity, we analyze item…
Descriptors: Knowledge Level, Construct Validity, Validity, Scaffolding (Teaching Technique)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Baghaei, Purya; Carstensen, Claus H. – Practical Assessment, Research & Evaluation, 2013
Standard unidimensional Rasch models assume that persons with the same ability parameters are comparable. That is, the same interpretation applies to persons with identical ability estimates as regards the underlying mental processes triggered by the test. However, research in cognitive psychology shows that persons at the same trait level may…
Descriptors: Item Response Theory, Models, Reading Comprehension, Reading Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Karim, Aidah Abdul; Shah, Parilah M.; Din, Rosseni; Ahmad, Mazalah; Lubis, Maimun Aqhsa – International Education Studies, 2014
This study explored the psychometric properties of a locally developed information skills test for youth students in Malaysia using Rasch analysis. The test was a combination of 24 structured and multiple choice items with a 4-point grading scale. The test was administered to 72 technical college students and 139 secondary school students. The…
Descriptors: Foreign Countries, Information Skills, Item Response Theory, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Cloonan, Carrie A.; Hutchinson, John S. – Chemistry Education Research and Practice, 2011
A Chemistry Concept Reasoning Test was created and validated providing an easy-to-use tool for measuring conceptual understanding and critical scientific thinking of general chemistry models and theories. The test is designed to measure concept understanding comparable to that found in free-response questions requiring explanations over…
Descriptors: Test Validity, Chemistry, Correlation, Multiple Choice Tests
Way, Walter D.; Murphy, Daniel; Powers, Sonya; Keng, Leslie – Pearson, 2012
Significant momentum exists for next-generation assessments to increasingly utilize technology to develop and deliver performance-based assessments. Many traditional challenges with this assessment approach still apply, including psychometric concerns related to performance-based tasks (PBTs), which include low reliability, efficiency of…
Descriptors: Task Analysis, Performance Based Assessment, Technology Uses in Education, Models
Kobrin, Jennifer L.; Kim, Rachel; Sackett, Paul – College Board, 2011
There is much debate on the merits and pitfalls of standardized tests for college admission, with questions regarding the format (multiple-choice versus constructed response), cognitive complexity, and content of these assessments (achievement versus aptitude) at the forefront of the discussion. This study addressed these questions by…
Descriptors: College Entrance Examinations, Mathematics Tests, Test Items, Predictive Validity
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Graf, Edith Aurora – ETS Research Report Series, 2008
Quantitative item models are item structures that may be expressed in terms of mathematical variables and constraints. An item model may be developed as a computer program from which large numbers of items are automatically generated. Item models can be used to produce large numbers of items for use in traditional, large-scale assessments. But…
Descriptors: Test Items, Models, Diagnostic Tests, Statistical Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Haberman, Shelby J. – ETS Research Report Series, 2007
In item-response theory, if a latent-structure model has an ability variable, then elementary information theory may be employed to provide a criterion for evaluation of the information the test provides concerning ability. This criterion may be considered even in cases in which the latent-structure model is not valid, although interpretation of…
Descriptors: Item Response Theory, Ability, Information Theory, Computation
Bennett, Randy Elliot; And Others – 1990
A framework for categorizing constructed-response items was developed in which items were ordered on a continuum from multiple-choice to presentation/performance according to the degree of constraint placed on the examinee's response. Two investigations were carried out to evaluate the validity of this framework. In the first investigation, 27…
Descriptors: Classification, Constructed Response, Models, Multiple Choice Tests
Sugrue, Brenda – 1993
This report describes a methodology for increasing the validity and reliability of inferences made about the problem-solving ability of science students that is based on performance on different kinds of tests. The generalizable cognitive components of problem solving that might be targeted by assessment are described, and specifications are…
Descriptors: Chemistry, Educational Assessment, Inferences, Models
Quellmalz, Edys; Capell, Frank – 1979
The purpose of this study was to examine the stability of measures of student writing performance across types of discourse (genres) and across response modes (selected response: multiple choice; constructed response: single paragraph, and full length essay). The study addressed the following: (1) the relationship/stability of writing scores…
Descriptors: Correlation, Essay Tests, Literary Genres, Models
Previous Page | Next Page ยป
Pages: 1  |  2