NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1,531 to 1,545 of 9,533 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A.; Dimitrov, Dimiter M.; Li, Tatyana – Educational and Psychological Measurement, 2018
This article extends the procedure outlined in the article by Raykov, Marcoulides, and Tong for testing congruence of latent constructs to the setting of binary items and clustering effects. In this widely used setting in contemporary educational and psychological research, the method can be used to examine if two or more homogeneous…
Descriptors: Tests, Psychometrics, Test Items, Construct Validity
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Metsämuuronen, Jari – International Electronic Journal of Mathematics Education, 2018
The article discusses the challenges of and solutions of developing a common, general standards referenced student assessment framework for mathematics. Two main challenges are faced. First, the main challenge is the lack of commonly accepted standards as the basis for the criterion- and standards-referenced assessment. Second challenge is that,…
Descriptors: Academic Standards, Mathematics Skills, Evaluation Criteria, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Gangur, Mikuláš; Plevny, Miroslav – Journal on Efficiency and Responsibility in Education and Science, 2018
The paper presents a possible way of solving the problem of creating more test variants for a large number of students divided into groups. The proposed solution may consist in introducing a parameterized automatic test generator. The principle of an automatic parameterized test generator is shown. The process of the question tree construction…
Descriptors: Computer Assisted Testing, Test Construction, Test Items, Heuristics
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Educational Measurement: Issues and Practice, 2018
The choice of anchor tests is crucial in applications of the nonequivalent groups with anchor test design of equating. Sinharay and Holland (2006, 2007) suggested "miditests," which are anchor tests that are content-representative and have the same mean item difficulty as the total test but have a smaller spread of item difficulties.…
Descriptors: Test Content, Difficulty Level, Test Items, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Kam, Chester Chun Seng – Sociological Methods & Research, 2018
The item wording (or keying) effect is respondents' differential response style to positively and negatively worded items. Despite decades of research, the nature of the effect is still unclear. This article proposes a potential reason; namely, that the item wording effect is scale-specific, and thus findings are applicable only to a particular…
Descriptors: Response Style (Tests), Test Items, Language Usage, College Students
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Howard, Matt C. – Practical Assessment, Research & Evaluation, 2018
Scale pretests analyze the suitability of individual scale items for further analysis, whether through judging their face validity, wording concerns, and/or other aspects. The current article reviews scale pretests, separated by qualitative and quantitative methods, in order to identify the differences, similarities, and even existence of the…
Descriptors: Pretesting, Measures (Individuals), Test Items, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
International Journal of Testing, 2018
The second edition of the International Test Commission Guidelines for Translating and Adapting Tests was prepared between 2005 and 2015 to improve upon the first edition, and to respond to advances in testing technology and practices. The 18 guidelines are organized into six categories to facilitate their use: pre-condition (3), test development…
Descriptors: Translation, Test Construction, Testing, Scoring
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sari, Halil Ibrahim; Karaman, Mehmet Akif – International Journal of Assessment Tools in Education, 2018
The current study shows the applications of both classical test theory (CTT) and item response theory (IRT) to the psychology data. The study discusses item level analyses of General Mattering Scale produced by the two theories as well as strengths and weaknesses of both measurement approaches. The survey consisted of a total of five Likert-type…
Descriptors: Measures (Individuals), Test Theory, Item Response Theory, Likert Scales
Peer reviewed Peer reviewed
Direct linkDirect link
Becker, Anthony; Nekrasova-Beker, Tatiana – Educational Assessment, 2018
While previous research has identified numerous factors that contribute to item difficulty, studies involving large-scale reading tests have provided mixed results. This study examined five selected-response item types used to measure reading comprehension in the Pearson Test of English Academic: a) multiple-choice (choose one answer), b)…
Descriptors: Reading Comprehension, Test Items, Reading Tests, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Rubright, Jonathan D. – Educational Measurement: Issues and Practice, 2018
Performance assessments, scenario-based tasks, and other groups of items carry a risk of violating the local item independence assumption made by unidimensional item response theory (IRT) models. Previous studies have identified negative impacts of ignoring such violations, most notably inflated reliability estimates. Still, the influence of this…
Descriptors: Performance Based Assessment, Item Response Theory, Models, Test Reliability
Tornabene, Robyn – ProQuest LLC, 2018
Genetics is universally recognized as a core aspect of biological and scientific literacy. Beyond genetics' own role as a major unifying topic within the biological sciences, understanding genetics is essential for understanding other integral ideas such as evolution and development. Genetics understanding also underlies public decision making…
Descriptors: Item Response Theory, Biology, Undergraduate Students, Majors (Students)
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Kyung Yong; Lee, Won-Chan – Applied Measurement in Education, 2017
This article provides a detailed description of three factors (specification of the ability distribution, numerical integration, and frame of reference for the item parameter estimates) that might affect the item parameter estimation of the three-parameter logistic model, and compares five item calibration methods, which are combinations of the…
Descriptors: Test Items, Item Response Theory, Comparative Analysis, Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Köhler, Carmen; Pohl, Steffi; Carstensen, Claus H. – Journal of Educational Measurement, 2017
Competence data from low-stakes educational large-scale assessment studies allow for evaluating relationships between competencies and other variables. The impact of item-level nonresponse has not been investigated with regard to statistics that determine the size of these relationships (e.g., correlations, regression coefficients). Classical…
Descriptors: Test Items, Cognitive Measurement, Testing Problems, Regression (Statistics)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Bayaydah, Areen Mohammad; Altwissi, Ahmad Issa – International Online Journal of Primary Education, 2020
This study aimed to identify and analyze the patterns of final exam questions prepared by English teachers for the 9th and 10th grades and to analyze all the revision questions presented in the English language textbooks in Jordan, based on Bloom's taxonomy to determine the nature and types of these questions. The sample of the study consisted of…
Descriptors: Taxonomy, Textbook Content, Content Analysis, Language Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chen, Michelle Y.; Flasko, Jennifer J. – Canadian Journal of Applied Linguistics / Revue canadienne de linguistique appliquée, 2020
Seeking evidence to support content validity is essential to test validation. This is especially the case in contexts where test scores are interpreted in relation to external proficiency standards and where new test content is constantly being produced to meet test administration and security demands. In this paper, we describe a modified…
Descriptors: Foreign Countries, Reading Tests, Language Tests, English (Second Language)
Pages: 1  |  ...  |  99  |  100  |  101  |  102  |  103  |  104  |  105  |  106  |  107  |  ...  |  636