NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 2,851 to 2,865 of 9,530 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Bramley, Tom – Research in Mathematics Education, 2017
This study compared models of assessment structure for achieving differentiation across the range of examinee attainment in the General Certificate of Secondary Education (GCSE) examination taken by 16-year-olds in England. The focus was on the "adjacent levels" model, where papers are targeted at three specific non-overlapping ranges of…
Descriptors: Foreign Countries, Mathematics Education, Student Certification, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Moshinsky, Avital; Ziegler, David; Gafni, Naomi – International Journal of Testing, 2017
Many medical schools have adopted multiple mini-interviews (MMI) as an advanced selection tool. MMIs are expensive and used to test only a few dozen candidates per day, making it infeasible to develop a different test version for each test administration. Therefore, some items are reused both within and across years. This study investigated the…
Descriptors: Interviews, Medical Schools, Test Validity, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Traynor, Anne – Applied Measurement in Education, 2017
It has long been argued that U.S. states' differential performance on nationwide assessments may reflect differences in students' opportunity to learn the tested content that is primarily due to variation in curricular content standards, rather than in instructional quality or educational investment. To quantify the effect of differences in…
Descriptors: Test Items, Difficulty Level, State Standards, Academic Standards
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Al-Habashneh, Maher Hussein; Najjar, Nabil Juma – Journal of Education and Practice, 2017
This study aimed at constructing a criterion-reference test to measure the research and statistical competencies of graduate students at the Jordanian governmental universities, the test has to be in its first form of (50) multiple choice items, then the test was introduced to (5) arbitrators with competence in measurement and evaluation to…
Descriptors: Foreign Countries, Criterion Referenced Tests, Graduate Students, Test Construction
Chen, Chieh-Yu – ProQuest LLC, 2017
Investigating the psychometric properties of a screening instrument for young children is necessary to ascertain its quality and accuracy. In light of the important role culture plays on human beliefs and parenting styles, a newly translated and adapted test needs to be studied. Evaluating outcomes on a translated version of a test may reveal…
Descriptors: Psychometrics, Screening Tests, Questionnaires, Social Development
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Bokyoung Park – English Teaching, 2017
This study investigated Korean college students' performance as measured by two different vocabulary assessment tools (the Productive Vocabulary Levels Test (PVLT) and the Productive Vocabulary Use Task (PVUT)) and the relationship these assessments have with students' writing proficiency. A total of 72 students participated in the study. The…
Descriptors: Foreign Countries, Vocabulary Development, Language Tests, Second Language Learning
Peer reviewed Peer reviewed
Direct linkDirect link
Bolt, Daniel M.; Wollack, James A.; Suh, Youngsuk – Psychometrika, 2012
Nested logit models have been presented as an alternative to multinomial logistic models for multiple-choice test items (Suh and Bolt in "Psychometrika" 75:454-473, 2010) and possess a mathematical structure that naturally lends itself to evaluating the incremental information provided by attending to distractor selection in scoring. One potential…
Descriptors: Test Items, Multiple Choice Tests, Models, Scoring
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zwick, Rebecca; Ye, Lei; Isham, Steven – ETS Research Report Series, 2013
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. Although it is often assumed that refinement of the matching criterion always provides more accurate DIF results, the actual situation proves to be more complex. To explore the effectiveness of refinement, we…
Descriptors: Test Bias, Statistical Analysis, Simulation, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Moses, Tim – International Journal of Testing, 2013
The major purpose of this study is to assess the conditions under which single scoring for constructed-response (CR) items is as effective as double scoring in the licensure testing context. We used both empirical datasets of five mixed-format licensure tests collected in actual operational settings and simulated datasets that allowed for the…
Descriptors: Scoring, Test Format, Licensing Examinations (Professions), Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Yasuda, Jun-ichiro; Taniguchi, Masa-aki – Physical Review Special Topics - Physics Education Research, 2013
In this study, we evaluate the structural validity of Q.16 and Q.7 in the Force Concept Inventory (FCI). We address whether respondents who answer Q.16 and Q.7 correctly actually have an understanding of the concepts of physics tested in the questions. To examine respondents' levels of understanding, we use subquestions that test them on concepts…
Descriptors: Physics, Scientific Concepts, Science Tests, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Patton, Jeffrey M.; Cheng, Ying; Yuan, Ke-Hai; Diao, Qi – Applied Psychological Measurement, 2013
Variable-length computerized adaptive testing (VL-CAT) allows both items and test length to be "tailored" to examinees, thereby achieving the measurement goal (e.g., scoring precision or classification) with as few items as possible. Several popular test termination rules depend on the standard error of the ability estimate, which in turn depends…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Length, Ability
Peer reviewed Peer reviewed
Direct linkDirect link
San Martin, Ernesto; Rolin, Jean-Marie; Castro, Luis M. – Psychometrika, 2013
In this paper, we study the identification of a particular case of the 3PL model, namely when the discrimination parameters are all constant and equal to 1. We term this model, 1PL-G model. The identification analysis is performed under three different specifications. The first specification considers the abilities as unknown parameters. It is…
Descriptors: Item Response Theory, Models, Identification, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Gierl, Mark J.; Lai, Hollis; Li, Johnson – Educational Research and Evaluation, 2013
The purpose of this study is to evaluate the performance of CATSIB (Computer Adaptive Testing-Simultaneous Item Bias Test) for detecting differential item functioning (DIF) when items in the matching and studied subtest are administered adaptively in the context of a realistic multi-stage adaptive test (MST). MST was simulated using a 4-item…
Descriptors: Adaptive Testing, Test Bias, Computer Assisted Testing, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Teker, Gulsen Tasdelen; Dogan, Nuri – Educational Sciences: Theory and Practice, 2015
Reliability and differential item functioning (DIF) analyses were conducted on testlets displaying local item dependence in this study. The data set employed in the research was obtained from the answers given by 1,500 students to the 20 items included in six testlets given in English Proficiency Exam by the School of Foreign Languages of a state…
Descriptors: Foreign Countries, Test Items, Test Bias, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Marie, S. Maria Josephine Arokia; Edannur, Sreekala – Journal of Educational Technology, 2015
This paper focused on the analysis of test items constructed in the paper of teaching Physical Science for B.Ed. class. It involved the analysis of difficulty level and discrimination power of each test item. Item analysis allows selecting or omitting items from the test, but more importantly item analysis is a tool to help the item writer improve…
Descriptors: Item Analysis, Relevance (Education), Standardized Tests, Achievement Tests
Pages: 1  |  ...  |  187  |  188  |  189  |  190  |  191  |  192  |  193  |  194  |  195  |  ...  |  636