NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 2,716 to 2,730 of 9,530 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Carnegie, Jacqueline A. – Canadian Journal for the Scholarship of Teaching and Learning, 2017
Summative evaluation for large classes of first- and second-year undergraduate courses often involves the use of multiple choice question (MCQ) exams in order to provide timely feedback. Several versions of those exams are often prepared via computer-based question scrambling in an effort to deter cheating. An important parameter to consider when…
Descriptors: Undergraduate Students, Student Evaluation, Multiple Choice Tests, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Dahlke, Katie; Yang, Rui; Martínez, Carmen; Chavez, Suzette; Martin, Alejandra; Hawkinson, Laura; Shields, Joseph; Garland, Marshall; Carle, Jill – Regional Educational Laboratory Southwest, 2017
The New Mexico Public Education Department developed the Kindergarten Observation Tool (KOT) as a multidimensional observational measure of students' knowledge and skills at kindergarten entry. The primary purpose of the KOT is to inform instruction, so that kindergarten teachers can use the information about their students' knowledge and skills…
Descriptors: Test Validity, Observation, Measures (Individuals), Kindergarten
Schoen, Robert C.; Yang, Xiaotong; Liu, Sicong; Paek, Insu – Grantee Submission, 2017
The Early Fractions Test v2.2 is a paper-pencil test designed to measure mathematics achievement of third- and fourth-grade students in the domain of fractions. The purpose, or intended use, of the Early Fractions Test v2.2 is to serve as a measure of student outcomes in a randomized trial designed to estimate the effect of an educational…
Descriptors: Psychometrics, Mathematics Tests, Mathematics Achievement, Fractions
Peer reviewed Peer reviewed
Direct linkDirect link
Hitchcock, John H.; Johanson, George A. – Research in the Schools, 2015
Understanding the reason(s) for Differential Item Functioning (DIF) in the context of measurement is difficult. Although identifying potential DIF items is typically a statistical endeavor, understanding the reasons for DIF (and item repair or replacement) might require investigations that can be informed by qualitative work. Such work is…
Descriptors: Mixed Methods Research, Test Items, Item Analysis, Measurement
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Pichette, François; Béland, Sébastien; Jolani, Shahab; Lesniewska, Justyna – Studies in Second Language Learning and Teaching, 2015
Researchers are frequently confronted with unanswered questions or items on their questionnaires and tests, due to factors such as item difficulty, lack of testing time, or participant distraction. This paper first presents results from a poll confirming previous claims (Rietveld & van Hout, 2006; Schafer & Graham, 2002) that data…
Descriptors: Language Research, Data Analysis, Simulation, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Janke, Vikki; Kolokonte, Marina – Second Language Research, 2015
Three profoundly deaf individuals undertook a low-frequency backward lexical translation task (French/English), where morphological structure was manipulated and orthographic distance between test items was measured. Conditions included monomorphemic items (simplex), polymorphemic items (complex), items whose French morphological structure…
Descriptors: Deafness, Interference (Language), Morphology (Languages), Phonology
Peer reviewed Peer reviewed
Direct linkDirect link
Oshima, T. C.; Wright, Keith; White, Nick – International Journal of Testing, 2015
Raju, van der Linden, and Fleer (1995) introduced a framework for differential functioning of items and tests (DFIT) for unidimensional dichotomous models. Since then, DFIT has been shown to be a quite versatile framework as it can handle polytomous as well as multidimensional models both at the item and test levels. However, DFIT is still limited…
Descriptors: Test Bias, Item Response Theory, Test Items, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Kannan, Priya; Sgammato, Adrienne; Tannenbaum, Richard J. – Applied Measurement in Education, 2015
Establishing cut scores using the Angoff method requires panelists to evaluate every item on a test and make a probability judgment. This can be time-consuming when there are large numbers of items on the test. Previous research using resampling studies suggest that it is possible to recommend stable Angoff-based cut score estimates using a…
Descriptors: Cutting Scores, Test Items, Standard Setting (Scoring), Feasibility Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Hardré, Patricia L.; Hackett, Shannon – Educational Assessment, Evaluation and Accountability, 2015
This manuscript chronicles the process and products of a redesign for evaluation of the graduate college experience (GCE) which was initiated by a university graduate college, based on its observed need to reconsider and update its measures and methods for assessing graduate students' experiences. We examined the existing instrumentation and…
Descriptors: Test Construction, Graduate Students, Student Experience, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Brennan, Robert L. – Measurement: Interdisciplinary Research and Perspectives, 2015
Koretz, in his article published in this issue, provides compelling arguments that the high stakes currently associated with accountability testing lead to behavioral changes in students, teachers, and other stakeholders that often have negative consequences, such as inflated scores. Koretz goes on to argue that these negative consequences require…
Descriptors: Accountability, High Stakes Tests, Behavior Change, Student Behavior
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Attali, Yigal; Saldivia, Luis; Jackson, Carol; Schuppan, Fred; Wanamaker, Wilbur – ETS Research Report Series, 2014
Previous investigations of the ability of content experts and test developers to estimate item difficulty have, for themost part, produced disappointing results. These investigations were based on a noncomparative method of independently rating the difficulty of items. In this article, we argue that, by eliciting comparative judgments of…
Descriptors: Test Items, Difficulty Level, Comparative Analysis, College Entrance Examinations
Peer reviewed Peer reviewed
Direct linkDirect link
Dickson, Chata A.; MacDonald, Rebecca P. F.; Mansfield, Renee; Guilhardi, Paulo; Johnson, Cammarie; Ahearn, William H. – Journal of Autism and Developmental Disorders, 2014
We investigated the social validity of the NECC Core Skills Assessment (NECC-CSA) with parents and professionals as participants. The NECC-CSA is a measurement tool consisting of direct and indirect measures of skills important to all individuals with autism, across the lifespan. Participants (N = 245) were provided with a list of 66 skills, 47 of…
Descriptors: Test Validity, Basic Skills, Measures (Individuals), Autism
Peer reviewed Peer reviewed
Direct linkDirect link
Scholz, Michael A.; Huynh, Niem Tu; Brysch, Carmen P.; Scholz, Ruojing Wang – Journal of Geography, 2014
Geography textbooks contain chapter or review questions that may engage students in spatial thinking. This research used Jo and Bednarz's (2009) "Taxonomy of Spatial Thinking" to evaluate the percentage of spatial thinking questions in four university-level world geography course textbooks. The results from this study were then…
Descriptors: World Geography, Textbooks, Spatial Ability, Content Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Roivainen, Eka – Journal of Psychoeducational Assessment, 2014
Research on secular trends in mean intelligence test scores shows smaller gains in vocabulary skills than in nonverbal reasoning. One possible explanation is that vocabulary test items become outdated faster compared to nonverbal tasks. The history of the usage frequency of the words on five popular vocabulary tests, the GSS Wordsum, Wechsler…
Descriptors: Vocabulary Skills, Word Frequency, Language Usage, Change
Peer reviewed Peer reviewed
Direct linkDirect link
Antal, Judit; Proctor, Thomas P.; Melican, Gerald J. – Applied Measurement in Education, 2014
In common-item equating the anchor block is generally built to represent a miniature form of the total test in terms of content and statistical specifications. The statistical properties frequently reflect equal mean and spread of item difficulty. Sinharay and Holland (2007) suggested that the requirement for equal spread of difficulty may be too…
Descriptors: Test Items, Equated Scores, Difficulty Level, Item Response Theory
Pages: 1  |  ...  |  178  |  179  |  180  |  181  |  182  |  183  |  184  |  185  |  186  |  ...  |  636