NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1,306 to 1,320 of 3,093 results Save | Export
Peer reviewed Peer reviewed
Hanson, Bradley A. – Applied Measurement in Education, 1996
Determining whether score distributions differ on two or more test forms administered to samples of examinees from a single population is explored using three statistical tests using loglinear models. Examples are presented of applying tests of distribution differences to decide if equating is needed for alternative forms of a test. (SLD)
Descriptors: Equated Scores, Scoring, Statistical Distributions, Test Format
Peer reviewed Peer reviewed
Feldt, Leonard S. – Applied Measurement in Education, 2002
Considers the degree of bias in testlet-based alpha (internal consistency reliability) through hypothetical examples and real test data from four tests of the Iowa Tests of Basic Skills. Presents a simple formula for computing a testlet-based congeneric coefficient. (SLD)
Descriptors: Estimation (Mathematics), Reliability, Statistical Bias, Test Format
Peer reviewed Peer reviewed
Fernald, Peter S.; Webster, Sandra – Journal of Humanistic Education and Development, 1991
Conducted two studies on take-home, closed-book examination. First study involved 20 college students and was designed to provide categorical, comprehensive outline of students' assessments of take-home, closed-book procedure. Second study involved 23 students and compared amount of learning achieved on in-class examination with that on take-home,…
Descriptors: College Students, Higher Education, Student Attitudes, Test Format
Peer reviewed Peer reviewed
Herring, Warren – Journal of Correctional Education, 1999
Analyzes the relationship between success on the two new practice-test forms (EE and FF) developed by Steck-Vaughn and success on the General Educational Development (GED) test. Success with practice-form EE correlated with GED test scores; form FF failed to correlate. (JOW)
Descriptors: Adult Education, Eligibility, Prediction, Scores
Peer reviewed Peer reviewed
van der Linden, Wim J.; Adema, Jos J. – Journal of Educational Measurement, 1998
Proposes an algorithm for the assembly of multiple test forms in which the multiple-form problem is reduced to a series of computationally less intensive two-form problems. Illustrates how the method can be implemented using 0-1 linear programming and gives two examples. (SLD)
Descriptors: Algorithms, Linear Programming, Test Construction, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Afolabi, E. R. I. – Educational Research and Reviews, 2007
The study examined the effects of item format, self-concept and anxiety on response changing behaviour. Four hundred undergraduate students who offered a counseling psychology course in a Nigerian university participated in the study. Students' answers in multiple--choice and true--false formats of an achievement test were observed for response…
Descriptors: Undergraduate Students, Test Items, Self Concept, Multiple Choice Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Mikhailova, Julia – Foreign Language Annals, 2007
This article describes the shortcomings of one of the major testing tools in the foreign languages field, the Simulated Oral Proficiency Interview (SOPI), with regard to the elicitation of the function of description. In doing so, the article raises questions about the applicability of the SOPI as a surrogate for the Oral Proficiency Interview…
Descriptors: Oral Language, Russian, Language Tests, Language Proficiency
Peer reviewed Peer reviewed
Direct linkDirect link
DeMars, Christine E. – Educational Assessment, 2007
A series of 8 tests was administered to university students over 4 weeks for program assessment purposes. The stakes of these tests were low for students; they received course points based on test completion, not test performance. Tests were administered in a counterbalanced order across 2 administrations. Response time effort, a measure of the…
Descriptors: Reaction Time, Guessing (Tests), Testing Programs, College Students
Peer reviewed Peer reviewed
Direct linkDirect link
Frey, Andreas; Hartig, Johannes; Rupp, Andre A. – Educational Measurement: Issues and Practice, 2009
In most large-scale assessments of student achievement, several broad content domains are tested. Because more items are needed to cover the content domains than can be presented in the limited testing time to each individual student, multiple test forms or booklets are utilized to distribute the items to the students. The construction of an…
Descriptors: Measures (Individuals), Test Construction, Theory Practice Relationship, Design
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Seonghoon; Kolen, Michael J. – Applied Measurement in Education, 2006
Four item response theory linking methods (2 moment methods and 2 characteristic curve methods) were compared to concurrent (CO) calibration with the focus on the degree of robustness to format effects (FEs) when applying the methods to multidimensional data that reflected the FEs associated with mixed-format tests. Based on the quantification of…
Descriptors: Item Response Theory, Robustness (Statistics), Test Format, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Singelis, Theodore M.; Yamada, Ann Marie; Barrio, Concepcion; Laney, Joshua Harrison; Her, Pa; Ruiz-Anaya, Alejandrina; Lennertz, Sara Terwilliger – Hispanic Journal of Behavioral Sciences, 2006
The metric equivalence of translated scales is often in question but seldom examined. This study presents test-retest data that support the metric equivalence of the Spanish and English language versions of three measures: the Bidimensional Acculturation Scale, the Satisfaction with Life Scale, and the Self-Construal Scale. Participants were…
Descriptors: Acculturation, Life Satisfaction, English, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Liu, Jinghua; Low, Albert C. – ETS Research Report Series, 2007
This study applied kernel equating (KE) in two scenarios: equating to a very similar population and equating to a very different population, referred to as a distant population, using SAT® data. The KE results were compared to the results obtained from analogous classical equating methods in both scenarios. The results indicate that KE results are…
Descriptors: College Entrance Examinations, Equated Scores, Comparative Analysis, Evaluation Methods
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Haynie, W. J., III – Journal of Technology Education, 2007
Commencing in 1985, a small body of experimental studies on the effects of test taking on delayed retention learning of technical subject matter has been completed in technology education settings. Much of the learning in technology education courses, especially the hands-on aspects, are best assessed via instruments and techniques other than…
Descriptors: Meta Analysis, Retention (Psychology), Technology Education, Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
National Center for Education Statistics, 2007
The purpose of this document is to provide background information that will be useful in interpreting the 2007 results from the Trends in International Mathematics and Science Study (TIMSS) by comparing its design, features, framework, and items with those of the U.S. National Assessment of Educational Progress and another international assessment…
Descriptors: National Competency Tests, Comparative Analysis, Achievement Tests, Test Items
Deborah Elizabeth Fox – ProQuest LLC, 2007
The purpose of this study was the development and testing of a novel method for assessment of white blood cell (WBC) identification skills used in the field of Clinical Laboratory Sciences (CLS). A dual format exam was administered to both novices (students) and experts (laboratory professionals). Format 1 was similar to current assessment…
Descriptors: Evaluation, Evaluation Methods, Health Sciences, Metabolism
Pages: 1  |  ...  |  84  |  85  |  86  |  87  |  88  |  89  |  90  |  91  |  92  |  ...  |  207