Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 4 |
Descriptor
| Comparative Analysis | 24 |
| Testing Problems | 24 |
| Scoring | 18 |
| Test Construction | 7 |
| Guessing (Tests) | 6 |
| Higher Education | 6 |
| Computer Assisted Testing | 5 |
| Educational Assessment | 5 |
| Multiple Choice Tests | 5 |
| Test Format | 5 |
| Test Interpretation | 5 |
| More ▼ | |
Source
Author
Publication Type
Education Level
| Higher Education | 2 |
| Postsecondary Education | 2 |
| Secondary Education | 1 |
Audience
| Practitioners | 1 |
| Teachers | 1 |
Location
| Iran | 1 |
| Netherlands | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 2 |
| Comprehensive Tests of Basic… | 1 |
| Graduate Management Admission… | 1 |
| Michigan Test of English… | 1 |
| Program for International… | 1 |
| SAT (College Admission Test) | 1 |
| Wechsler Preschool and… | 1 |
What Works Clearinghouse Rating
Emadian, Farzaneh; Gholami, Javad; Sarkhosh, Mehdi – Journal of Teacher Education for Sustainability, 2018
The first and most crucial step towards developing a sustainable curriculum for instructors teaching English for Specific Academic Purposes (ESAP) is a needs analysis. Therefore, the main aim of conducting this study was to investigate the in-service needs of language instructors and content specialists teaching ESAP and to spot the differences…
Descriptors: English for Academic Purposes, Second Language Learning, Second Language Instruction, Inservice Teacher Education
Chen, Haiwen H.; von Davier, Matthias; Yamamoto, Kentaro; Kong, Nan – ETS Research Report Series, 2015
One major issue with large-scale assessments is that the respondents might give no responses to many items, resulting in less accurate estimations of both assessed abilities and item parameters. This report studies how the types of items affect the item-level nonresponse rates and how different methods of treating item-level nonresponses have an…
Descriptors: Achievement Tests, Foreign Countries, International Assessment, Secondary School Students
Ventouras, Errikos; Triantis, Dimos; Tsiakas, Panagiotis; Stergiopoulos, Charalampos – Computers & Education, 2011
The aim of the present research was to compare the use of multiple-choice questions (MCQs) as an examination method against the oral examination (OE) method. MCQs are widely used and their importance seems likely to grow, due to their inherent suitability for electronic assessment. However, MCQs are influenced by the tendency of examinees to guess…
Descriptors: Grades (Scholastic), Scoring, Multiple Choice Tests, Test Format
Ventouras, Errikos; Triantis, Dimos; Tsiakas, Panagiotis; Stergiopoulos, Charalampos – Computers & Education, 2010
The aim of the present research was to compare the use of multiple-choice questions (MCQs) as an examination method, to the examination based on constructed-response questions (CRQs). Despite that MCQs have an advantage concerning objectivity in the grading process and speed in production of results, they also introduce an error in the final…
Descriptors: Computer Assisted Instruction, Scoring, Grading, Comparative Analysis
Larkin, Kevin C.; Weiss, David J. – 1975
A 15-stage pyramidal test and a 40-item two-stage test were constructed and administered by computer to 111 college undergraduates. The two-stage test was found to utilize a smaller proportion of its potential score range than the pyramidal test. Score distributions for both tests were positively skewed but not significantly different from the…
Descriptors: Ability, Aptitude Tests, Comparative Analysis, Computer Programs
Peer reviewedvan der Linden, Wim J. – Review of Educational Research, 1981
Using criterion-referenced test item data collected in an empirical study, differences in item selection between Cox and Vargas' pretest-posttest validity index and a latent trait approach (evaluation of the item information function for the mastery score) are analyzed. (Author/GK)
Descriptors: Comparative Analysis, Criterion Referenced Tests, Foreign Countries, Latent Trait Theory
Peer reviewedLord, Frederic M. – Educational and Psychological Measurement, 1971
A number of empirical studies are suggested to answer certain questions in connection with flexilevel tests. (MS)
Descriptors: Comparative Analysis, Difficulty Level, Guessing (Tests), Item Analysis
Peer reviewedKoehler, Roger A. – Journal of Educational Measurement, 1971
Descriptors: Achievement Tests, Comparative Analysis, Confidence Testing, Grade 11
Schnipke, Deborah L.; Reese, Lynda M. – 1997
Two-stage and multistage test designs provide a way of roughly adapting item difficulty to test-taker ability. All test takers take a parallel stage-one test, and, based on their scores, they are routed to tests of different difficulty levels in subsequent stages. These designs provide some of the benefits of standard computerized adaptive testing…
Descriptors: Ability, Adaptive Testing, Algorithms, Comparative Analysis
Peer reviewedGreen, Bert F. – Educational Measurement: Issues and Practice, 1995
If annual performance assessments are to yield results that can be compared from year to year, many technical problems must be addressed. It is essential that tests to be equated measure the same construct. Methods of equating performance assessment scores, ways of equating system assessments, and standard setting are discussed. (SLD)
Descriptors: Comparative Analysis, Educational Assessment, Educational Change, Equated Scores
Angoff, William H.; Schrader, William B. – 1982
In a study to determine whether a shift from Formula scoring to Rights scoring can be made without causing a discontinuity in the test scale, the analysis of special administrations of the Scholastic Aptitude Test and Chemistry Achievement Test and the variable section of an operational form of the Graduate Management Admission Test (GMAT) is…
Descriptors: Comparative Analysis, Equated Scores, Guessing (Tests), Higher Education
Phillips, Gary W., Ed. – 1996
Recently, there has been a significant expansion in the use of performance assessment in large scale testing programs. Although there has been significant support from curriculum and policy stakeholders, the technical feasibility of large scale performance assessments has remained a question. This report is intended to contribute to the debate by…
Descriptors: Comparative Analysis, Generalizability Theory, Performance Based Assessment, Psychometrics
Crews, William E., Jr. – 1991
As part of a study of teacher evaluation of student replies to open-ended questions, a second question--the best method of determining interrater reliability--was examined. The standard method, the Pearson Product-Moment correlation, overestimated the degree of match between researchers' and teachers' scoring of tests. The simpler percent…
Descriptors: Comparative Analysis, Elementary School Teachers, Evaluation Methods, Evaluators
Cross, Lawrence H. – 1975
A novel scoring procedure was investigated in order to obtain scores from a conventional multiple-choice test that would be free of the guessing component or contain a known guessing component even though examinees were permitted to guess at will. Scores computed with the experimental procedure are based not only on the number of items answered…
Descriptors: Algebra, Comparative Analysis, Guessing (Tests), High Schools
Peer reviewedSattler, Jerome M. – Journal of School Psychology, 1976
The study investigated levels of agreement among graduate students (n=14) and school psychologists (n=18) in scoring drawings for the 10 designs on the WPPSI Geometric Design subtest. Considerable scoring disagreement occurred within each group. Results suggest careful study of the WPPSI scoring criteria is needed to achieve scoring proficiency.…
Descriptors: Comparative Analysis, Criteria, Criterion Referenced Tests, Elementary Secondary Education
Previous Page | Next Page ยป
Pages: 1 | 2
Direct link
