Publication Date
In 2025 | 1 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 9 |
Descriptor
Source
Author
Publication Type
Reports - Evaluative | 25 |
Journal Articles | 14 |
Speeches/Meeting Papers | 9 |
Reports - Research | 2 |
Tests/Questionnaires | 1 |
Education Level
Higher Education | 7 |
Postsecondary Education | 2 |
Adult Education | 1 |
Elementary Secondary Education | 1 |
High Schools | 1 |
Two Year Colleges | 1 |
Audience
Researchers | 2 |
Laws, Policies, & Programs
Elementary and Secondary… | 1 |
No Child Left Behind Act 2001 | 1 |
Race to the Top | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Ole J. Kemi – Advances in Physiology Education, 2025
Students are assessed by coursework and/or exams, all of which are marked by assessors (markers). Student and marker performances are then subject to end-of-session board of examiner handling and analysis. This occurs annually and is the basis for evaluating students but also the wider learning and teaching efficiency of an academic institution.…
Descriptors: Undergraduate Students, Evaluation Methods, Evaluation Criteria, Academic Standards
Turgut, Guliz – Clearing House: A Journal of Educational Strategies, Issues and Ideas, 2013
The ranking of the United States in major international tests such as the Progress in International Reading Literacy Study (PIRLS), Trends in International Mathematics and Science Study (TIMSS), and Program for International Student Assessment (PISA) is used as the driving force and rationale for the current educational reforms in the United…
Descriptors: Educational Change, Success, Educational Strategies, Educational Indicators
Feuer, Michael J. – Educational Testing Service, 2011
Few arguments about education are as effective at galvanizing public attention and motivating political action as those that compare the performance of students with their counterparts in other countries and that connect academic achievement to economic performance. Because data from international large-scale assessments (ILSA) have a powerful…
Descriptors: International Assessment, Test Interpretation, Testing Problems, Comparative Testing
Steedle, Jeffrey; Kugelmass, Heather; Nemeth, Alex – Change: The Magazine of Higher Learning, 2010
Many postsecondary institutions currently administer standardized tests of general college outcomes; more than a quarter of Association of American Colleges and Universities (AAC&U) member institutions do so. Using standardized tests for accountability purposes has been contentious mainly because these tests do not measure every important…
Descriptors: Test Results, Standardized Tests, Test Validity, Educational Testing
Liow, Jong-Leng – European Journal of Engineering Education, 2008
Peer assessment has been studied in various situations and actively pursued as a means by which students are given more control over their learning and assessment achievement. This study investigated the reliability of staff and student assessments in two oral presentations with limited feedback for a school-based thesis course in engineering…
Descriptors: Feedback (Response), Student Evaluation, Grade Point Average, Peer Evaluation
Mueller, Karsten; Liebig, Christian; Hattrup, Keith – Educational and Psychological Measurement, 2007
Two quasi-experimental field studies were conducted to evaluate the psychometric equivalence of computerized and paper-and-pencil job satisfaction measures. The present research extends previous work in the area by providing better control of common threats to validity in quasi-experimental research on test mode effects and by evaluating a more…
Descriptors: Psychometrics, Field Studies, Job Satisfaction, Computer Assisted Testing
Bower, Anna; Hayes, Alan – American Journal on Mental Retardation, 1995
This study compared global scores of 26 Australian students (ages 4 to 16) with mental retardation on the third and fourth editions of the Stanford Binet Intelligence Scale. Analysis indicated a fairly strong positive relation between the two tests, suggesting that the fourth edition may be substituted for the older instrument in longitudinal…
Descriptors: Comparative Testing, Concurrent Validity, Intelligence Tests, Mental Retardation
Kong, Xiaojing J.; Wise, Steven L.; Bhola, Dennison S. – Educational and Psychological Measurement, 2007
This study compared four methods for setting item response time thresholds to differentiate rapid-guessing behavior from solution behavior. Thresholds were either (a) common for all test items, (b) based on item surface features such as the amount of reading required, (c) based on visually inspecting response time frequency distributions, or (d)…
Descriptors: Test Items, Reaction Time, Timed Tests, Item Response Theory

Sevier, Robert; And Others – Roeper Review, 1994
Thirty-five gifted elementary children were tested with the Wechsler Intelligence Scale for Children III (WISC-III) and results were compared to a previous administration of the WISC (Revised). Scores were significantly higher on all three WISC-R global scales and most subtest scaled scores. Correlation coefficients between the tests were…
Descriptors: Ability Identification, Comparative Testing, Correlation, Elementary Education
Kirkup, Catherine; Schagen, Ian; Wheater, Rebecca; Morrison, Jo; Whetton, Chris – National Foundation for Educational Research, 2007
In September 2005 the National Foundation for Educational Research (NFER) in association with the Department for Education and Skills (DfES), the Sutton Trust and the College Board, began a five-year research study to examine the validity of an aptitude test in higher education admissions. This report describes and explores the relationships…
Descriptors: Educational Research, Academic Achievement, Student Surveys, Aptitude Tests
Allen, Nancy L.; And Others – 1992
Many testing programs include a section of optional questions in addition to mandatory parts of a test. These optional parts of a test are not often truly parallel to one another, and groups of examinees selecting each optional test section are not equivalent to one another. This paper provides a general method based on missing-data methods for…
Descriptors: Comparative Testing, Estimation (Mathematics), Graphs, Scaling
Squires, David; Trevisan, Michael S.; Canney, George F. – Studies in Educational Evaluation, 2006
The Idaho Comprehensive Literacy Assessment (ICLA) is a faculty-developed, state-wide, high-stakes assessment of pre-service teachers' knowledge and application of research based literacy practices. The literacy faculty control all aspects of the test, including construction, refinement, administration, scoring and reporting. The test development…
Descriptors: Test Construction, Comparative Testing, Investigations, Test Reliability

Wainer, Howard; And Others – Journal of Educational Measurement, 1992
Computer simulations were run to measure the relationship between testlet validity and factors of item pool size and testlet length for both adaptive and linearly constructed testlets. Making a testlet adaptive yields only modest increases in aggregate validity because of the peakedness of the typical proficiency distribution. (Author/SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation
Laird, Barbara B. – Inquiry, 2003
Laird studies the relationship between two computerized nursing tests and finds a relationship between the two sets of scores. (Contains 2 tables.)
Descriptors: Nursing Education, Nurses, Computer Assisted Testing, Comparative Testing
DeMars, Christine E. – Online Submission, 2005
Several methods for estimating item response theory scores for multiple subtests were compared. These methods included two multidimensional item response theory models: a bi-factor model where each subtest was a composite score based on the primary trait measured by the set of tests and a secondary trait measured by the individual subtest, and a…
Descriptors: Item Response Theory, Multidimensional Scaling, Correlation, Scoring Rubrics
Previous Page | Next Page ยป
Pages: 1 | 2