NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)1
Since 2006 (last 20 years)5
Audience
Researchers1
Laws, Policies, & Programs
No Child Left Behind Act 20011
What Works Clearinghouse Rating
Showing 1 to 15 of 27 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Vaheoja, Monika; Verhelst, N. D.; Eggen, T.J.H.M. – European Journal of Science and Mathematics Education, 2019
In this article, the authors applied profile analysis to Maths exam data to demonstrate how different exam forms, differing in difficulty and length, can be reported and easily interpreted. The results were presented for different groups of participants and for different institutions in different Maths domains by evaluating the balance. Some…
Descriptors: Feedback (Response), Foreign Countries, Statistical Analysis, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Spears, Karen; Wilson, Mary – Journal of Extension, 2010
Evaluation is an essential component of any Extension education program. One tool, the pre- and post-test, provides measurable evaluation data. Yet often the answer "I don't know" or all possible answers to a multiple choice question are not included in the repeated measure analysis. Because more than two answers are offered, the test of marginal…
Descriptors: Extension Education, Computer Assisted Instruction, Statistical Analysis, Pretests Posttests
Stoneberg, Bert D. – Online Submission, 2009
Test developers are responsible to define how test scores should be interpreted and used. The No Child Left Behind Act of 2001 (NCLB) directed the Secretary of Education to use results from the National Assessment of Educational Progress (NAEP) to confirm the proficiency scores from state developed tests. There are two sets of federal definitions…
Descriptors: National Competency Tests, State Programs, Achievement Tests, Scores
Lang, W. Steve; Wilkerson, Judy R. – Online Submission, 2008
The National Council for Accreditation of Teacher Education (NCATE, 2002) requires teacher education units to develop assessment systems and evaluate both the success of candidates and unit operations. Because of a stated, but misguided, fear of statistics, NCATE fails to use accepted terminology to assure the quality of institutional evaluative…
Descriptors: State Standards, Validity, Resource Materials, Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Brown, James Dean – Language Assessment Quarterly, 2008
In keeping with the theme of the International Language Testing Association/Language Testing Research Colloquium Conference in 2008, "Focusing on the Core: Justifying the Use of Language Assessments to Stakeholders," I define "stakeholder-friendly tests," "defensible testing," and "testing-context analysis."…
Descriptors: Language Usage, Curriculum Development, Testing, Language Tests
Peer reviewed Peer reviewed
Clauser, Brian E.; Mazor, Kathleen M. – Educational Measurement: Issues and Practice, 1998
This module prepares the reader to use statistical procedures to detect differentially functioning test items. The Mantel-Haenszel statistic, logistic regression, the SIBTEST procedure, the Standardization procedure, and various item response theory-based procedures are presented. Theoretical frameworks, strengths and weaknesses, and…
Descriptors: Item Bias, Item Response Theory, Statistical Analysis, Teaching Methods
Mislevy, Robert J. – 1995
Educational test theory consists of statistical and methodological tools to support inferences about examinees' knowledge, skills, and accomplishments. The evolution of test theory has been shaped by the nature of users' inferences which, until recently, have been framed almost exclusively in terms of trait and behavioral psychology. Progress in…
Descriptors: Cognitive Psychology, Developmental Psychology, Educational Testing, Inferences
Crossman, Leslie L. – 1994
The present paper suggests that multivariate techniques are very important in social science research, and that canonical correlation analysis may be particularly useful. The logic of canonical analysis is explained and discussed. The necessity of using replicability/generalizability analyses is argued. It is suggested that cross-validation…
Descriptors: Correlation, Generalizability Theory, Heuristics, Multivariate Analysis
Gardner, Eric – 1989
Five of the common misuses of tests are reviewed: (1) acceptance of the test title as an accurate and complete description of the variable being measured (failure to examine the manual and the items carefully to know the specific aspects to be tested can result in misuse through selection of an inappropriate test for a particular purpose or…
Descriptors: Error of Measurement, Evaluation Problems, Examiners, Scoring
Peer reviewed Peer reviewed
Larson, Gerald E. – Intelligence, 1990
Sternberg and Gastel have provided an example of a major principle of intelligence research--the relationship between a task's working memory demands and its sensitivity to individual differences in fluid intelligence and "g." There is no need to invoke additional constructs such as "novelty." (SLD)
Descriptors: Cognitive Processes, Individual Differences, Intelligence, Intelligence Tests
Peer reviewed Peer reviewed
Sternberg, Robert J. – Intelligence, 1990
It is asserted that the statistical arguments proposed by Humphreys are not supported by the formula he presents. The disagreement with Larson is with the contention that there exists one correct theoretical framework in which intelligence research should be posed. The respective viewpoints are seen as complementary, not competing. (SLD)
Descriptors: Intelligence, Intelligence Tests, Memory, Novelty (Stimulus Dimension)
Thackrey, Michael – American Journal on Mental Retardation, 1991
Principal components analysis of standardization sample data for the Comprehensive Test of Adaptive Behavior, involving 6,647 persons with mental retardation, found that a single factor accounted for 86 percent of the variance in category scores. Use of the Total Score and idiographic item clusters is psychometrically preferable to use of category…
Descriptors: Adaptive Behavior (of Disabled), Evaluation Methods, Factor Analysis, Mental Retardation
Peer reviewed Peer reviewed
MacCann, Robert G. – Journal of Educational Statistics, 1990
For anchor test equating, 3 linear observed score methods are derived for populations differing in ability. Each version requires that the correlations of the tests with the selection variable be known. Five sets of assumptions are made for each model--yielding 15 methods--which are then related to existing methods. (SLD)
Descriptors: Ability, Ability Grouping, Equated Scores, Equations (Mathematics)
Beaton, Albert E.; Johnson, Eugene G. – 1990
When the Educational Testing Service became the administrator of the National Assessment of Educational Progress (NAEP) in 1983, it introduced scales based on item response theory (IRT) as a way of presenting results of the assessment to the general public. Some properties of the scales and their uses are discussed. Initial attempts at presenting…
Descriptors: Academic Achievement, Data Interpretation, Educational Assessment, Elementary Secondary Education
PDF pending restoration PDF pending restoration
Fennessey, James; Salganik, Laura Hersh – 1982
An explicit model identifying 10 relevant components of achievement gain scores has been developed. Based on that model, all students under consideration are stratified according to individual observed pretest score, and achievement gains are measured relative to the average and range of gains among students in the same prescore stratum. The…
Descriptors: Achievement Gains, Elementary Secondary Education, Models, Pretests Posttests
Previous Page | Next Page ยป
Pages: 1  |  2