Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 2 |
Descriptor
Comparative Analysis | 34 |
Mathematical Models | 34 |
Test Reliability | 34 |
Error of Measurement | 12 |
Statistical Analysis | 12 |
Higher Education | 8 |
Test Items | 8 |
Test Construction | 6 |
Criterion Referenced Tests | 5 |
Item Analysis | 5 |
Analysis of Variance | 4 |
More ▼ |
Source
Educational and Psychological… | 3 |
Journal of Educational… | 3 |
Journal of Experimental… | 2 |
Psychometrika | 2 |
International Association for… | 1 |
Multivariate Behavioral… | 1 |
Sociological Methods &… | 1 |
Author
Bashaw, W. L. | 2 |
Benson, Jeri | 2 |
Huynh, Huynh | 2 |
Reckase, Mark D. | 2 |
Rentz, R. Robert | 2 |
Ackerman, Terry A. | 1 |
Agunwamba, Christian C. | 1 |
Barton, Mark A. | 1 |
Braumoeller, Bear F. | 1 |
Brennan, Robert L. | 1 |
Brown, R. L. | 1 |
More ▼ |
Publication Type
Reports - Research | 20 |
Journal Articles | 9 |
Speeches/Meeting Papers | 9 |
Reports - Evaluative | 8 |
Collected Works - Proceedings | 1 |
Information Analyses | 1 |
Numerical/Quantitative Data | 1 |
Education Level
Elementary Secondary Education | 1 |
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Asia | 1 |
Australia | 1 |
Brazil | 1 |
Connecticut | 1 |
Denmark | 1 |
Egypt | 1 |
Estonia | 1 |
Florida | 1 |
Germany | 1 |
Greece | 1 |
Hawaii | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
SAT (College Admission Test) | 2 |
Comprehensive Tests of Basic… | 1 |
General Educational… | 1 |
NEO Personality Inventory | 1 |
School and College Ability… | 1 |
What Works Clearinghouse Rating
Braumoeller, Bear F. – Sociological Methods & Research, 2017
Fuzzy-set qualitative comparative analysis (fsQCA) has become one of the most prominent methods in the social sciences for capturing causal complexity, especially for scholars with small- and medium-"N" data sets. This research note explores two key assumptions in fsQCA's methodology for testing for necessary and sufficient…
Descriptors: Qualitative Research, Comparative Analysis, Social Science Research, Research Methodology

Jackson, Paul H.; Agunwamba, Christian C. – Psychometrika, 1977
Finding and interpreting lower bounds for reliability coefficients for tests with nonhomogenous items has been a problem for psychometricians. This paper presents a mathematical formula for finding the greatest lower bound for such a coefficient. (Author/JKS)
Descriptors: Comparative Analysis, Mathematical Models, Measurement, Test Interpretation

Zumbo, Bruno D.; And Others – Journal of Experimental Education, 1992
An error in an essential equation within the article by Williams and Zimmerman is corrected, and the algebraic inequalities are translated into questions a researcher can ask about simple or residualized difference scores. Williams and Zimmerman acknowledge the error and note that main conclusions are not affected. (SLD)
Descriptors: Algebra, Comparative Analysis, Equations (Mathematics), Mathematical Models

Raju, Nambury S. – Educational and Psychological Measurement, 1977
A rederivation of Lord's formula for estimating variance in multiple matrix sampling is presented as well as the ways Cronbach's coefficient alpha and the Spearman-Brown prophecy formula are related in this context. (Author/JKS)
Descriptors: Analysis of Variance, Comparative Analysis, Item Sampling, Mathematical Models
Huynh, Huynh – 1977
Three techniques for estimating Kuder Richardson reliability (KR20) coefficients for incomplete data are contrasted. The methods are: (1) Henderson's Method 1 (analysis of variance, or ANOVA); (2) Henderson's Method 3 (FITCO); and (3) Koch's method of symmetric sums (SYSUM). A Monte Carlo simulation was used to assess the precision of the three…
Descriptors: Analysis of Variance, Comparative Analysis, Mathematical Models, Monte Carlo Methods
Barton, Mark A.; Lord, Frederic M. – 1981
An upper-asymptote parameter was added to the three-parameter logistic item response model. This four-parameter model was compared to the three-parameter model on four data sets. The fourth parameter increased the likelihood in only two of the four sets. Ability estimates for the students were generally unchanged by the introduction of the fourth…
Descriptors: College Entrance Examinations, Comparative Analysis, Latent Trait Theory, Mathematical Formulas
Mandeville, Garrett K.
Results of a comparative study of F and Q tests, in a randomized block design with one replication per cell, are presented. In addition to these two procedures, a multivariate test was also considered. The model and test statistics, data generation and parameter selection, results, summary and conclusions are presented. Ten tables contain the…
Descriptors: Comparative Analysis, Data Analysis, Mathematical Models, Models
Petersen, Nancy S.; And Others – 1981
Three equating methods were compared in terms of magnitude of scale drift: equipercentile equating, linear equating, and item response theory (IRT) equating. A sample of approximately 2670 cases was selected for each pairing of a form of the Scholastic Aptitude Tests (SAT) and an anchor test. Of the two conventional equating methods,…
Descriptors: College Entrance Examinations, Comparative Analysis, Equated Scores, Latent Trait Theory
Koch, Bill R.; Reckase, Mark D. – 1978
A live tailored testing study was conducted to compare the results of using either the one-parameter logistic model or the three-parameter logistic model to measure the performance of college students on multiple choice vocabulary items. The results of the study showed the three-parameter tailored testing procedure to be superior to the…
Descriptors: Adaptive Testing, Comparative Analysis, Goodness of Fit, Higher Education

Zimmerman, Donald W.; And Others – Journal of Experimental Education, 1984
Three types of test were compared: a completion test, a matching test, and a multiple-choice test. The completion test was more reliable than the matching test, and the matching test was more reliable than the multiple-choice test. (Author/BW)
Descriptors: Comparative Analysis, Error of Measurement, Higher Education, Mathematical Models

Huynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models

Lovett, Hubert T. – 1975
The reliability of a criterion referenced test was defined as a measure of the degree to which the test discriminates between an individual's level of performance and a predetermined criterion level. The variances of observed and true scores were defined as the squared deviation of the score from the criterion. Based on these definitions and the…
Descriptors: Career Development, Comparative Analysis, Criterion Referenced Tests, Mathematical Models

Brennan, Robert L.; Kane, Michael T. – Psychometrika, 1977
Using the assumption of randomly parallel tests and concepts from generalizability theory, three signal/noise ratios for domain-referenced tests are developed, discussed, and compared. The three ratios have the same noise but different signals depending upon the kind of decision to be made as a result of measurement. (Author/JKS)
Descriptors: Comparative Analysis, Criterion Referenced Tests, Error of Measurement, Mathematical Models
Linacre, John M. – 1993
Generalizability theory (G-theory) and many-facet Rasch measurement (Rasch) manage the variability inherent when raters rate examinees on test items. The purpose of G-theory is to estimate test reliability in a raw score metric. Unadjusted examinee raw scores are reported as measures. A variance component is estimated for the examinee…
Descriptors: Comparative Analysis, Equations (Mathematics), Estimation (Mathematics), Evaluators

Kolen, Michael J.; Whitney, Douglas R. – Journal of Educational Measurement, 1982
The adequacy of equipercentile, linear, one-parameter (Rasch), and three-parameter logistic item-response theory procedures for equating 12 forms of five tests of general educational development were compared. Results indicated the equating method adequacy depends on a variety of factors such as test characteristics, equating design, and sample…
Descriptors: Achievement Tests, Comparative Analysis, Equated Scores, Equivalency Tests