NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)2
Since 2006 (last 20 years)10
Publication Type
Journal Articles12
Reports - Research12
Information Analyses1
Audience
Location
Oklahoma1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 12 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Jewsbury, Paul A. – ETS Research Report Series, 2019
When an assessment undergoes changes to the administration or instrument, bridge studies are typically used to try to ensure comparability of scores before and after the change. Among the most common and powerful is the common population linking design, with the use of a linear transformation to link scores to the metric of the original…
Descriptors: Evaluation Research, Scores, Error Patterns, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Mather, Nancy; Wendling, Barbara J. – Journal of Psychoeducational Assessment, 2017
We reviewed 13 studies that focused on analyzing student errors on achievement tests from the Kaufman Test of Educational Achievement-Third edition (KTEA-3). The intent was to determine what instructional implications could be derived from in-depth error analysis. As we reviewed these studies, several themes emerged. We explain how a careful…
Descriptors: Achievement Tests, Educational Research, Evaluation Research, Error Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Eun Sook; Kwok, Oi-man; Yoon, Myeongsun – Structural Equation Modeling: A Multidisciplinary Journal, 2012
Testing factorial invariance has recently gained more attention in different social science disciplines. Nevertheless, when examining factorial invariance, it is generally assumed that the observations are independent of each other, which might not be always true. In this study, we examined the impact of testing factorial invariance in multilevel…
Descriptors: Monte Carlo Methods, Testing, Social Science Research, Factor Structure
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Eun Sook; Yoon, Myeongsun; Lee, Taehun – Educational and Psychological Measurement, 2012
Multiple-indicators multiple-causes (MIMIC) modeling is often used to test a latent group mean difference while assuming the equivalence of factor loadings and intercepts over groups. However, this study demonstrated that MIMIC was insensitive to the presence of factor loading noninvariance, which implies that factor loading invariance should be…
Descriptors: Test Items, Simulation, Testing, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Ying; Rupp, Andre A. – Educational and Psychological Measurement, 2011
This study investigated the Type I error rate and power of the multivariate extension of the S - [chi][squared] statistic using unidimensional and multidimensional item response theory (UIRT and MIRT, respectively) models as well as full-information bifactor (FI-bifactor) models through simulation. Manipulated factors included test length, sample…
Descriptors: Test Length, Item Response Theory, Statistical Analysis, Error Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Mrazik, Martin; Janzen, Troy M.; Dombrowski, Stefan C.; Barford, Sean W.; Krawchuk, Lindsey L. – Canadian Journal of School Psychology, 2012
A total of 19 graduate students enrolled in a graduate course conducted 6 consecutive administrations of the Wechsler Intelligence Scale for Children, 4th edition (WISC-IV, Canadian version). Test protocols were examined to obtain data describing the frequency of examiner errors, including administration and scoring errors. Results identified 511…
Descriptors: Intelligence Tests, Intelligence, Statistical Analysis, Scoring
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Hathcoat, John D.; Penn, Jeremy D. – Research & Practice in Assessment, 2012
Critics of standardized testing have recommended replacing standardized tests with more authentic assessment measures, such as classroom assignments, projects, or portfolios rated by a panel of raters using common rubrics. Little research has examined the consistency of scores across multiple authentic assignments or the implications of this…
Descriptors: Generalizability Theory, Performance Based Assessment, Writing Across the Curriculum, Standardized Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Forero, Carlos G.; Maydeu-Olivares, Alberto – Psychological Methods, 2009
The performance of parameter estimates and standard errors in estimating F. Samejima's graded response model was examined across 324 conditions. Full information maximum likelihood (FIML) was compared with a 3-stage estimator for categorical item factor analysis (CIFA) when the unweighted least squares method was used in CIFA's third stage. CIFA…
Descriptors: Factor Analysis, Least Squares Statistics, Computation, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Athy, Jeremy; Friedrich, Jeff; Delany, Eileen – Science & Education, 2008
Egon Brunswik (1903-1955) first made an interesting distinction between perception and explicit reasoning, arguing that perception included quick estimates of an object's size, nearly always resulting in good approximations in uncertain environments, whereas explicit reasoning, while better at achieving exact estimates, could often fail by wide…
Descriptors: Psychology, Logical Thinking, Perception, Psychological Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Williams, Jason; MacKinnon, David P. – Structural Equation Modeling: A Multidisciplinary Journal, 2008
Recent advances in testing mediation have found that certain resampling methods and tests based on the mathematical distribution of 2 normal random variables substantially outperform the traditional "z" test. However, these studies have primarily focused only on models with a single mediator and 2 component paths. To address this limitation, a…
Descriptors: Intervals, Testing, Predictor Variables, Effect Size
Peer reviewed Peer reviewed
Direct linkDirect link
Serlin, Ronald C.; Harwell, Michael R. – Psychological Methods, 2004
It is well-known that for normally distributed errors parametric tests are optimal statistically, but perhaps less well-known is that when normality does not hold, nonparametric tests frequently possess greater statistical power than parametric tests, while controlling Type I error rate. However, the use of nonparametric procedures has been…
Descriptors: Multiple Regression Analysis, Monte Carlo Methods, Nonparametric Statistics, Error Patterns
Peer reviewed Peer reviewed
Goldberg, Gail Lynn; Kapinus, Barbara – Applied Measurement in Education, 1993
Using responses of 123 elementary school teachers, a battery of performance-assessment tasks designed to generate responses to reading tests was evaluated from task development and scoring perspectives. More than one dozen types of errors were identified. Practical outcomes of the study and improvement of task development and scoring are…
Descriptors: Educational Assessment, Educational Practices, Elementary Education, Elementary School Teachers