NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 46 to 60 of 1,164 results Save | Export
Gilraine, Michael; Gu, Jiaying; McMillan, Robert – National Bureau of Economic Research, 2020
This paper proposes a new methodology for estimating teacher value-added. Rather than imposing a normality assumption on unobserved teacher quality (as in the standard empirical Bayes approach), our nonparametric estimator permits the underlying distribution to be estimated directly and in a computationally feasible way. The resulting estimates…
Descriptors: Value Added Models, Teacher Effectiveness, Nonparametric Statistics, Computation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Simsek, Ahmet Salih – International Journal of Assessment Tools in Education, 2023
Likert-type item is the most popular response format for collecting data in social, educational, and psychological studies through scales or questionnaires. However, there is no consensus on whether parametric or non-parametric tests should be preferred when analyzing Likert-type data. This study examined the statistical power of parametric and…
Descriptors: Error of Measurement, Likert Scales, Nonparametric Statistics, Statistical Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mor, Ezgi; Kula-Kartal, Seval – International Journal of Assessment Tools in Education, 2022
The dimensionality is one of the most investigated concepts in the psychological assessment, and there are many ways to determine the dimensionality of a measured construct. The Automated Item Selection Procedure (AISP) and the DETECT are non-parametric methods aiming to determine the factorial structure of a data set. In the current study,…
Descriptors: Psychological Evaluation, Nonparametric Statistics, Test Items, Item Analysis
Zebing Wu – ProQuest LLC, 2024
Response style, one common aberrancy in non-cognitive assessments in psychological fields, is problematic in terms of inaccurate estimation of item and person parameters, which leads to serious reliability, validity, and fairness issues (Baumgartner & Steenkamp, 2001; Bolt & Johnson, 2009; Bolt & Newton, 2011). Response style refers to…
Descriptors: Response Style (Tests), Accuracy, Preferences, Psychological Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Calver, Michael; Fletcher, Douglas – American Biology Teacher, 2020
Data collected in many biology laboratory classes are on ratio or interval scales where the size interval between adjacent units on the scale is constant, which is a critical requirement for analysis with parametric statistics such as t-tests or analysis of variance. In other cases, such as ratings of disease or behavior, data are collected on…
Descriptors: Statistical Analysis, Data Collection, Biology, Science Laboratories
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Myszkowski, Nils – Journal of Intelligence, 2020
Raven's Standard Progressive Matrices (Raven 1941) is a widely used 60-item long measure of general mental ability. It was recently suggested that, for situations where taking this test is too time consuming, a shorter version, comprised of only the last series of the Standard Progressive Matrices (Myszkowski and Storme 2018) could be used, while…
Descriptors: Intelligence Tests, Psychometrics, Nonparametric Statistics, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Metsämuuronen, Jari – International Journal of Educational Methodology, 2020
Kelley's Discrimination Index (DI) is a simple and robust, classical non-parametric short-cut to estimate the item discrimination power (IDP) in the practical educational settings. Unlike item-total correlation, DI can reach the ultimate values of +1 and -1, and it is stable against the outliers. Because of the computational easiness, DI is…
Descriptors: Test Items, Computation, Item Analysis, Nonparametric Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Yu; Chiu, Chia-Yi; Köhn, Hans Friedrich – Journal of Educational and Behavioral Statistics, 2023
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Nonparametric Statistics, Test Format, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, HyeSun; Smith, Weldon; Martinez, Angel; Ferris, Heather; Bova, Joe – Applied Measurement in Education, 2021
The aim of the current research was to provide recommendations to facilitate the development and use of anchoring vignettes (AVs) for cross-cultural comparisons in education. Study 1 identified six factors leading to order violations and ties in AV responses based on cognitive interviews with 15-year-old students. The factors were categorized into…
Descriptors: Vignettes, Test Items, Equated Scores, Nonparametric Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Yukhymenko-Lescroart, Mariya A.; Goldman, Susan R.; Lawless, Kimberly A.; Pellegrino, James W.; Shanahan, Cynthia R. – Educational Psychology, 2022
To extend the existing research examining multiple text comprehension and its assessment, we developed a verification task approach to assessing of information that was "explicitly" and "implicitly" presented "within" and across nine texts. A nonparametric form of signal detection theory was used to analyse the…
Descriptors: Task Analysis, Reading Comprehension, Middle School Students, Nonparametric Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A. – Language Testing, 2019
Differences in rater judgments that are systematically related to construct-irrelevant characteristics threaten the fairness of rater-mediated writing assessments. Accordingly, it is essential that researchers and practitioners examine the degree to which the psychometric quality of rater judgments is comparable across test-taker subgroups.…
Descriptors: Nonparametric Statistics, Interrater Reliability, Differences, Writing Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A. – Measurement: Interdisciplinary Research and Perspectives, 2020
A major challenge in the widespread application of Mokken scale analysis (MSA) to educational performance assessments is the requirement of complete data, where every rater rates every student. In this study, simulated and real data are used to demonstrate a method by which researchers and practitioners can apply MSA to incomplete rating designs.…
Descriptors: Item Response Theory, Scaling, Nonparametric Statistics, Performance Based Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Walter M. Stroup; Anthony Petrosino; Corey Brady; Karen Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Tests of statistical significance often play a decisive role in establishing the empirical warrant of evidence-based research in education. The results from pattern-based assessment items, as introduced in this paper, are categorical and multimodal and do not immediately support the use of measures of central tendency as typically related to…
Descriptors: Statistical Significance, Comparative Analysis, Research Methodology, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Ghaemi, Hamed – Language Testing in Asia, 2022
Listening comprehension in English, as one of the most fundamental skills, has an essential role in the process of learning English. Mokken scale analysis (MSA) is a probabilistic-nonparametric approach to item response theory (IRT) which determines the one-dimensionality and scalability of test. Mokken scaling techniques are a useful tool for…
Descriptors: Second Language Learning, English (Second Language), Nonparametric Statistics, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Dimitrov, Dimiter M.; Atanasov, Dimitar V.; Luo, Yong – Measurement: Interdisciplinary Research and Perspectives, 2020
This study examines and compares four person-fit statistics (PFSs) in the framework of the "D"- scoring method (DSM): (a) van der Flier's "U3" statistic; (b) "Ud" statistic, as a modification of "U3" under the DSM; (c) "Zd" statistic, as a modification of the "Z3 (l[subscript z])"…
Descriptors: Goodness of Fit, Item Analysis, Item Response Theory, Scoring
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  78