Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 5 |
Since 2006 (last 20 years) | 10 |
Descriptor
Comparative Analysis | 24 |
Research Methodology | 24 |
Test Items | 24 |
Item Response Theory | 9 |
Simulation | 7 |
Evaluation Methods | 6 |
Foreign Countries | 5 |
Higher Education | 5 |
Test Construction | 5 |
Achievement Tests | 4 |
Educational Research | 4 |
More ▼ |
Source
Author
Anthony Petrosino | 1 |
Boldt, R. F. | 1 |
Chang, Lei | 1 |
Chen, Jie | 1 |
Chissom, Brad S. | 1 |
Clauser, Brian | 1 |
Cohen, Allan S. | 1 |
Corey Brady | 1 |
Dallimore, Elise J. | 1 |
Dossey, John A. | 1 |
Earley, Mark A. | 1 |
More ▼ |
Publication Type
Reports - Research | 12 |
Speeches/Meeting Papers | 10 |
Journal Articles | 9 |
Reports - Evaluative | 8 |
Reports - Descriptive | 2 |
Tests/Questionnaires | 2 |
Books | 1 |
Collected Works - General | 1 |
Dissertations/Theses -… | 1 |
Education Level
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Walter M. Stroup; Anthony Petrosino; Corey Brady; Karen Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Tests of statistical significance often play a decisive role in establishing the empirical warrant of evidence-based research in education. The results from pattern-based assessment items, as introduced in this paper, are categorical and multimodal and do not immediately support the use of measures of central tendency as typically related to…
Descriptors: Statistical Significance, Comparative Analysis, Research Methodology, Evaluation Methods
Golino, Hudson F.; Gomes, Cristiano M. A. – International Journal of Research & Method in Education, 2016
This paper presents a non-parametric imputation technique, named random forest, from the machine learning field. The random forest procedure has two main tuning parameters: the number of trees grown in the prediction and the number of predictors used. Fifty experimental conditions were created in the imputation procedure, with different…
Descriptors: Item Response Theory, Regression (Statistics), Difficulty Level, Goodness of Fit
Wagemaker, Hans, Ed. – International Association for the Evaluation of Educational Achievement, 2020
Although International Association for the Evaluation of Educational Achievement-pioneered international large-scale assessment (ILSA) of education is now a well-established science, non-practitioners and many users often substantially misunderstand how large-scale assessments are conducted, what questions and challenges they are designed to…
Descriptors: International Assessment, Achievement Tests, Educational Assessment, Comparative Analysis
Dallimore, Elise J.; Hertenstein, Julie H.; Platt, Marjorie B. – Journal on Excellence in College Teaching, 2016
Prior research indicates that the frequency with which students participate in classroom discussion is positively related to learning. This study examines students' perceptions of the relationship between their class discussion participation and learning and between preparation and both comfort with participating and actual discussion…
Descriptors: Discussion (Teaching Technique), Student Attitudes, Student Participation, Research Methodology
Gündüz, Mevlüt – Journal of Education and Training Studies, 2016
The aim of this study is to make a new classification regarding the fact that the current classifications may change constantly because of values? gaining a different dimension and importance every single day. In this research descriptive research, which was used frequently in qualitative research methods, was preferred. This research was…
Descriptors: Classification, Values, Qualitative Research, Research Methodology
Chen, Jie – ProQuest LLC, 2012
Living in an era of test-based accountability systems, how do we hold accountability tests accountable? Many accountability decisions made today are based on the assumption that test scores successfully reflect the effect of instruction. However, only instructionally sensitive assessments, not the instructionally insensitive ones, reflect the…
Descriptors: High Stakes Tests, Achievement Tests, Comparative Analysis, Research Methodology
Groß Ophoff, Jana; Schladitz, Sandra; Leuders, Juliane; Leuders, Timo; Wirtz, Markus A. – Peabody Journal of Education, 2015
The ability to purposefully access, reflect, and use evidence from educational research (Educational Research Literacy) is expected of future professionals in educational practice. Based on the presented conceptual framework, a test instrument was developed to assess the different competency aspects: Information Literacy, Statistical Literacy, and…
Descriptors: Educational Research, Research Methodology, Literacy, Educational Development
Stuive, Ilse; Kiers, Henk A. L.; Timmerman, Marieke E. – Educational and Psychological Measurement, 2009
A common question in test evaluation is whether an a priori assignment of items to subtests is supported by empirical data. If the analysis results indicate the assignment of items to subtests under study is not supported by data, the assignment is often adjusted. In this study the authors compare two methods on the quality of their suggestions to…
Descriptors: Simulation, Item Response Theory, Test Items, Factor Analysis

Tate, Richard – Applied Psychological Measurement, 2003
Compared selected methods of assessing the structure of tests with dichotomous items using real data from a 62-item test of reading ability and computer-generated data for multiple unidimensional and multidimensional cases. All methods performed reasonably well over a relatively wide range of conditions. (SLD)
Descriptors: Comparative Analysis, Reading Ability, Research Methodology, Test Construction
Hancock, Dawson R. – Assessment & Evaluation in Higher Education, 2007
This study explored the impact of individual versus two-person group testing on graduate students' achievement and motivation to learn while enrolled in a 16-lesson educational research methods course. Students in two sections of the course were exposed to the same content and instructional methods, with one exception: students in one section took…
Descriptors: Teaching Methods, Test Items, Learning Motivation, Group Testing

Nandakumar, Ratna – Journal of Educational Measurement, 1994
Using simulated and real data, this study compares the performance of three methodologies for assessing unidimensionality: (1) DIMTEST; (2) the approach of Holland and Rosenbaum; and (3) nonlinear factor analysis. All three models correctly confirm unidimensionality, but they differ in their ability to detect the lack of unidimensionality.…
Descriptors: Ability, Comparative Analysis, Evaluation Methods, Factor Analysis

Clauser, Brian; And Others – Applied Measurement in Education, 1993
The usefulness of a two-step version of the Mantel Haenszel procedure for distinguishing between differential item functioning (DIF) and item impact was studied by comparing the single-step and two-step procedures using a simulated data set. Results show changes in the identification rate for the two-step methods. (SLD)
Descriptors: Comparative Analysis, Evaluation Methods, Identification, Item Bias
Mertler, Craig A.; Earley, Mark A. – 2003
A study was conducted to compare the psychometric qualities of two forms of an identical survey: one administered in a paper-and-pencil format and the other administered in Web format. The survey addressed the topic of college course anxiety and used to survey a sample of 236 undergraduate students. The psychometric qualities investigated included…
Descriptors: Anxiety, Comparative Analysis, Higher Education, Psychometrics
Schumacker, Randall E.; And Others – 1994
Rasch between and total weighted and unweighted fit statistics were compared using varying test lengths and sample sizes. Two test lengths (20 and 50 items) and three sample sizes (150, 500, and 1,000 were crossed. Each of the six combinations were replicated 100 times. In addition, power comparisons were made. Results indicated that there were no…
Descriptors: Comparative Analysis, Goodness of Fit, Item Response Theory, Power (Statistics)
Boldt, R. F. – 1994
The comparison of item response theory models for the Test of English as a Foreign Language (TOEFL) was extended to an equating context as simulation trials were used to "equate the test to itself." Equating sample data were generated from administration of identical item sets. Equatings that used procedures based on each model (simple…
Descriptors: Comparative Analysis, Cutting Scores, English (Second Language), Equated Scores
Previous Page | Next Page »
Pages: 1 | 2