Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 6 |
Since 2016 (last 10 years) | 23 |
Since 2006 (last 20 years) | 75 |
Descriptor
Item Response Theory | 89 |
Psychometrics | 89 |
Models | 35 |
Test Items | 25 |
Test Construction | 22 |
Evaluation Methods | 15 |
Student Evaluation | 13 |
Computation | 12 |
Educational Assessment | 12 |
Measurement | 10 |
Measurement Techniques | 10 |
More ▼ |
Source
Author
Boone, William J. | 3 |
Borsboom, Denny | 2 |
De Boeck, Paul | 2 |
Embretson, Susan E. | 2 |
Ferrando, Pere J. | 2 |
Ferrara, Steve | 2 |
Maris, Gunter | 2 |
Rupp, Andre A. | 2 |
Toland, Michael D. | 2 |
White, Sheida | 2 |
Zumbo, Bruno D. | 2 |
More ▼ |
Publication Type
Education Level
Elementary Secondary Education | 11 |
Elementary Education | 10 |
Secondary Education | 7 |
Higher Education | 6 |
Grade 5 | 5 |
Middle Schools | 5 |
Early Childhood Education | 4 |
Grade 4 | 4 |
High Schools | 4 |
Intermediate Grades | 4 |
Grade 3 | 3 |
More ▼ |
Audience
Researchers | 6 |
Practitioners | 2 |
Policymakers | 1 |
Students | 1 |
Teachers | 1 |
Location
New Mexico | 2 |
United Kingdom (England) | 2 |
Australia | 1 |
China | 1 |
India | 1 |
Indiana | 1 |
Italy | 1 |
Netherlands | 1 |
Turkey | 1 |
United States | 1 |
Wisconsin | 1 |
More ▼ |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Kim, Yunsung; Sreechan; Piech, Chris; Thille, Candace – International Educational Data Mining Society, 2023
Dynamic Item Response Models extend the standard Item Response Theory (IRT) to capture temporal dynamics in learner ability. While these models have the potential to allow instructional systems to actively monitor the evolution of learner proficiency in real time, existing dynamic item response models rely on expensive inference algorithms that…
Descriptors: Item Response Theory, Accuracy, Inferences, Algorithms
Dahl, Laura S.; Staples, B. Ashley; Mayhew, Matthew J.; Rockenbach, Alyssa N. – Innovative Higher Education, 2023
Surveys with rating scales are often used in higher education research to measure student learning and development, yet testing and reporting on the longitudinal psychometric properties of these instruments is rare. Rasch techniques allow scholars to map item difficulty and individual aptitude on the same linear, continuous scale to compare…
Descriptors: Surveys, Rating Scales, Higher Education, Educational Research
Chen, Yunxiao; Lee, Yi-Hsuan; Li, Xiaoou – Journal of Educational and Behavioral Statistics, 2022
In standardized educational testing, test items are reused in multiple test administrations. To ensure the validity of test scores, the psychometric properties of items should remain unchanged over time. In this article, we consider the sequential monitoring of test items, in particular, the detection of abrupt changes to their psychometric…
Descriptors: Standardized Tests, Test Items, Test Validity, Scores
Abbitt, Jason T.; Boone, William J. – Journal of Computing in Higher Education, 2021
This article presents the results of evaluating a dataset collected with the Community of inquiry (CoI) survey (Arbaugh, The International Review of Research in Open and Distributed Learning 9:1-21, 2008) using Rasch psychometric techniques to evaluate instrument functioning. Data were collected over a two-year period yielding a sample of 704…
Descriptors: Communities of Practice, Student Surveys, Item Response Theory, Psychometrics
Student, Sanford R. – Educational Researcher, 2022
Empirical growth benchmarks, as introduced by Hill, Bloom, Black, and Lipsey (2008), are a well-known way to contextualize effect sizes in education research. Past work on these benchmarks, both positive and negative, has largely avoided confronting the role of vertical scales, yet technical issues with vertical scales trouble the use of such…
Descriptors: Computer Simulation, Benchmarking, Effect Size, Intervention
Schumacker, Randall – Measurement: Interdisciplinary Research and Perspectives, 2019
The R software provides packages and functions that provide data analysis in classical true score, generalizability theory, item response theory, and Rasch measurement theories. A brief list of notable articles in each measurement theory and the first measurement journals is followed by a list of R psychometric software packages. Each psychometric…
Descriptors: Psychometrics, Computer Software, Measurement, Item Response Theory
Levy, Roy – Educational Measurement: Issues and Practice, 2020
In this digital ITEMS module, Dr. Roy Levy describes Bayesian approaches to psychometric modeling. He discusses how Bayesian inference is a mechanism for reasoning in a probability-modeling framework and is well-suited to core problems in educational measurement: reasoning from student performances on an assessment to make inferences about their…
Descriptors: Bayesian Statistics, Psychometrics, Item Response Theory, Statistical Inference
Kortemeyer, Gerd – Physics Teacher, 2019
Item Response Theory (IRT) has proven useful in physics education research to examine the validity of concept tests (e.g., Refs. 1-3) and online homework (e.g. Refs. 4-6), yet as a tool for the improvement of physics instruction (particularly exams), it is oftentimes perceived as (a) mysterious, (b) unjustified, and (c) impractical. This article…
Descriptors: Item Response Theory, Physics, Science Education, Psychometrics
Komperda, Regis; Pentecost, Thomas C.; Barbera, Jack – Journal of Chemical Education, 2018
This methodological paper examines current conceptions of reliability in chemistry education research (CER) and provides recommendations for moving beyond the current reliance on reporting coefficient alpha (a) as reliability evidence without regard to its appropriateness for the research context. To help foster a better understanding of…
Descriptors: Chemistry, Science Instruction, Teaching Methods, Reliability
Zehner, Fabian; Eichmann, Beate; Deribo, Tobias; Harrison, Scott; Bengs, Daniel; Andersen, Nico; Hahnel, Carolin – Journal of Educational Data Mining, 2021
The NAEP EDM Competition required participants to predict efficient test-taking behavior based on log data. This paper describes our top-down approach for engineering features by means of psychometric modeling, aiming at machine learning for the predictive classification task. For feature engineering, we employed, among others, the Log-Normal…
Descriptors: National Competency Tests, Engineering Education, Data Collection, Data Analysis
Oranje, Andreas; Kolstad, Andrew – Journal of Educational and Behavioral Statistics, 2019
The design and psychometric methodology of the National Assessment of Educational Progress (NAEP) is constantly evolving to meet the changing interests and demands stemming from a rapidly shifting educational landscape. NAEP has been built on strong research foundations that include conducting extensive evaluations and comparisons before new…
Descriptors: National Competency Tests, Psychometrics, Statistical Analysis, Computation
von Davier, Matthias; Khorramdel, Lale; He, Qiwei; Shin, Hyo Jeong; Chen, Haiwen – Journal of Educational and Behavioral Statistics, 2019
International large-scale assessments (ILSAs) transitioned from paper-based assessments to computer-based assessments (CBAs) facilitating the use of new item types and more effective data collection tools. This allows implementation of more complex test designs and to collect process and response time (RT) data. These new data types can be used to…
Descriptors: International Assessment, Computer Assisted Testing, Psychometrics, Item Response Theory
Sadeghi, Karim; Abolfazli Khonbi, Zainab – Language Testing in Asia, 2017
As perfectly summarised by Ida Lawrence, "Testing is growing by leaps and bounds across the world. There is a realization that a nation's well-being depends crucially on the educational achievement of its population. Valid tests are an essential tool to evaluate a nation's educational standing and to implement efficacious educational reforms.…
Descriptors: Test Items, Item Response Theory, Computer Assisted Testing, Adaptive Testing
Embretson, Susan E. – Educational Measurement: Issues and Practice, 2016
Examinees' thinking processes have become an increasingly important concern in testing. The responses processes aspect is a major component of validity, and contemporary tests increasingly involve specifications about the cognitive complexity of examinees' response processes. Yet, empirical research findings on examinees' cognitive processes are…
Descriptors: Testing, Cognitive Processes, Test Construction, Test Items
Ames, Allison; Myers, Aaron – Educational Measurement: Issues and Practice, 2019
Drawing valid inferences from modern measurement models is contingent upon a good fit of the data to the model. Violations of model-data fit have numerous consequences, limiting the usefulness and applicability of the model. As Bayesian estimation is becoming more common, understanding the Bayesian approaches for evaluating model-data fit models…
Descriptors: Bayesian Statistics, Psychometrics, Models, Predictive Measurement