NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)1
Since 2007 (last 20 years)14
What Works Clearinghouse Rating
Showing 1 to 15 of 33 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A.; Huber, Chuck – Measurement: Interdisciplinary Research and Perspectives, 2020
It is demonstrated that the popular three-parameter logistic model can lead to markedly inaccurate individual ability level estimates for mixture populations. A theoretically and empirically important setting is initially considered where (a) in one of two subpopulations (latent classes) the two-parameter logistic model holds for each item in a…
Descriptors: Item Response Theory, Models, Measurement Techniques, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Dietel, Ron – Kappa Delta Pi Record, 2012
Even as the Internet and i-devices have changed so many parts of people's lives, educational testing has not changed substantially in decades. Though researchers and educators have for years raised ample concerns about existing tests, assessments have barely evolved. Further, despite years of expanded testing and greater school accountability,…
Descriptors: Student Evaluation, Teacher Effectiveness, Accountability, National Competency Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Baker, Eva L.; O'Neil, Harry – Educational Psychologist, 2010
We describe some of Merl's more subtle contributions that extended beyond his own prodigious scholarly record, his high-quality teaching, and his mentoring of students. The context of our remarks on Programmatic R&D focuses on his contribution to the University of California, Los Angeles, Department of Education. Merl was the first director of…
Descriptors: Mentors, Educational Testing, Research and Development, Evaluation
Educational Testing Service, 2011
Choosing whether to test via computer is the most difficult and consequential decision the designers of a testing program can make. The decision is difficult because of the wide range of choices available. Designers can choose where and how often the test is made available, how the test items look and function, how those items are combined into…
Descriptors: Test Items, Testing Programs, Testing, Computer Assisted Testing
Scherrer, Jimmy – Phi Delta Kappan, 2012
The use of value-added modeling (VAM) in school accountability is expanding, but deciding how to embrace VAM is difficult. Various experts say it's too unreliable, causes more harm than good, and has a big margin for error. Others assert VAM is imperfect but useful, and provides valuable feedback. A closer look at the models, and their use,…
Descriptors: Teaching (Occupation), Accountability, Models, Data
Peer reviewed Peer reviewed
Direct linkDirect link
Low, Ee-Ling; Lee, Sing-Kong – Educational Research for Policy and Practice, 2012
In recent years, education systems around the world have been keeping a keen eye on rankings of student achievement as measured by internationally benchmarked tests. This has led to considerable attention being paid to teasing out success factors that may account for countries that have emerged top of the ranks or those that have shown the most…
Descriptors: Foreign Countries, Expertise, Teacher Education, Academic Achievement
Liu, Xiufeng – IAP - Information Age Publishing, Inc., 2010
This book meets a demand in the science education community for a comprehensive and introductory measurement book in science education. It describes measurement instruments reported in refereed science education research journals, and introduces the Rasch modeling approach to developing measurement instruments in common science assessment domains,…
Descriptors: Graduate Students, Textbooks, Research Methodology, Science Tests
Braun, Henry, Ed.; Chudowsky, Naomi, Ed.; Koenig, Judith, Ed. – National Academies Press, 2010
Value-added methods refer to efforts to estimate the relative contributions of specific teachers, schools, or programs to student test performance. In recent years, these methods have attracted considerable attention because of their potential applicability for educational accountability, teacher pay-for-performance systems, school and teacher…
Descriptors: Accountability, Teacher Improvement, Workshops, Program Evaluation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Regional Educational Laboratory Southeast, 2009
Since the passage of the No Child Left Behind Act of 2001 (2002), there has been increased interest in using student achievement data (through standardized tests) to evaluate teacher effectiveness. Two U.S. Department of Education secretaries, Secretary Spellings and Secretary Duncan, have expressed interest in growth models and the need to…
Descriptors: Evidence, Educational Research, Teacher Effectiveness, Teacher Evaluation
Herman, Joan L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2009
In this report, Joan Herman, director for the National Center for Research, on Evaluation, Standards, & Student Testing (CRESST) recommends that the new generation of science standards be based on lessons learned from current practice and on recent examples of standards-development methodology. In support of this, recent, promising efforts to…
Descriptors: Research and Development, Advanced Placement, Educational Research, State Standards
Peer reviewed Peer reviewed
Direct linkDirect link
Bartolucci, Francesco; Forcina, Antonio – Psychometrika, 2005
The assumptions underlying item response theory (IRT) models may be expressed as a set of equality and inequality constraints on the parameters of a latent class model. It is well known that the same assumptions imply that the parameters of the manifest distribution have to satisfy a more complicated set of inequality constraints which, however,…
Descriptors: Inferences, Educational Testing, Item Response Theory, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Von Davier, Matthias; Yamamoto, Kentaro – Applied Psychological Measurement, 2004
The generalized partial-credit model (GPCM) is used frequently in educational testing and in large-scale assessments for analyzing polytomous data. Special cases of the generalized partial-credit model are the partial-credit model--or Rasch model for ordinal data--and the two parameter logistic (2PL) model. This article extends the GPCM to the…
Descriptors: Educational Testing, Item Response Theory, Models, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Baker, Eva L. – Educational Assessment, 2007
This article describes the history, evidence warrants, and evolution of the Center for Research on Evaluation, Standards, and Student Testing's (CRESST) model-based assessments. It considers alternative interpretations of scientific or practical models and illustrates how model-based assessment addresses both definitions. The components of the…
Descriptors: Educational Testing, Computer Assisted Testing, Validity, Test Construction
Hashway, Robert M. – Educational Technology, 1979
Four testing schemes are outlined: post-test only, pretest/post-test, the Individually Prescribed Instruction (IPI) Model, and the EIPI Model which takes into account the learning style and the psychosocial background of the learner. The EIPI model has three phases: placement, counseling, and intervention. (RAO)
Descriptors: Educational Testing, Individual Testing, Individualized Instruction, Models
Fletcher, Jason; Wolfe, Barbara L. – National Bureau of Economic Research, 2007
Recently, Currie and Stabile (2006) made a significant contribution to our understanding of the influence of ADHD symptoms on a variety of school outcomes including participation in special education, grade repetition and test scores. Their contributions include using a broad sample of children and estimating sibling fixed effects models to…
Descriptors: Human Capital, Grade Repetition, Mental Health, Child Health
Previous Page | Next Page ยป
Pages: 1  |  2  |  3