Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 14 |
Descriptor
Computation | 33 |
Testing Programs | 33 |
State Programs | 17 |
Test Items | 15 |
Mathematics Tests | 14 |
Problem Solving | 14 |
Geometry | 13 |
Behavioral Objectives | 10 |
Minimum Competency Testing | 9 |
Elementary School Mathematics | 8 |
Number Concepts | 8 |
More ▼ |
Source
Author
Cai, Li | 3 |
Carter, Betsy Y. | 3 |
Yang, Ji Seung | 3 |
Jaciw, Andrew P. | 2 |
Wyse, Adam E. | 2 |
Abari, Kálmán | 1 |
Babcock, Ben | 1 |
Biró, Piroska | 1 |
Chan, Tsze | 1 |
Chen, Hui-Fang | 1 |
Childs, Ruth A. | 1 |
More ▼ |
Publication Type
Reports - Research | 14 |
Guides - Classroom - Teacher | 11 |
Journal Articles | 11 |
Tests/Questionnaires | 6 |
Reports - Evaluative | 4 |
Numerical/Quantitative Data | 3 |
Reports - Descriptive | 2 |
Education Level
Secondary Education | 3 |
Elementary Education | 2 |
Elementary Secondary Education | 2 |
Grade 8 | 2 |
Junior High Schools | 2 |
Middle Schools | 2 |
Grade 6 | 1 |
Grade 7 | 1 |
Higher Education | 1 |
Intermediate Grades | 1 |
Postsecondary Education | 1 |
More ▼ |
Audience
Practitioners | 14 |
Teachers | 7 |
Researchers | 3 |
Location
Connecticut | 4 |
Arizona | 1 |
California | 1 |
Florida | 1 |
Hong Kong | 1 |
Hungary | 1 |
Missouri | 1 |
Oregon | 1 |
Romania | 1 |
United States | 1 |
Wisconsin | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Texas Educational Assessment… | 9 |
Program for International… | 5 |
College Level Academic Skills… | 1 |
Comprehensive Tests of Basic… | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Wyse, Adam E.; Babcock, Ben – Educational and Psychological Measurement, 2016
Continuously administered examination programs, particularly credentialing programs that require graduation from educational programs, often experience seasonality where distributions of examine ability may differ over time. Such seasonality may affect the quality of important statistical processes, such as item response theory (IRT) item…
Descriptors: Test Items, Item Response Theory, Computation, Licensing Examinations (Professions)
Tindal, Gerald; Nese, Joseph F. T.; Stevens, Joseph J. – Educational Assessment, 2017
For the past decade, the accountability model associated with No Child Left Behind (NCLB) emphasized proficiency on end of year tests; with Every Student Succeeds Act (ESSA) the emphasis on proficiency within statewide testing programs, though now integrated with other measures of student learning, nevertheless remains a primary metric for…
Descriptors: Testing Programs, Middle School Students, Models, State Standards
Csernoch, Mária; Biró, Piroska; Abari, Kálmán; Máth, János – Acta Didactica Napocensia, 2015
Within the framework of the Testing Algorithmic and Application Skills project we tested first year students of Informatics at the beginning of their tertiary education. We were focusing on the students' level of understanding in different programming environments. In the present paper we provide the results from the University of Debrecen, the…
Descriptors: Foreign Countries, Information Science Education, College Freshmen, Mathematics
Yang, Ji Seung; Cai, Li – Journal of Educational and Behavioral Statistics, 2014
The main purpose of this study is to improve estimation efficiency in obtaining maximum marginal likelihood estimates of contextual effects in the framework of nonlinear multilevel latent variable model by adopting the Metropolis-Hastings Robbins-Monro algorithm (MH-RM). Results indicate that the MH-RM algorithm can produce estimates and standard…
Descriptors: Computation, Hierarchical Linear Modeling, Mathematics, Context Effect
Yang, Ji Seung; Cai, Li – Grantee Submission, 2014
The main purpose of this study is to improve estimation efficiency in obtaining maximum marginal likelihood estimates of contextual effects in the framework of nonlinear multilevel latent variable model by adopting the Metropolis-Hastings Robbins-Monro algorithm (MH-RM; Cai, 2008, 2010a, 2010b). Results indicate that the MH-RM algorithm can…
Descriptors: Computation, Hierarchical Linear Modeling, Mathematics, Context Effect
Wang, Wen-Chung; Chen, Hui-Fang; Jin, Kuan-Yu – Educational and Psychological Measurement, 2015
Many scales contain both positively and negatively worded items. Reverse recoding of negatively worded items might not be enough for them to function as positively worded items do. In this study, we commented on the drawbacks of existing approaches to wording effect in mixed-format scales and used bi-factor item response theory (IRT) models to…
Descriptors: Item Response Theory, Test Format, Language Usage, Test Items
Yang, Ji Seung; Cai, Li – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2013
The main purpose of this study is to improve estimation efficiency in obtaining full-information maximum likelihood (FIML) estimates of contextual effects in the framework of a nonlinear multilevel latent variable model by adopting the Metropolis-Hastings Robbins-Monro algorithm (MH-RM; Cai, 2008, 2010a, 2010b). Results indicate that the MH-RM…
Descriptors: Context Effect, Computation, Hierarchical Linear Modeling, Mathematics
Keller, Lisa A.; Keller, Robert R. – Educational and Psychological Measurement, 2011
This article investigates the accuracy of examinee classification into performance categories and the estimation of the theta parameter for several item response theory (IRT) scaling techniques when applied to six administrations of a test. Previous research has investigated only two administrations; however, many testing programs equate tests…
Descriptors: Item Response Theory, Scaling, Sustainability, Classification
Paek, Insu; Guo, Hongwen – Applied Psychological Measurement, 2011
This study examined how much improvement was attainable with respect to accuracy of differential item functioning (DIF) measures and DIF detection rates in the Mantel-Haenszel procedure when employing focal and reference groups with notably unbalanced sample sizes where the focal group has a fixed small sample which does not satisfy the minimum…
Descriptors: Test Bias, Accuracy, Reference Groups, Investigations
Olsen, Robert B.; Unlu, Fatih; Price, Cristofer; Jaciw, Andrew P. – National Center for Education Evaluation and Regional Assistance, 2011
This report examines the differences in impact estimates and standard errors that arise when these are derived using state achievement tests only (as pre-tests and post-tests), study-administered tests only, or some combination of state- and study-administered tests. State tests may yield different evaluation results relative to a test that is…
Descriptors: Achievement Tests, Standardized Tests, State Standards, Reading Achievement
Haberman, Shelby J. – Journal of Educational and Behavioral Statistics, 2008
In educational tests, subscores are often generated from a portion of the items in a larger test. Guidelines based on mean squared error are proposed to indicate whether subscores are worth reporting. Alternatives considered are direct reports of subscores, estimates of subscores based on total score, combined estimates based on subscores and…
Descriptors: Testing Programs, Regression (Statistics), Scores, Student Evaluation
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Cohen, Jon; Chan, Tsze; Jiang, Tao; Seburn, Mary – Applied Psychological Measurement, 2008
U.S. state educational testing programs administer tests to track student progress and hold schools accountable for educational outcomes. Methods from item response theory, especially Rasch models, are usually used to equate different forms of a test. The most popular method for estimating Rasch models yields inconsistent estimates and relies on…
Descriptors: Testing Programs, Educational Testing, Item Response Theory, Computation
Childs, Ruth A.; Jaciw, Andrew P.; Saunders, Kelsey – International Journal of Testing, 2007
Many approaches to standard-setting use item calibration and student score estimation results to structure panelists' tasks. However, this requires collecting standard-setting judgments after the item analysis results are available. The Scoring Guide Alignment approach collects standard-setting judgments during the scoring sessions from teachers…
Descriptors: Testing Programs, Scoring, Item Analysis, Test Items
Tal, Joseph – 1987
An experimental test battery (the Johnson O'Connor Research Foundation battery) designed to measure numerical facility was administered to 1,451 subjects at 12 testing centers across the United States over a 5-month period. Five work samples were included: (1) arithmetic; (2) counting backwards; (3) number reasoning; (4) rule learning; and (5)…
Descriptors: Aptitude Tests, Arithmetic, Computation, Factor Analysis