NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 987 results Save | Export
Jiajing Huang – ProQuest LLC, 2022
The nonequivalent-groups anchor-test (NEAT) data-collection design is commonly used in large-scale assessments. Under this design, different test groups take different test forms. Each test form has its own unique items and all test forms share a set of common items. If item response theory (IRT) models are applied to analyze the test data, the…
Descriptors: Item Response Theory, Test Format, Test Items, Test Construction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Merchant, Stefan; Rich, Jessica; Klinger, Don A. – Canadian Journal of Educational Administration and Policy, 2022
Both school and district administrators use the results of standardized, large-scale tests to inform decisions about the need for, or success of, educational programs and interventions. However, test results at the school level are subject to random fluctuations due to changes in cohort, test items, and other factors outside of the school's…
Descriptors: Standardized Tests, Foreign Countries, Generalizability Theory, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Onur Demirkaya; Sharon Frey; Sid Sharairi; JongPil Kim – International Electronic Journal of Elementary Education, 2025
This study compares latent profiles derived from student subgroups of varying levels of mathematical skills defined by achievement and ability assessment scores. Achievement and ability cut scores for identifying students at both ends of the mathematics spectrum were applied and the resulting latent profiles within each condition were compared.…
Descriptors: Profiles, Statistical Analysis, Academic Achievement, Mathematics Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Mark Wilson – Journal of Educational and Behavioral Statistics, 2024
This article introduces a new framework for articulating how educational assessments can be related to teacher uses in the classroom. It articulates three levels of assessment: macro (use of standardized tests), meso (externally developed items), and micro (on-the-fly in the classroom). The first level is the usual context for educational…
Descriptors: Educational Assessment, Measurement, Standardized Tests, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Erbeli, Florina; He, Kai; Cheek, Connor; Rice, Marianne; Qian, Xiaoning – Scientific Studies of Reading, 2023
Purpose: Researchers have developed a constellation model of decodingrelated reading disabilities (RD) to improve the RD risk determination. The model's hallmark is its inclusion of various RD indicators to determine RD risk. Classification methods such as logistic regression (LR) might be one way to determine RD risk within the constellation…
Descriptors: At Risk Students, Reading Difficulties, Classification, Comparative Analysis
Bray, Karen – ProQuest LLC, 2023
The purpose of this quantitative study was to test the relationship between instructional related expenditures and student performance on a state mandated exam while controlling for various district-level and student-level variables for school districts in the state of Texas. The study had two research questions regarding student performance in…
Descriptors: Academic Achievement, Standardized Tests, Achievement Tests, Statistical Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Hao, Jiangang; Liu, Lei; Kyllonen, Patrick; Flor, Michael; von Davier, Alina A. – ETS Research Report Series, 2019
Collaborative problem solving (CPS) is an important 21st-century skill that is crucial for both career and academic success. However, developing a large-scale and standardized assessment of CPS that can be administered on a regular basis is very challenging. In this report, we introduce a set of psychometric considerations and a general scoring…
Descriptors: Scoring, Psychometrics, Cooperation, Problem Solving
Peer reviewed Peer reviewed
Direct linkDirect link
Tsaousis, Ioannis; Sideridis, Georgios; Al-Saawi, Fahad – International Journal of Testing, 2018
The aim of the present study was to examine Differential Distractor Functioning (DDF) as a means of improving the quality of a measure through understanding biased responses across groups. A DDF analysis could shed light on the potential sources of construct-irrelevant variance by examining whether the differential selection of incorrect choices…
Descriptors: Foreign Countries, College Entrance Examinations, Test Bias, Chemistry
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ssemakula, Mukasa E.; Liao, Gene Y.; Sawilowsky, Shlomo – American Journal of Engineering Education, 2018
There is a major trend in engineering education to provide students with realistic hands-on learning experiences. This paper reports on the results of work done to develop standardized test instruments to use for student learning outcomes assessment in an experiential hands-on manufacturing engineering and technology environment. The specific…
Descriptors: Test Construction, Psychometrics, Test Validity, Standardized Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Xueman Lucy; de Villiers, Jill; Ning, Chunyan; Rolfhus, Eric; Hutchings, Teresa; Lee, Wendy; Jiang, Fan; Zhang, Yi Wen – Journal of Speech, Language, and Hearing Research, 2017
Purpose: With no existing gold standard for comparison, challenges arise for establishing the validity of a new standardized Mandarin language assessment normed in mainland China. Method: A new assessment, Diagnostic Receptive and Expressive Assessment of Mandarin (DREAM), was normed with a stratified sample of 969 children ages 2;6 (years;months)…
Descriptors: Mandarin Chinese, Correlation, Language Tests, Diagnostic Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Lovett, Benjamin J.; Lewandowski, Lawrence J.; Potts, Heather E. – Journal of Psychoeducational Assessment, 2017
Students often feel time pressure when taking tests, and students with disabilities are sometimes given extended time testing accommodations, but little research has been done on the factors that affect students' test-taking speed. In the present study, 253 students at two colleges completed measures of processing speed, reading fluency, and…
Descriptors: Reading Fluency, Test Wiseness, Reading Skills, Standardized Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Assessment in Education: Principles, Policy & Practice, 2017
In response to an argument by Baird, Andrich, Hopfenbeck and Stobart (2017), Michael Kane states that there needs to be a better fit between educational assessment and learning theory. In line with this goal, Kane will examine how psychometric constraints might be loosened by relaxing some psychometric "rules" in some assessment…
Descriptors: Educational Assessment, Psychometrics, Standards, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Chow, Jason C.; Hollo, Alexandra – Assessment for Effective Intervention, 2018
Language impairment often goes unidentified in students with behavioral disorders, perhaps in part because different forms of problem behavior deflect adult attention from more subtle language deficits. Therefore, attention to teachers' perception of students' language and behavioral performance is merited. The present study examines agreement…
Descriptors: Emotional Disturbances, Language Proficiency, Language Impairments, Norm Referenced Tests
Konstantopoulos, Spyros; Li, Wei; Miller, Shazia; van der Ploeg, Arie – Educational and Psychological Measurement, 2019
This study discusses quantile regression methodology and its usefulness in education and social science research. First, quantile regression is defined and its advantages vis-à-vis vis ordinary least squares regression are illustrated. Second, specific comparisons are made between ordinary least squares and quantile regression methods. Third, the…
Descriptors: Regression (Statistics), Statistical Analysis, Educational Research, Social Science Research
Peer reviewed Peer reviewed
Direct linkDirect link
Huggins-Manley, Anne Corinne – Educational and Psychological Measurement, 2017
This study defines subpopulation item parameter drift (SIPD) as a change in item parameters over time that is dependent on subpopulations of examinees, and hypothesizes that the presence of SIPD in anchor items is associated with bias and/or lack of invariance in three psychometric outcomes. Results show that SIPD in anchor items is associated…
Descriptors: Psychometrics, Test Items, Item Response Theory, Hypothesis Testing
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  66