Publication Date
In 2025 | 0 |
Since 2024 | 3 |
Since 2021 (last 5 years) | 23 |
Since 2016 (last 10 years) | 66 |
Since 2006 (last 20 years) | 227 |
Descriptor
Raw Scores | 452 |
Comparative Analysis | 111 |
Statistical Analysis | 96 |
Equated Scores | 94 |
Scores | 78 |
Standardized Tests | 71 |
Correlation | 69 |
Foreign Countries | 66 |
Reading Tests | 59 |
Test Items | 58 |
Item Response Theory | 56 |
More ▼ |
Source
Author
Bianchini, John C. | 31 |
Loret, Peter G. | 30 |
Livingston, Samuel A. | 8 |
Kim, Sooyeon | 7 |
Prediger, Dale J. | 7 |
Liu, Jinghua | 6 |
Kolen, Michael J. | 5 |
Puhan, Gautam | 5 |
Guo, Hongwen | 4 |
Gallas, Edwin J. | 3 |
Hanson, Gary R. | 3 |
More ▼ |
Publication Type
Education Level
Location
Canada | 7 |
Australia | 6 |
Turkey | 6 |
United States | 6 |
United Kingdom | 5 |
Florida | 4 |
Germany | 4 |
Illinois | 4 |
South Africa | 4 |
Texas | 4 |
United Kingdom (England) | 4 |
More ▼ |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 4 |
Elementary and Secondary… | 3 |
Assessments and Surveys
What Works Clearinghouse Rating
James D. Weese; Ronna C. Turner; Allison Ames; Xinya Liang; Brandon Crawford – Journal of Experimental Education, 2024
In this study a standardized effect size was created for use with the SIBTEST procedure. Using this standardized effect size, a single set of heuristics was developed that are appropriate for data fitting different item response models (e.g., 2-parameter logistic, 3-parameter logistic). The standardized effect size rescales the raw beta-uni value…
Descriptors: Test Bias, Test Items, Item Response Theory, Effect Size
Jessica Stinson – ProQuest LLC, 2024
Intelligence tests have been used in the United States since the early 1900s for assessing soldiers during World War I (Kaufman & Harrison, 2008; White & Hall, 1980). Presently, cognitive assessments are used in school, civil service, military, clinical, and industry settings (White & Hall, 1980). Although the results of these…
Descriptors: Graduate Students, Masters Programs, Doctoral Programs, Comparative Analysis
Li, Dongmei; Kapoor, Shalini – Educational Measurement: Issues and Practice, 2022
Population invariance is a desirable property of test equating which might not hold when significant changes occur in the test population, such as those brought about by the COVID-19 pandemic. This research aims to investigate whether equating functions are reasonably invariant when the test population is impacted by the pandemic. Based on…
Descriptors: Test Items, Equated Scores, COVID-19, Pandemics
Vecchini, Aurora; Buratta, Livia; Fogassi, Leonardo – Cogent Education, 2023
Imitation is a crucial process for learning and brain development. It is based on the mirror neuron mechanism and underlies our understanding of actions and the gestures of others. Some researchers hypothesized a possible correlation between a low functioning mirror neuron system (MNS) and developmental coordination disorder, including dysgraphic…
Descriptors: Handwriting, Difficulty Level, Nonverbal Communication, Imitation
Öztürk Gübes, Nese – Eurasian Journal of Educational Research, 2021
Purpose: In grading, one of the most common errors is made in combining two or more different test scores. This study aimed to investigate the agreement of grades calculated by weighting raw scores and standard scores. Research Methods: In this simulation study, data were simulated for midterm and final measurements. Nine conditions [3 (class…
Descriptors: Grading, Raw Scores, Weighted Scores, Norm Referenced Tests
Emily Relkin; Sara K. Johnson; Marina U. Bers – Educational Technology & Society, 2023
"TechCheck" is an assessment of Computational Thinking (CT) for early elementary school children consisting of fifteen developmentally appropriate unplugged challenges that probe six CT domains. The first version of "TechCheck" showed good psychometric properties as well as ease of administration and scoring in a validation…
Descriptors: Elementary School Students, Developmentally Appropriate Practices, Computation, Thinking Skills
Lu, Ru; Kim, Sooyeon – ETS Research Report Series, 2021
This study evaluated the impact of subgroup weighting for equating through a common-item anchor. We used data from a single test form to create two research forms for which the equating relationship was known. The results showed that equating was most accurate when the new form and reference form samples were weighted to be similar to the target…
Descriptors: Equated Scores, Weighted Scores, Raw Scores, Test Items
Arslan, Deniz; Sak, Ugur – Journal of Psychoeducational Assessment, 2023
The purpose of this study was to examine the factorial invariance of the ASIS (Anadolu Sak Intelligence Scale) across gender. Raw scores from the ASIS standardization study (N = 4641) were used in the analyses. Factorial invariance was analyzed by gender across three age groups: 4-7 (N = 2528), 8-9 (N = 848), and 10-12 (N = 1265) using the…
Descriptors: Gender Differences, Intelligence Tests, Factor Analysis, Raw Scores
Kim, Sooyeon; Walker, Michael E. – ETS Research Report Series, 2021
Equating the scores from different forms of a test requires collecting data that link the forms. Problems arise when the test forms to be linked are given to groups that are not equivalent and the forms share no common items by which to measure or adjust for this group nonequivalence. We compared three approaches to adjusting for group…
Descriptors: Equated Scores, Weighted Scores, Sampling, Multiple Choice Tests
The Relationship between Speech Perception and Speech Production in Children with Visual Impairments
Brouwer, Kyle; Gordon-Pershey, Monica – Journal of Visual Impairment & Blindness, 2021
The influence of visual impairment (i.e., blindness or low vision) on the development of children's speech sound production has undergone some speculation for over half a century. Children learn to produce speech sounds by utilizing and integrating several cognitive-linguistic and sensory-perceptual processes, which may include processing visual…
Descriptors: Auditory Perception, Speech Communication, Articulation (Speech), Speech Tests
Selcuk Acar; Lindsay E. Lee; Jaret Hodges – Creativity Research Journal, 2023
Numerous primary studies and a recent meta-analytic confirmatory factor analysis (Meta-CFA; Said-Metwaly, Fernández-Castilla, Kyndt, & Van den Noortgate, 2018) have shown that Torrance Tests of Creative Thinking -- Figural (TTCT-F) consists of two factors. However, recent research has raised questions regarding factor analysis of the TTCT-F…
Descriptors: Creativity, Creative Thinking, Creativity Tests, Factor Structure
Betts, Joe; Muntean, William; Kim, Doyoung; Kao, Shu-chuan – Educational and Psychological Measurement, 2022
The multiple response structure can underlie several different technology-enhanced item types. With the increased use of computer-based testing, multiple response items are becoming more common. This response type holds the potential for being scored polytomously for partial credit. However, there are several possible methods for computing raw…
Descriptors: Scoring, Test Items, Test Format, Raw Scores
Li, Daoxin; Schuler, Kathryn D. – Language Acquisition: A Journal of Developmental Linguistics, 2023
Languages differ regarding the depth, structure, and syntactic domains of recursive structures. Even within a single language, some structures allow infinite self-embedding while others are more restricted. For example, when expressing ownership relation, English allows infinite embedding of the prenominal genitive "-s," whereas the…
Descriptors: Language Acquisition, Linguistic Input, Artificial Languages, Learning Processes
Jin, Kuan-Yu; Wang, Wen-Chung – Journal of Educational Measurement, 2018
The Rasch facets model was developed to account for facet data, such as student essays graded by raters, but it accounts for only one kind of rater effect (severity). In practice, raters may exhibit various tendencies such as using middle or extreme scores in their ratings, which is referred to as the rater centrality/extremity response style. To…
Descriptors: Scoring, Models, Interrater Reliability, Computation
Kwok, Elaine; Feiner, Hannah; Grauzer, Jeffrey; Kaat, Aaron; Roberts, Megan Y. – Journal of Speech, Language, and Hearing Research, 2022
Purpose: Norm-referenced, standardized measures are tools designed to characterize a child's abilities relative to their same-age peers, but they also have been used to measure changes in skills during intervention. This study compared the psychometric properties of four types of available scores from one commonly used standardized measure, the…
Descriptors: Language Tests, Preschool Children, Norm Referenced Tests, Standardized Tests