Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 7 |
Since 2006 (last 20 years) | 17 |
Descriptor
Scores | 18 |
Test Items | 18 |
Mathematics Tests | 10 |
Test Construction | 6 |
Grade 4 | 5 |
Grade 5 | 5 |
Regression (Statistics) | 5 |
Elementary School Students | 4 |
English (Second Language) | 4 |
Test Validity | 4 |
Differences | 3 |
More ▼ |
Source
Educational Assessment | 18 |
Author
Solano-Flores, Guillermo | 3 |
Li, Min | 2 |
Barnett-Clarke, Carne | 1 |
Braslow, David | 1 |
Briggs, Derek C. | 1 |
Bulut, Hatice Cigdem | 1 |
Bulut, Okan | 1 |
Buono, Stephanie | 1 |
Cormier, Damien C. | 1 |
Custer, Michael | 1 |
Dadey, Nathan | 1 |
More ▼ |
Publication Type
Journal Articles | 18 |
Reports - Research | 14 |
Reports - Evaluative | 3 |
Reports - Descriptive | 1 |
Tests/Questionnaires | 1 |
Education Level
Elementary Education | 8 |
Grade 4 | 6 |
Intermediate Grades | 6 |
Middle Schools | 6 |
Grade 5 | 5 |
Grade 8 | 4 |
Junior High Schools | 4 |
Secondary Education | 4 |
Grade 6 | 3 |
Grade 7 | 3 |
Elementary Secondary Education | 2 |
More ▼ |
Audience
Location
Massachusetts | 2 |
Oregon | 2 |
Alabama | 1 |
Canada | 1 |
Idaho | 1 |
Indiana | 1 |
Kansas | 1 |
Michigan | 1 |
Minnesota | 1 |
Missouri | 1 |
New Jersey | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Bulut, Okan; Bulut, Hatice Cigdem; Cormier, Damien C.; Ilgun Dibek, Munevver; Sahin Kursad, Merve – Educational Assessment, 2023
Some statewide testing programs allow students to receive corrective feedback and revise their answers during testing. Despite its pedagogical benefits, the effects of providing revision opportunities remain unknown in the context of alternate assessments. Therefore, this study examined student data from a large-scale alternate assessment that…
Descriptors: Error Correction, Alternative Assessment, Feedback (Response), Multiple Choice Tests
Moon, Jung Aa; Keehner, Madeleine; Katz, Irvin R. – Educational Assessment, 2020
We investigated how item formats influence test takers' response tendencies under uncertainty. Adult participants solved content-equivalent math items in three formats: multiple-selection multiple-choice, grid with forced-choice (true-false) options, and grid with non-forced-choice options. Participants showed a greater tendency to commit (rather…
Descriptors: College Students, Test Wiseness, Test Format, Test Items
Walker, A. Adrienne; Jennings, Jeremy Kyle; Engelhard, George, Jr. – Educational Assessment, 2018
Individual person fit analyses provide important information regarding the validity of test score inferences for an "individual" test taker. In this study, we use data from an undergraduate statistics test (N = 1135) to illustrate a two-step method that researchers and practitioners can use to examine individual person fit. First, person…
Descriptors: Test Items, Test Validity, Scores, Statistics
Buono, Stephanie; Jang, Eunice Eunhee – Educational Assessment, 2021
Increasing linguistic diversity in classrooms has led researchers to examine the validity and fairness of standardized achievement tests, specifically concerning whether test score interpretations are free of bias and score use is fair for all students. This study examined whether mathematics achievement test items that contain complex language…
Descriptors: English Language Learners, Standardized Tests, Achievement Tests, Culture Fair Tests
Ralston, Nicole C.; Li, Min; Taylor, Catherine – Educational Assessment, 2018
Elementary school students often exhibit a variety of conceptions associated with algebraic thinking that their teachers fail to recognize or understand. It is crucial that elementary school teachers possess knowledge of the variety of student conceptions and also have abilities to address varying states of conceptions. Otherwise, students who are…
Descriptors: Elementary School Students, Student Evaluation, Mathematics Tests, Test Construction
Tannenbaum, Richard J.; Kannan, Priya – Educational Assessment, 2015
Angoff-based standard setting is widely used, especially for high-stakes licensure assessments. Nonetheless, some critics have claimed that the judgment task is too cognitively complex for panelists, whereas others have explicitly challenged the consistency in (replicability of) standard-setting outcomes. Evidence of consistency in item judgments…
Descriptors: Standard Setting (Scoring), Reliability, Scores, Licensing Examinations (Professions)
Briggs, Derek C.; Dadey, Nathan – Educational Assessment, 2015
This study focuses on an instance in which the mean grade-to-grade scale scores on a vertical scale showed evidence of common test items that do not get easier from one grade to the next. The issue was examined as part of a 2-day workshop in which participants were asked to predict the growth on all linking items used in the construction of…
Descriptors: Test Items, Grading, Scores, Scaling
Solano-Flores, Guillermo; Wang, Chao; Kachchaf, Rachel; Soltero-Gonzalez, Lucinda; Nguyen-Le, Khanh – Educational Assessment, 2014
We address valid testing for English language learners (ELLs)--students in the United States who are schooled in English while they are still acquiring English as a second language. Also, we address the need for procedures for systematically developing ELL testing accommodations--changes in tests intended to support ELLs to gain access to the…
Descriptors: English Language Learners, Testing Accommodations, Illustrations, Educational Testing
Koretz, Daniel; Jennings, Jennifer L.; Ng, Hui Leng; Yu, Carol; Braslow, David; Langi, Meredith – Educational Assessment, 2016
Test-based accountability often produces score inflation. Most studies have evaluated inflation by comparing trends on a high-stakes test and a lower stakes audit test. However, Koretz and Beguin (2010) noted weaknesses of audit tests and suggested self-monitoring assessments (SMAs), which incorporate audit items into high-stakes tests. This…
Descriptors: Audits (Verification), Scores, Grade Inflation, Self Evaluation (Individuals)
Traynor, Anne – Educational Assessment, 2017
Variation in test performance among examinees from different regions or national jurisdictions is often partially attributed to differences in the degree of content correspondence between local school or training program curricula, and the test of interest. This posited relationship between test-curriculum correspondence, or "alignment,"…
Descriptors: Test Items, Test Construction, Alignment (Education), Curriculum
Solano-Flores, Guillermo; Barnett-Clarke, Carne; Kachchaf, Rachel R. – Educational Assessment, 2013
We examined the performance of English language learners (ELLs) and non-ELLs on Grade 4 and Grade 5 mathematics content knowledge (CK) and academic language (AL) tests. CK and AL items had different semiotic loads (numbers of different types of semiotic features) and different semiotic structures (relative frequencies of different semiotic…
Descriptors: English Language Learners, Performance, Mathematics Tests, Semiotics
Pae, Hye K. – Educational Assessment, 2014
This study investigated the role of item formats in the performance of 206 nonnative speakers of English on expressive skills (i.e., speaking and writing). Test scores were drawn from the field test of the "Pearson Test of English Academic" for Chinese, French, Hebrew, and Korean native speakers. Four item formats, including…
Descriptors: Test Items, Test Format, Speech Skills, Writing Skills
Kim, Do-Hong; Huynh, Huynh – Educational Assessment, 2010
This study investigated whether scores obtained from the online and paper-and-pencil administrations of the statewide end-of-course English test were equivalent for students with and without disabilities. Score comparability was evaluated by examining equivalence of factor structure (measurement invariance) and differential item and bundle…
Descriptors: Computer Assisted Testing, Language Tests, English, Scores
Ketterlin-Geller, Leanne R.; Yovanoff, Paul; Jung, EunJu; Liu, Kimy; Geller, Josh – Educational Assessment, 2013
In this article, we highlight the need for a precisely defined construct in score-based validation and discuss the contribution of cognitive theories to accurately and comprehensively defining the construct. We propose a framework for integrating cognitively based theoretical and empirical evidence to specify and evaluate the construct. We apply…
Descriptors: Test Validity, Construct Validity, Scores, Evidence
Wolf, Mikyung Kim; Leon, Seth – Educational Assessment, 2009
The purpose of the present study is to examine the language characteristics of a few states' large-scale assessments of mathematics and science and investigate whether the language demands of the items are associated with the degree of differential item functioning (DIF) for English language learner (ELL) students. A total of 542 items from 11…
Descriptors: Mathematics Tests, Science Tests, Measurement, Test Bias
Previous Page | Next Page »
Pages: 1 | 2