Publication Date
| In 2026 | 0 |
| Since 2025 | 222 |
| Since 2022 (last 5 years) | 1091 |
| Since 2017 (last 10 years) | 2601 |
| Since 2007 (last 20 years) | 4962 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 227 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Kaliski, Pamela K.; Wind, Stefanie A.; Engelhard, George, Jr.; Morgan, Deanna L.; Plake, Barbara S.; Reshetar, Rosemary A. – Educational and Psychological Measurement, 2013
The many-faceted Rasch (MFR) model has been used to evaluate the quality of ratings on constructed response assessments; however, it can also be used to evaluate the quality of judgments from panel-based standard setting procedures. The current study illustrates the use of the MFR model for examining the quality of ratings obtained from a standard…
Descriptors: Item Response Theory, Models, Standard Setting (Scoring), Science Tests
DeStefano, Lizanne; Johnson, Jeremiah – American Institutes for Research, 2013
This paper describes one of the first efforts by the National Assessment of Educational Progress (NAEP) to improve measurement at the lower end of the distribution, including measurement for students with disabilities (SD) and English language learners (ELLs). One way to improve measurement at the lower end is to introduce one or more…
Descriptors: National Competency Tests, Measures (Individuals), Disabilities, English Language Learners
Rogers, Angela – Mathematics Education Research Group of Australasia, 2013
As we move into the 21st century, educationalists are exploring the myriad of possibilities associated with Computer Based Assessment (CBA). At first glance this mode of assessment seems to provide many exciting opportunities in the mathematics domain, yet one must question the validity of CBA and whether our school systems, students and teachers…
Descriptors: Mathematics Tests, Student Evaluation, Computer Assisted Testing, Test Validity
Terry, Rodney L.; Fond, Marissa – Hispanic Journal of Behavioral Sciences, 2013
People of Hispanic origin, including monolingual Spanish speakers, have experienced difficulty identifying with a race category on U.S. demographic surveys. As part of a larger research effort by the U.S. Census Bureau to improve race and Hispanic origin questions for the 2020 Census, we tested experimental versions of race and Hispanic origin…
Descriptors: Census Figures, Experimental Programs, Spanish Speaking, Demography
Cho, Yeonsuk; Rijmen, Frank; Novák, Jakub – Language Testing, 2013
This study examined the influence of prompt characteristics on the averages of all scores given to test taker responses on the TOEFL iBT[TM] integrated Read-Listen-Write (RLW) writing tasks for multiple administrations from 2005 to 2009. In the context of TOEFL iBT RLW tasks, the prompt consists of a reading passage and a lecture. To understand…
Descriptors: English (Second Language), Language Tests, Writing Tests, Cues
Zu, Jiyun; Liu, Jinghua – Journal of Educational Measurement, 2010
Equating of tests composed of both discrete and passage-based multiple choice items using the nonequivalent groups with anchor test design is popular in practice. In this study, we compared the effect of discrete and passage-based anchor items on observed score equating via simulation. Results suggested that an anchor with a larger proportion of…
Descriptors: Equated Scores, Test Items, Multiple Choice Tests, Comparative Analysis
Penfield, Randall D. – Journal of Educational Measurement, 2010
In this article, I address two competing conceptions of differential item functioning (DIF) in polytomously scored items. The first conception, referred to as net DIF, concerns between-group differences in the conditional expected value of the polytomous response variable. The second conception, referred to as global DIF, concerns the conditional…
Descriptors: Test Bias, Test Items, Evaluation Methods, Item Response Theory
Raykov, Tenko – Measurement: Interdisciplinary Research and Perspectives, 2010
Mroch, Suh, Kane, & Ripkey (2009); Suh, Mroch, Kane, & Ripkey (2009); and Kane, Mroch, Suh, & Ripkey (2009) provided elucidating discussions on critical properties of linear equating methods under the nonequivalent groups with anchor test (NEAT) design. In this popular equating design, two test forms are administered to different…
Descriptors: Equated Scores, Test Items, Factor Analysis, Models
Chen, Shu-Ying – Applied Psychological Measurement, 2010
To date, exposure control procedures that are designed to control test overlap in computerized adaptive tests (CATs) are based on the assumption of item sharing between pairs of examinees. However, in practice, examinees may obtain test information from more than one previous test taker. This larger scope of information sharing needs to be…
Descriptors: Computer Assisted Testing, Adaptive Testing, Methods, Test Items
van der Linden, Wim J.; Glas, Cees A. W. – Psychometrika, 2010
Three plausible assumptions of conditional independence in a hierarchical model for responses and response times on test items are identified. For each of the assumptions, a Lagrange multiplier test of the null hypothesis of conditional independence against a parametric alternative is derived. The tests have closed-form statistics that are easy to…
Descriptors: Test Items, Computation, Responses, Reaction Time
Maryland State Department of Education, 2018
Based on Maryland's 2017-2018 Kindergarten Readiness Assessment (KRA) results, nearly half of all entering kindergarten children show foundational skills indicating they are fully ready for kindergarten, more than a third are approaching readiness, and 18% have emerging readiness skills. Results for the 2017-2018 school year show a slight increase…
Descriptors: Kindergarten, School Readiness, Academic Standards, Gender Differences
Haberman, Shelby J.; Sinharay, Sandip; Lee, Yi-Hsuan – Educational Testing Service, 2011
Providing information to test takers and test score users about the abilities of test takers at different score levels has been a persistent problem in educational and psychological measurement (Carroll, 1993). Scale anchoring (Beaton & Allen, 1992), a technique that describes what students at different points on a score scale know and can do,…
Descriptors: Statistical Analysis, Scores, Regression (Statistics), Item Response Theory
Carvajal-Espinoza, Jorge E. – ProQuest LLC, 2011
The Non-Equivalent groups with Anchor Test equating (NEAT) design is a widely used equating design in large scale testing that involves two groups that do not have to be of equal ability. One group P gets form X and a group of items A and the other group Q gets form Y and the same group of items A. One of the most commonly used equating methods in…
Descriptors: Sample Size, Equated Scores, Psychometrics, Measurement
Ling, Guangming; Rijmen, Frank – Educational Testing Service, 2011
The factorial structure of the Time Management (TM) scale of the Student 360: Insight Program (S360) was evaluated based on a national sample. A general procedure with a variety of methods was introduced and implemented, including the computation of descriptive statistics, exploratory factor analysis (EFA), and confirmatory factor analysis (CFA).…
Descriptors: Time Management, Measures (Individuals), Statistical Analysis, Factor Analysis
Brese, Falk; Jung, Michael; Mirazchiyski, Plamen; Schulz, Wolfram; Zuehlke, Olaf – International Association for the Evaluation of Educational Achievement, 2011
This document contains released items used in the ICCS 2009 main survey. Seven clusters of items were used in the study in a rotated booklet design with three clusters per test booklet. Clusters 1 and 7 comprise the released item set. Cluster 1 comprises items that were newly developed for ICCS 2009 and Cluster 7 is drawn from previously secure…
Descriptors: Foreign Countries, Civics, Citizenship Education, Academic Achievement

Peer reviewed
Direct link
