Publication Date
| In 2026 | 0 |
| Since 2025 | 38 |
| Since 2022 (last 5 years) | 225 |
| Since 2017 (last 10 years) | 570 |
| Since 2007 (last 20 years) | 1377 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 110 |
| Practitioners | 107 |
| Teachers | 46 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 61 |
| Canada | 60 |
| United States | 57 |
| Turkey | 47 |
| Australia | 43 |
| Florida | 34 |
| Germany | 26 |
| Texas | 26 |
| China | 25 |
| Netherlands | 25 |
| Iran | 22 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Rikoon, Samuel H.; Liebtag, Travis; Olivera-Aguilar, Margarita; Steinberg, Jonathan; Robbins, Steven B. – ETS Research Report Series, 2015
In this report, we describe the development of an extension of the "SuccessNavigator"® assessment for late high school settings. We discuss the assessment's conceptualization and support its application with psychometric studies detailing scale development in terms of structural analyses, reliability, and several other aspects of…
Descriptors: College Preparation, College Attendance, High School Students, Test Construction
New York State Education Department, 2016
This technical report provides detailed information regarding the technical, statistical, and measurement attributes of the New York State Testing Program (NYSTP) for the Grades 3-8 Common Core English Language Arts (ELA) and Mathematics 2016 Operational Tests. This report includes information about test content and test development, item (i.e.,…
Descriptors: Testing Programs, English, Language Arts, Mathematics Tests
Dlaska, Andrea; Krekeler, Christian – Language Learning in Higher Education, 2012
This paper draws on previous research to identify a specific set of criteria for evaluating assessments for learning which are not trialled and for which the teacher is the only rater. The paper suggests that four criteria be used for the evaluation of assessment for learning in the language classroom: impact, fairness, activity, and feedback. The…
Descriptors: Student Evaluation, Evaluation Methods, College Faculty, Second Language Instruction
Lakin, Joni M.; Elliott, Diane Cardenas; Liu, Ou Lydia – Educational and Psychological Measurement, 2012
Outcomes assessments are gaining great attention in higher education because of increased demand for accountability. These assessments are widely used by U.S. higher education institutions to measure students' college-level knowledge and skills, including students who speak English as a second language (ESL). For the past decade, the increasing…
Descriptors: College Outcomes Assessment, Achievement Tests, English Language Learners, College Students
Beretvas, S. Natasha; Cawthon, Stephanie W.; Lockhart, L. Leland; Kaye, Alyssa D. – Educational and Psychological Measurement, 2012
This pedagogical article is intended to explain the similarities and differences between the parameterizations of two multilevel measurement model (MMM) frameworks. The conventional two-level MMM that includes item indicators and models item scores (Level 1) clustered within examinees (Level 2) and the two-level cross-classified MMM (in which item…
Descriptors: Test Bias, Comparative Analysis, Test Items, Difficulty Level
Liu, Yan; Zumbo, Bruno D. – Educational and Psychological Measurement, 2012
There is a lack of research on the effects of outliers on the decisions about the number of factors to retain in an exploratory factor analysis, especially for outliers arising from unintended and unknowingly included subpopulations. The purpose of the present research was to investigate how outliers from an unintended and unknowingly included…
Descriptors: Factor Analysis, Factor Structure, Evaluation Research, Evaluation Methods
Chae, Soo Eun; Kim, Doyoung; Han, Jae-Ho – IEEE Transactions on Education, 2012
Those items or test characteristics that are likely to result in differential item functioning (DIF) across accommodated test forms in statewide tests have received little attention. An examination of elementary-level student performance across accommodated test forms in a large-scale mathematics assessment revealed DIF variations by grades,…
Descriptors: Test Bias, Mathematics Tests, Testing Accommodations, Elementary School Mathematics
Ho, Ya-Ting – Online Submission, 2014
There is a continuing increase in the African American and Hispanic student populations in public schools. The students who are invited to gifted programs are overwhelmingly White. This is the situation in schools in the United States and also in Taiwan. Misunderstanding or unawareness of culture difference among educators might contribute to…
Descriptors: Foreign Countries, Cross Cultural Studies, Interviews, Academically Gifted
Moses, Tim; Miao, Jing; Dorans, Neil J. – Journal of Educational and Behavioral Statistics, 2010
In this study, the accuracies of four strategies were compared for estimating conditional differential item functioning (DIF), including raw data, logistic regression, log-linear models, and kernel smoothing. Real data simulations were used to evaluate the estimation strategies across six items, DIF and No DIF situations, and four sample size…
Descriptors: Test Bias, Statistical Analysis, Computation, Comparative Analysis
Penfield, Randall D. – Journal of Educational Measurement, 2010
In this article, I address two competing conceptions of differential item functioning (DIF) in polytomously scored items. The first conception, referred to as net DIF, concerns between-group differences in the conditional expected value of the polytomous response variable. The second conception, referred to as global DIF, concerns the conditional…
Descriptors: Test Bias, Test Items, Evaluation Methods, Item Response Theory
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
Using data from a large-scale exam, in this study we compared various designs for equating constructed-response (CR) tests to determine which design was most effective in producing equivalent scores across the two tests to be equated. In the context of classical equating methods, four linking designs were examined: (a) an anchor set containing…
Descriptors: Equated Scores, Responses, Tests, Measurement
Reddy, Linda A.; Fabiano, Gregory; Dudek, Christopher M.; Hsu, Louis – School Psychology Quarterly, 2013
Research on progress monitoring has almost exclusively focused on student behavior and not on teacher practices. This article presents the development and validation of a new teacher observational assessment (Classroom Strategies Scale) of classroom instructional and behavioral management practices. The theoretical underpinnings and empirical…
Descriptors: Test Construction, Construct Validity, Test Validity, Observation
Aslanides, J. S.; Savage, C. M. – Physical Review Special Topics - Physics Education Research, 2013
We report on a concept inventory for special relativity: the development process, data analysis methods, and results from an introductory relativity class. The Relativity Concept Inventory tests understanding of relativistic concepts. An unusual feature is confidence testing for each question. This can provide additional information; for example,…
Descriptors: Physics, Science Tests, Scientific Concepts, Confidence Testing
Nankervis, Bryan – Journal of College Admission, 2013
This article investigates gender inequity in the National Merit Scholarship Competition. Results suggest the competition favors males due to their higher mean score on the mathematics section and greater variability on all sections of the PSAT, which shares differential validity concerns with the SAT in terms of gender. These instruments are…
Descriptors: Gender Bias, Scholarships, Scores, College Entrance Examinations
Gerritson, Michael – ProQuest LLC, 2013
The practice of grading student writing often seems arbitrary and unfair to students. It is important to students and educators that writing is fairly and accurately assessed in order to facilitate demonstrable improvement in their composition. This research addressed a specific writing rubric as a method to mitigate implicit or subconscious…
Descriptors: Grading, Scoring Rubrics, Writing Assignments, Writing (Composition)

Peer reviewed
Direct link
