Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 9 |
Since 2006 (last 20 years) | 47 |
Descriptor
Test Items | 136 |
College Entrance Examinations | 107 |
Difficulty Level | 40 |
High School Students | 38 |
Scores | 33 |
Mathematics Tests | 32 |
Test Construction | 32 |
Test Bias | 31 |
High Schools | 29 |
Comparative Analysis | 26 |
Item Analysis | 26 |
More ▼ |
Source
Author
Dorans, Neil J. | 16 |
Liu, Jinghua | 10 |
Lawrence, Ida M. | 7 |
Eignor, Daniel R. | 6 |
Feigenbaum, Miriam | 6 |
Kostin, Irene | 6 |
Bridgeman, Brent | 5 |
Curley, Edward | 5 |
Freedle, Roy | 5 |
Gierl, Mark J. | 5 |
Kobrin, Jennifer L. | 5 |
More ▼ |
Publication Type
Education Level
Higher Education | 48 |
Postsecondary Education | 44 |
High Schools | 22 |
Secondary Education | 21 |
Elementary Secondary Education | 3 |
Grade 10 | 1 |
Grade 11 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Two Year Colleges | 1 |
Audience
Practitioners | 2 |
Researchers | 2 |
Administrators | 1 |
Teachers | 1 |
Laws, Policies, & Programs
Individuals with Disabilities… | 1 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Domingue, Benjamin W.; Kanopka, Klint; Stenhaug, Ben; Sulik, Michael J.; Beverly, Tanesia; Brinkhuis, Matthieu; Circi, Ruhan; Faul, Jessica; Liao, Dandan; McCandliss, Bruce; Obradovic, Jelena; Piech, Chris; Porter, Tenelle; Soland, James; Weeks, Jon; Wise, Steven L.; Yeatman, Jason – Journal of Educational and Behavioral Statistics, 2022
The speed-accuracy trade-off (SAT) suggests that time constraints reduce response accuracy. Its relevance in observational settings--where response time (RT) may not be constrained but respondent speed may still vary--is unclear. Using 29 data sets containing data from cognitive tasks, we use a flexible method for identification of the SAT (which…
Descriptors: Accuracy, Reaction Time, Task Analysis, College Entrance Examinations
McDaniel, Mark A.; Cahill, Michael J.; Frey, Regina F.; Limeri, Lisa B.; Lemons, Paula P. – CBE - Life Sciences Education, 2022
Previous studies have found that students' concept-building approaches, identified a priori with a cognitive psychology laboratory task, are associated with student exam performances in chemistry classes. Abstraction learners (those who extract the principles underlying related examples) performed better than exemplar learners (those who focus on…
Descriptors: Introductory Courses, Concept Formation, Scientific Concepts, Biology
College Board, 2023
Over the past several years, content experts, psychometricians, and researchers have been hard at work developing, refining, and studying the digital SAT. The work is grounded in foundational best practices and advances in measurement and assessment design, with fairness for students informing all of the work done. This paper shares learnings from…
Descriptors: College Entrance Examinations, Psychometrics, Computer Assisted Testing, Best Practices
Wedman, Jonathan – Scandinavian Journal of Educational Research, 2018
Gender fairness in testing can be impeded by the presence of differential item functioning (DIF), which potentially causes test bias. In this study, the presence and causes of gender-related DIF were investigated with real data from 800 items answered by 250,000 test takers. DIF was examined using the Mantel-Haenszel and logistic regression…
Descriptors: Gender Differences, College Entrance Examinations, Test Items, Vocabulary
Liu, Jinghua; Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2014
Maintaining score interchangeability and scale consistency is crucial for any testing programs that administer multiple forms across years. The use of a multiple linking design, which involves equating a new form to multiple old forms and averaging the conversions, has been proposed to control scale drift. However, the use of multiple linking…
Descriptors: Comparative Analysis, Reliability, Test Construction, Equated Scores
Rios, Joseph A.; Sparks, Jesse R.; Zhang, Mo; Liu, Ou Lydia – ETS Research Report Series, 2017
Proficiency with written communication (WC) is critical for success in college and careers. As a result, institutions face a growing challenge to accurately evaluate their students' writing skills to obtain data that can support demands of accreditation, accountability, or curricular improvement. Many current standardized measures, however, lack…
Descriptors: Test Construction, Test Validity, Writing Tests, College Outcomes Assessment
Attali, Yigal; Saldivia, Luis; Jackson, Carol; Schuppan, Fred; Wanamaker, Wilbur – ETS Research Report Series, 2014
Previous investigations of the ability of content experts and test developers to estimate item difficulty have, for themost part, produced disappointing results. These investigations were based on a noncomparative method of independently rating the difficulty of items. In this article, we argue that, by eliciting comparative judgments of…
Descriptors: Test Items, Difficulty Level, Comparative Analysis, College Entrance Examinations
Bacon, Donald R.; Hartley, Steven W. – Marketing Education Review, 2015
Many educators and researchers have suggested that some students learn more effectively with visual stimuli (e.g., pictures, graphs), whereas others learn more effectively with verbal information (e.g., text) (Felder & Brent, 2005). In two studies, the present research seeks to improve popular self-reported (indirect) learning style measures…
Descriptors: Cognitive Style, Test Items, Individual Differences, Mathematical Aptitude
Chubbuck, Kay; Curley, W. Edward; King, Teresa C. – ETS Research Report Series, 2016
This study gathered quantitative and qualitative evidence concerning gender differences in performance by using critical reading material on the "SAT"® test with sports and science content. The fundamental research questions guiding the study were: If sports and science are to be included in a skills test, what kinds of material are…
Descriptors: College Entrance Examinations, Gender Differences, Critical Reading, Reading Tests
Liu, Jinghua; Zu, Jiyun; Curley, Edward; Carey, Jill – ETS Research Report Series, 2014
The purpose of this study is to investigate the impact of discrete anchor items versus passage-based anchor items on observed score equating using empirical data.This study compares an "SAT"® critical reading anchor that contains more discrete items proportionally, compared to the total tests to be equated, to another anchor that…
Descriptors: Equated Scores, Test Items, College Entrance Examinations, Comparative Analysis
Alexander, Patricia A.; Dumas, Denis; Grossnickle, Emily M.; List, Alexandra; Firetto, Carla M. – Journal of Experimental Education, 2016
Relational reasoning is the foundational cognitive ability to discern meaningful patterns within an informational stream, but its reliable and valid measurement remains problematic. In this investigation, the measurement of relational reasoning unfolded in three stages. Stage 1 entailed the establishment of a research-based conceptualization of…
Descriptors: Cognitive Ability, Logical Thinking, Thinking Skills, Cognitive Processes
Liu, Ou Lydia; Mao, Liyang; Frankel, Lois; Xu, Jun – Assessment & Evaluation in Higher Education, 2016
Critical thinking is a learning outcome highly valued by higher education institutions and the workforce. The Educational Testing Service (ETS) has designed a next generation assessment, the HEIghten™ critical thinking assessment, to measure students' critical thinking skills in analytical and synthetic dimensions. This paper introduces the…
Descriptors: Critical Thinking, Cognitive Tests, Test Validity, Thinking Skills
Warne, Russell T.; Doty, Kristine J.; Malbica, Anne Marie; Angeles, Victor R.; Innes, Scott; Hall, Jared; Masterson-Nixon, Kelli – Journal of Psychoeducational Assessment, 2016
"Above-level testing" (also called "above-grade testing," "out-of-level testing," and "off-level testing") is the practice of administering to a child a test that is designed for an examinee population that is older or in a more advanced grade. Above-level testing is frequently used to help educators design…
Descriptors: Test Items, Testing, Academically Gifted, Talent Identification
Engelhard, George, Jr.; Kobrin, Jennifer L.; Wind, Stefanie A. – International Journal of Testing, 2014
The purpose of this study is to explore patterns in model-data fit related to subgroups of test takers from a large-scale writing assessment. Using data from the SAT, a calibration group was randomly selected to represent test takers who reported that English was their best language from the total population of test takers (N = 322,011). A…
Descriptors: College Entrance Examinations, Writing Tests, Goodness of Fit, English
Guo, Hongwen; Liu, Jinghua; Dorans, Neil; Feigenbaum, Miriam – ETS Research Report Series, 2011
Maintaining score stability is crucial for an ongoing testing program that administers several tests per year over many years. One way to stall the drift of the score scale is to use an equating design with multiple links. In this study, we use the operational and experimental SAT® data collected from 44 administrations to investigate the effect…
Descriptors: Equated Scores, College Entrance Examinations, Reliability, Testing Programs