Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 5 |
Since 2016 (last 10 years) | 11 |
Since 2006 (last 20 years) | 15 |
Descriptor
Test Items | 45 |
Difficulty Level | 16 |
Higher Education | 13 |
Test Construction | 13 |
Multiple Choice Tests | 11 |
Test Format | 11 |
Foreign Countries | 9 |
Test Reliability | 9 |
Cognitive Processes | 7 |
Scores | 7 |
Test Validity | 7 |
More ▼ |
Source
Journal of Experimental… | 45 |
Author
Plake, Barbara S. | 3 |
Bulut, Okan | 2 |
Gierl, Mark J. | 2 |
Weiten, Wayne | 2 |
Alexander, Patricia A. | 1 |
Ali, Usama S. | 1 |
Allison Ames | 1 |
Arce-Ferrer, Alvaro J. | 1 |
Attali, Yigal | 1 |
Bateson, David J. | 1 |
Berk, Ronald A. | 1 |
More ▼ |
Publication Type
Journal Articles | 42 |
Reports - Research | 40 |
Reports - Evaluative | 3 |
Education Level
Higher Education | 8 |
Postsecondary Education | 5 |
High Schools | 4 |
Secondary Education | 4 |
Grade 10 | 1 |
Grade 9 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
COMPASS (Computer Assisted… | 1 |
Graduate Record Examinations | 1 |
Motivated Strategies for… | 1 |
SAT (College Admission Test) | 1 |
State Trait Anxiety Inventory | 1 |
Watson Glaser Critical… | 1 |
What Works Clearinghouse Rating
James D. Weese; Ronna C. Turner; Allison Ames; Xinya Liang; Brandon Crawford – Journal of Experimental Education, 2024
In this study a standardized effect size was created for use with the SIBTEST procedure. Using this standardized effect size, a single set of heuristics was developed that are appropriate for data fitting different item response models (e.g., 2-parameter logistic, 3-parameter logistic). The standardized effect size rescales the raw beta-uni value…
Descriptors: Test Bias, Test Items, Item Response Theory, Effect Size
Schweizer, Karl; Wang, Tengfei; Ren, Xuezhu – Journal of Experimental Education, 2022
The essay reports two studies on confirmatory factor analysis of speeded data with an effect of selective responding. This response strategy leads test takers to choose their own working order instead of completing the items along with the given order. Methods for detecting speededness despite such a deviation from the given order are proposed and…
Descriptors: Factor Analysis, Response Style (Tests), Decision Making, Test Items
van Rijn, Peter W.; Attali, Yigal; Ali, Usama S. – Journal of Experimental Education, 2023
We investigated whether and to what extent different scoring instructions, timing conditions, and direct feedback affect performance and speed. An experimental study manipulating these factors was designed to address these research questions. According to the factorial design, participants were randomly assigned to one of twelve study conditions.…
Descriptors: Scoring, Time, Feedback (Response), Performance
Bieleke, Maik; Goetz, Thomas; Krannich, Maike; Roos, Anna-Lena; Yanagida, Takuya – Journal of Experimental Education, 2023
Tests in educational contexts often start with easy tasks, assuming that this fosters positive experiences--a sense of control, higher valuing of the test, and more positive and less negative emotions. Although intuitive and widespread, this assumption lacks an empirical basis and a theoretical framework. We conducted a field experiment and…
Descriptors: Foreign Countries, Secondary School Students, Mathematics Tests, Test Construction
Gladstone, Jessica R.; Morell, Monica; Yang, Ji Seung; Ponnock, Annette; Turci Faust, Lara; Wigfield, Allan – Journal of Experimental Education, 2023
Researchers developing questionnaire measures of personality, motivation, and self-regulation constructs related to students' achievement and persistence in STEM or other fields rarely have examined whether the items on the measures used are functioning differently across groups, which is necessary for accurate group comparison. The present study…
Descriptors: Test Bias, STEM Education, Test Items, Student Characteristics
Shin, Jinnie; Bulut, Okan; Gierl, Mark J. – Journal of Experimental Education, 2020
The arrangement of response options in multiple-choice (MC) items, especially the location of the most attractive distractor, is considered critical in constructing high-quality MC items. In the current study, a sample of 496 undergraduate students taking an educational assessment course was given three test forms consisting of the same items but…
Descriptors: Foreign Countries, Undergraduate Students, Multiple Choice Tests, Item Response Theory
Wang, Xiaolin; Svetina, Dubravka; Dai, Shenghai – Journal of Experimental Education, 2019
Recently, interest in test subscore reporting for diagnosis purposes has been growing rapidly. The two simulation studies here examined factors (sample size, number of subscales, correlation between subscales, and three factors affecting subscore reliability: number of items per subscale, item parameter distribution, and data generating model)…
Descriptors: Value Added Models, Scores, Sample Size, Correlation
Roelle, Julian; Roelle, Detlev; Berthold, Kirsten – Journal of Experimental Education, 2019
Providing test questions after an initial study phase is a common instructional technique. In theory, questions that require higher-level (deep) processing should be more beneficial than those that require lower-level (shallow) processing. However, empirical evidence on the matter is inconsistent. To shed light on two potential reasons for these…
Descriptors: Testing Problems, Test Items, Cognitive Processes, Problem Based Learning
Arce-Ferrer, Alvaro J.; Bulut, Okan – Journal of Experimental Education, 2019
This study investigated the performance of four widely used data-collection designs in detecting test-mode effects (i.e., computer-based versus paper-based testing). The experimental conditions included four data-collection designs, two test-administration modes, and the availability of an anchor assessment. The test-level and item-level results…
Descriptors: Data Collection, Test Construction, Test Format, Computer Assisted Testing
Leroux, Audrey J.; Dodd, Barbara G. – Journal of Experimental Education, 2016
The current study compares the progressive-restricted standard error (PR-SE) exposure control method with the Sympson-Hetter, randomesque, and no exposure control (maximum information) procedures using the generalized partial credit model with fixed- and variable-length CATs and two item pools. The PR-SE method administered the entire item pool…
Descriptors: Computer Assisted Testing, Adaptive Testing, Comparative Analysis, Error of Measurement
DiBattista, David; Sinnige-Egger, Jo-Anne; Fortuna, Glenda – Journal of Experimental Education, 2014
The authors assessed the effects of using "none of the above" as an option in a 40-item, general-knowledge multiple-choice test administered to undergraduate students. Examinees who selected "none of the above" were given an incentive to write the correct answer to the question posed. Using "none of the above" as the…
Descriptors: Multiple Choice Tests, Testing, Undergraduate Students, Test Items
Alexander, Patricia A.; Dumas, Denis; Grossnickle, Emily M.; List, Alexandra; Firetto, Carla M. – Journal of Experimental Education, 2016
Relational reasoning is the foundational cognitive ability to discern meaningful patterns within an informational stream, but its reliable and valid measurement remains problematic. In this investigation, the measurement of relational reasoning unfolded in three stages. Stage 1 entailed the establishment of a research-based conceptualization of…
Descriptors: Cognitive Ability, Logical Thinking, Thinking Skills, Cognitive Processes
Bonner, Sarah M. – Journal of Experimental Education, 2013
Although test scores from similar tests in multiple choice and constructed response formats are highly correlated, equivalence in rankings may mask differences in substantive strategy use. The author used an experimental design and participant think-alouds to explore cognitive processes in mathematical problem solving among undergraduate examinees…
Descriptors: Scores, Multiple Choice Tests, Correlation, Protocol Analysis
Wasylkiw, Louise; Tomes, Jennifer L.; Smith, Francine – Journal of Experimental Education, 2008
In 3 studies, the authors examined the prevalence and effects of a testing strategy whereby they gave a set of items to participants in advance and subsequently tested them on a portion of those items (i.e., subset testing). In a survey of university instructors, Study 1 showed that subset testing is a commonly used testing strategy. In this…
Descriptors: Undergraduate Students, Incidence, Definitions, Testing
Elosua, Paula; Lopez-Jauregui, Alicia – Journal of Experimental Education, 2008
The comparison of scores from linguistically different tests is a twofold matter: the adaptation of tests and the comparison of scores. These 2 aspects of measurement invariance intersect at the need to guarantee the psychometric equivalence between the original and adapted versions. In this study, the authors examined comparability in 2 stages.…
Descriptors: Psychometrics, Item Response Theory, Equated Scores, Comparative Analysis