NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 11 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L.; Im, Sukkeun; Lee, Jay – Educational Assessment, 2021
This study investigated test-taking engagement on the Spring 2019 administration of a large-scale state summative assessment. Through the identification of rapid-guessing behavior -- which is a validated indicator of disengagement -- the percentage of Grade 8 test events with meaningful amounts of rapid guessing was 5.5% in mathematics, 6.7% in…
Descriptors: Accountability, Test Results, Guessing (Tests), Summative Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L.; Kuhfeld, Megan R.; Cronin, John – Educational Assessment, 2022
The arrival of the COVID-19 pandemic had a profound effect on K-12 education. Most schools transitioned to remote instruction, and some used remote testing to assess student learning. Remote testing, however, is less controlled than in-school testing, leading to concerns regarding test-taking engagement. This study compared the disengagement of…
Descriptors: Computer Assisted Testing, COVID-19, Pandemics, Learner Engagement
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L.; Kuhfeld, Megan R.; Soland, James – Applied Measurement in Education, 2019
When we administer educational achievement tests, we want to be confident that the resulting scores validly indicate what the test takers know and can do. However, if the test is perceived as low stakes by the test taker, disengaged test taking sometimes occurs, which poses a serious threat to score validity. When computer-based tests are used,…
Descriptors: Guessing (Tests), Computer Assisted Testing, Achievement Tests, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L. – Measurement: Interdisciplinary Research and Perspectives, 2015
The growing presence of computer-based testing has brought with it the capability to routinely capture the time that test takers spend on individual test items. This, in turn, has led to an increased interest in potential applications of response time in measuring intellectual ability and achievement. Goldhammer (this issue) provides a very useful…
Descriptors: Reaction Time, Measurement, Computer Assisted Testing, Achievement Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Kingsbury, G. Gage; Wise, Steven L. – Journal of Applied Testing Technology, 2011
Development of adaptive tests used in K-12 settings requires the creation of stable measurement scales to measure the growth of individual students from one grade to the next, and to measure change in groups from one year to the next. Accountability systems like No Child Left Behind require stable measurement scales so that accountability has…
Descriptors: Elementary Secondary Education, Adaptive Testing, Academic Achievement, Measures (Individuals)
Finney, Sara J.; Smith, Russell W.; Wise, Steven L. – 1999
Two operational item pools were used to investigate the performance of stratum computerized adaptive tests (CATs) when items were assigned to strata based on empirical estimates of item difficulty or human judgments of item difficulty. Items from the first data set consisted of 54 5-option multiple choice items from a form of the ACT mathematics…
Descriptors: Adaptive Testing, Classification, Computer Assisted Testing, High School Students
Wise, Steven L.; Owens, Kara M.; Yang, Sheng-Ta; Weiss, Brandi; Kissel, Hilary L.; Kong, Xiaojing; Horst, Sonia J. – Online Submission, 2005
There are a variety of situations in which low-stakes achievement tests--which are defined as those having few or no consequences for examinee performance--are used in applied measurement. A problem inherent in such testing is that we often cannot assume that all examinees give their best effort to their test, which suggests that the test scores…
Descriptors: Psychology, Mathematics Tests, Reaction Time, Achievement Tests
PDF pending restoration PDF pending restoration
Wise, Steven L.; And Others – 1993
A new testing strategy that provides protection against the problem of having examinees in adaptive testing choose difficulty levels that are not matched to their proficiency levels was introduced and evaluated. The method, termed restricted self-adapted testing (RSAT), still provides examinees with a degree of control over the difficulty levels…
Descriptors: Achievement Tests, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Wise, Steven L.; And Others – 1993
This study assessed whether providing examinees with a choice between computerized adaptive testing (CAT) and self-adaptive testing (SAT) affects test performance in comparison with being assigned a CAT or SAT, and evaluated variables influencing examinee choice of either test form. The relative influences of test type and test choice on examinee…
Descriptors: Ability, Adaptive Testing, Algebra, College Students
Plake, Barbara S.; Wise, Steven L. – 1986
One question regarding the utility of adaptive testing is the effect of individualized item arrangements on examinee test scores. The purpose of this study was to analyze the item difficulty choices by examinees as a function of previous item performance. The examination was a 25-item test of basic algebra skills given to 36 students in an…
Descriptors: Adaptive Testing, Algebra, College Students, Computer Assisted Testing
Peer reviewed Peer reviewed
Wise, Steven L.; And Others – Journal of Educational Measurement, 1992
Performance of 156 undergraduate and 48 graduate students on a self-adapted test (SFAT)--students choose the difficulty level of their test items--was compared with performance on a computer-adapted test (CAT). Those taking the SFAT obtained higher ability scores and reported lower posttest state anxiety than did CAT takers. (SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Difficulty Level