Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 7 |
Descriptor
Source
Author
Reckase, Mark D. | 36 |
McKinley, Robert L. | 12 |
He, Wei | 2 |
Patience, Wayne M. | 2 |
Ackerman, Terry A. | 1 |
Ferrini-Mundy, Joan | 1 |
Floden, Robert E. | 1 |
Luo, Xin | 1 |
McCrory, Raven | 1 |
Papanastasiou, Elena C. | 1 |
Senk, Sharon L. | 1 |
More ▼ |
Publication Type
Reports - Research | 28 |
Speeches/Meeting Papers | 17 |
Journal Articles | 10 |
Reports - Evaluative | 8 |
Guides - Non-Classroom | 1 |
Information Analyses | 1 |
Opinion Papers | 1 |
Education Level
Audience
Researchers | 8 |
Location
Australia | 1 |
Canada | 1 |
China | 1 |
Hong Kong | 1 |
India | 1 |
Japan | 1 |
South Korea | 1 |
Taiwan | 1 |
United Kingdom | 1 |
United States | 1 |
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 3 |
National Assessment of… | 2 |
COMPASS (Computer Assisted… | 1 |
Iowa Tests of Educational… | 1 |
School and College Ability… | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating
Luo, Xin; Reckase, Mark D.; He, Wei – AERA Online Paper Repository, 2016
While dichotomous item dominates the application of computerized adaptive testing (CAT), polytomous item and set-based item hold promises for being incorporated in CAT. However, how to assemble a CAT containing mixed item formats is challenging. This study investigated: (1) how the mixed CAT works compared with the dichotomous-item-based CAT; (2)…
Descriptors: Test Items, Test Format, Computer Assisted Testing, Adaptive Testing
Reckase, Mark D.; McCrory, Raven; Floden, Robert E.; Ferrini-Mundy, Joan; Senk, Sharon L. – Educational Assessment, 2015
Numerous researchers have suggested that there are multiple mathematical knowledge and skill areas needed by teachers in order for them to be effective teachers of mathematics: knowledge of the mathematics that are the goals of instruction, advanced mathematics beyond the instructional material, and mathematical knowledge that is specific to what…
Descriptors: Algebra, Knowledge Base for Teaching, Multidimensional Scaling, Psychometrics
Wyse, Adam E.; Reckase, Mark D. – Educational and Psychological Measurement, 2012
This study investigates how different rounding rules and ways of providing Angoff standard-setting judgments affect cut-scores. A simulation design based on data from the National Assessment of Education Progress was used to investigate how rounding judgments to the nearest whole number (e.g., 0, 1, 2, etc.), nearest 0.05, or nearest two decimal…
Descriptors: Standard Setting, Cutting Scores, Statistical Bias, Numbers
He, Wei; Reckase, Mark D. – Educational and Psychological Measurement, 2014
For computerized adaptive tests (CATs) to work well, they must have an item pool with sufficient numbers of good quality items. Many researchers have pointed out that, in developing item pools for CATs, not only is the item pool size important but also the distribution of item parameters and practical considerations such as content distribution…
Descriptors: Item Banks, Test Length, Computer Assisted Testing, Adaptive Testing
Reckase, Mark D.; Xu, Jing-Ru – Educational and Psychological Measurement, 2015
How to compute and report subscores for a test that was originally designed for reporting scores on a unidimensional scale has been a topic of interest in recent years. In the research reported here, we describe an application of multidimensional item response theory to identify a subscore structure in a test designed for reporting results using a…
Descriptors: English, Language Skills, English Language Learners, Scores
Papanastasiou, Elena C.; Reckase, Mark D. – International Journal of Testing, 2007
Because of the increased popularity of computerized adaptive testing (CAT), many admissions tests, as well as certification and licensure examinations, have been transformed from their paper-and-pencil versions to computerized adaptive versions. A major difference between paper-and-pencil tests and CAT from an examinee's point of view is that in…
Descriptors: Simulation, Adaptive Testing, Computer Assisted Testing, Test Items
Reckase, Mark D. – 1988
The requirements for adaptive testing are reviewed, and the question of why implementation has taken so long is examined. The concept of a testing procedure that selects items to match the level of performance of an examinee during the administration of a test had to wait for the technology necessary to apply the idea. Current procedures were…
Descriptors: Adaptive Testing, Computer Assisted Testing, Latent Trait Theory, Test Items
Reckase, Mark D. – Educational Measurement: Issues and Practice, 2006
Schulz (2006) provides a different perspective on standard setting than that provided in Reckase (2006). He also suggests a modification to the bookmark procedure and some alternative models for errors in panelists' judgments than those provided by Reckase. This article provides a response to some of the points made by Schulz and reports some…
Descriptors: Evaluation Methods, Standard Setting, Reader Response, Regression (Statistics)

Reckase, Mark D.; McKinley, Robert L. – 1984
A new indicator of item difficulty, which identifies effectiveness ranges, overcomes the limitations of other item difficulty indexes in describing the difficulty of an item or a test as a whole and in aiding the selection of appropriate ability level items for a test. There are three common uses of the term "item difficulty": (1) the probability…
Descriptors: Difficulty Level, Evaluation Methods, Item Analysis, Latent Trait Theory

Reckase, Mark D.; And Others – Journal of Educational Measurement, 1988
It is demonstrated, theoretically and empirically, that item sets can be selected that meet the unidimensionality assumption of most item response theory models, even though they require more than one ability for a correct response. A method for identifying such item sets for test development purposes is presented. (SLD)
Descriptors: Computer Simulation, Item Analysis, Latent Trait Theory, Mathematical Models

Reckase, Mark D. – Educational Measurement: Issues and Practice, 1989
Requirements for adaptive testing are reviewed, and the reasons implementation has taken so long are explored. The adaptive test is illustrated through the Stanford-Binet Intelligence Scale of L. M. Terman and M. A. Merrill (1960). Current adaptive testing is tied to the development of item response theory. (SLD)
Descriptors: Adaptive Testing, Educational Development, Elementary Secondary Education, Latent Trait Theory

Reckase, Mark D.; McKinley, Robert L. – Applied Psychological Measurement, 1991
The concept of item discrimination is generalized to the case in which more than one ability is required to determine the correct response to an item, using the conceptual framework of item response theory and the definition of multidimensional item difficulty previously developed by M. Reckase (1985). (SLD)
Descriptors: Ability, Definitions, Difficulty Level, Equations (Mathematics)
Reckase, Mark D.; McKinley, Robert L. – 1984
The purpose of this paper is to present a generalization of the concept of item difficulty to test items that measure more than one dimension. Three common definitions of item difficulty were considered: the proportion of correct responses for a group of individuals; the probability of a correct response to an item for a specific person; and the…
Descriptors: Difficulty Level, Item Analysis, Latent Trait Theory, Mathematical Models
Reckase, Mark D. – 1981
The purpose of this paper is to examine the capabilities of various procedures for sorting dichotomously-scored items into unidimensional subjects. The procedures include: factor analysis, nonmetric multidimensional scaling, cluster analysis, and latent trait analysis. Both simulated and real data sets of known structure were used to evaluate the…
Descriptors: Cluster Analysis, Factor Analysis, Guessing (Tests), Latent Trait Theory

Reckase, Mark D. – Psychological Assessment, 1996
Summarizes the current state of the art in test construction and contrasts it with previous conceptual models, some of which are wrong or misleading. New methodologies for item selection and review are presented, with current thinking on the specification of technical characteristics of tests. (Author/SLD)
Descriptors: Mathematical Models, Psychological Testing, Selection, State of the Art Reviews