Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 5 |
Descriptor
Source
Author
Reckase, Mark D. | 33 |
McKinley, Robert L. | 5 |
He, Wei | 2 |
Patience, Wayne M. | 2 |
Ackerman, Terry A. | 1 |
Ferrini-Mundy, Joan | 1 |
Floden, Robert E. | 1 |
Koch, William R. | 1 |
Luo, Xin | 1 |
McCrory, Raven | 1 |
Senk, Sharon L. | 1 |
More ▼ |
Publication Type
Reports - Research | 19 |
Speeches/Meeting Papers | 16 |
Reports - Evaluative | 13 |
Journal Articles | 10 |
Information Analyses | 3 |
Opinion Papers | 2 |
Numerical/Quantitative Data | 1 |
Reports - General | 1 |
Tests/Questionnaires | 1 |
Education Level
Audience
Researchers | 3 |
Location
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 4 |
COMPASS (Computer Assisted… | 1 |
National Assessment of… | 1 |
School and College Ability… | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating
Yang, Lihong; Reckase, Mark D. – Educational and Psychological Measurement, 2020
The present study extended the "p"-optimality method to the multistage computerized adaptive test (MST) context in developing optimal item pools to support different MST panel designs under different test configurations. Using the Rasch model, simulated optimal item pools were generated with and without practical constraints of exposure…
Descriptors: Item Banks, Adaptive Testing, Computer Assisted Testing, Item Response Theory
Reckase, Mark D. – ETS Research Report Series, 2017
A common interpretation of achievement test results is that they provide measures of achievement that are much like other measures we commonly use for height, weight, or the cost of goods. In a limited sense, such interpretations are correct, but some nuances of these interpretations have important implications for the use of achievement test…
Descriptors: Models, Achievement Tests, Test Results, Test Construction
Luo, Xin; Reckase, Mark D.; He, Wei – AERA Online Paper Repository, 2016
While dichotomous item dominates the application of computerized adaptive testing (CAT), polytomous item and set-based item hold promises for being incorporated in CAT. However, how to assemble a CAT containing mixed item formats is challenging. This study investigated: (1) how the mixed CAT works compared with the dichotomous-item-based CAT; (2)…
Descriptors: Test Items, Test Format, Computer Assisted Testing, Adaptive Testing
Reckase, Mark D.; McCrory, Raven; Floden, Robert E.; Ferrini-Mundy, Joan; Senk, Sharon L. – Educational Assessment, 2015
Numerous researchers have suggested that there are multiple mathematical knowledge and skill areas needed by teachers in order for them to be effective teachers of mathematics: knowledge of the mathematics that are the goals of instruction, advanced mathematics beyond the instructional material, and mathematical knowledge that is specific to what…
Descriptors: Algebra, Knowledge Base for Teaching, Multidimensional Scaling, Psychometrics
He, Wei; Reckase, Mark D. – Educational and Psychological Measurement, 2014
For computerized adaptive tests (CATs) to work well, they must have an item pool with sufficient numbers of good quality items. Many researchers have pointed out that, in developing item pools for CATs, not only is the item pool size important but also the distribution of item parameters and practical considerations such as content distribution…
Descriptors: Item Banks, Test Length, Computer Assisted Testing, Adaptive Testing

Reckase, Mark D. – Educational Measurement: Issues and Practice, 1998
Considers what a responsible test developer would do to gain information to support the consequential basis of validity for a test early in the development. How the consequential basis of validity of the program would be monitored and reported during the life of the program is examined. The validity of the ACT Assessment is considered as if it…
Descriptors: Evaluation Methods, Program Evaluation, Test Construction, Validity
Unidimensional Data from Multidimensional Tests and Multidimensional Data from Unidimensional Tests.
Reckase, Mark D. – 1990
Although the issue of dimensionality of the data obtained from educational and psychological tests has received considerable attention, the terms "unidimensional" and "multidimensional" have not been used very precisely. One use of the term dimensionality is to refer to the number of hypothesized psychological constructs…
Descriptors: Item Response Theory, Matrices, Statistical Analysis, Test Construction
Reckase, Mark D. – 1981
Definition of the issues to the use of latent trait models, specifically one- and three-parameter logistic models, in conjunction with multi-level achievement batteries, forms the basis of this paper. Research results related to these issues are also documented in an attempt to provide a rational basis for model selection. The application of the…
Descriptors: Achievement Tests, Comparative Analysis, Latent Trait Theory, Scores

Reckase, Mark D.; McKinley, Robert L. – 1984
A new indicator of item difficulty, which identifies effectiveness ranges, overcomes the limitations of other item difficulty indexes in describing the difficulty of an item or a test as a whole and in aiding the selection of appropriate ability level items for a test. There are three common uses of the term "item difficulty": (1) the probability…
Descriptors: Difficulty Level, Evaluation Methods, Item Analysis, Latent Trait Theory

McKinley, Robert L.; Reckase, Mark D. – AEDS Journal, 1980
Describes tailored testing (in which a computer selects appropriate items from an item bank while an examinee is taking a test) and shows it to be superior to paper-and-pencil tests in such areas as reliability, security, and appropriateness of items. (IRT)
Descriptors: Adaptive Testing, Computer Assisted Testing, Higher Education, Program Evaluation

Reckase, Mark D.; And Others – Journal of Educational Measurement, 1988
It is demonstrated, theoretically and empirically, that item sets can be selected that meet the unidimensionality assumption of most item response theory models, even though they require more than one ability for a correct response. A method for identifying such item sets for test development purposes is presented. (SLD)
Descriptors: Computer Simulation, Item Analysis, Latent Trait Theory, Mathematical Models

Reckase, Mark D. – Educational Measurement: Issues and Practice, 1989
Requirements for adaptive testing are reviewed, and the reasons implementation has taken so long are explored. The adaptive test is illustrated through the Stanford-Binet Intelligence Scale of L. M. Terman and M. A. Merrill (1960). Current adaptive testing is tied to the development of item response theory. (SLD)
Descriptors: Adaptive Testing, Educational Development, Elementary Secondary Education, Latent Trait Theory

Reckase, Mark D. – Psychological Assessment, 1996
Summarizes the current state of the art in test construction and contrasts it with previous conceptual models, some of which are wrong or misleading. New methodologies for item selection and review are presented, with current thinking on the specification of technical characteristics of tests. (Author/SLD)
Descriptors: Mathematical Models, Psychological Testing, Selection, State of the Art Reviews

Reckase, Mark D. – 1979
Because latent trait models require that large numbers of items be calibrated or that testing of the same large group be repeated, item parameter estimates are often obtained by administering separate tests to different groups and "linking" the results to construct an adequate item pool. Four issues were studied, based upon the analysis…
Descriptors: Achievement Tests, High Schools, Item Banks, Mathematical Models
Reckase, Mark D. – 1981
This report summarizes the research findings of a four year contract investigating the applicability of item response theory and tailored testing to criterion-referenced measurement. Six major areas were studied on the project: (1) techniques for forming unidimensional item sets; (2) techniques for calibrating items; (3) item parameter linking…
Descriptors: Achievement Tests, Adaptive Testing, Criterion Referenced Tests, Decision Making