Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 2 |
Descriptor
Source
Applied Psychological… | 22 |
Author
Baker, A. Harvey | 1 |
Barnes, Janet L. | 1 |
Brennan, Robert L. | 1 |
Cliff, Norman | 1 |
Cudeck, Robert | 1 |
Divgi, D. R. | 1 |
Downey, Ronald G. | 1 |
Dunlap, William P. | 1 |
Eiting, Mindert H. | 1 |
Funke, Joachim | 1 |
Greiff, Samuel | 1 |
More ▼ |
Publication Type
Journal Articles | 15 |
Reports - Research | 6 |
Reports - Evaluative | 5 |
Collected Works - Serials | 2 |
Reports - Descriptive | 1 |
Reports - General | 1 |
Tests/Questionnaires | 1 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
West Germany | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Armed Forces Qualification… | 1 |
Armed Services Vocational… | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating
Yao, Lihua – Applied Psychological Measurement, 2013
Through simulated data, five multidimensional computerized adaptive testing (MCAT) selection procedures with varying test lengths are examined and compared using different stopping rules. Fixed item exposure rates are used for all the items, and the Priority Index (PI) method is used for the content constraints. Two stopping rules, standard error…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Selection
Greiff, Samuel; Wustenberg, Sascha; Funke, Joachim – Applied Psychological Measurement, 2012
This article addresses two unsolved measurement issues in dynamic problem solving (DPS) research: (a) unsystematic construction of DPS tests making a comparison of results obtained in different studies difficult and (b) use of time-intensive single tasks leading to severe reliability problems. To solve these issues, the MicroDYN approach is…
Descriptors: Problem Solving, Tests, Measurement, Structural Equation Models

Divgi, D. R. – Applied Psychological Measurement, 1989
Two methods for estimating the reliability of a computerized adaptive test (CAT) without using item response theory are presented. The data consist of CAT and paper-and-pencil scores from identical or equivalent samples, and scores for all examinees on one or more covariates, using the Armed Services Vocational Aptitude Battery. (TJH)
Descriptors: Adaptive Testing, Computer Assisted Testing, Estimation (Mathematics), Predictive Validity

Cudeck, Robert; And Others – Applied Psychological Measurement, 1980
Tailored testing by Cliff's method of implied orders was simulated through the use of responses gathered during conventional administration of the Stanford-Binet Intelligence Scale. Tailoring eliminated approximately half the responses with only modest decreases in score reliability. (Author/BW)
Descriptors: Adaptive Testing, Computer Assisted Testing, Elementary Secondary Education, Intelligence Tests

Luecht, Richard M. – Applied Psychological Measurement, 1996
The example of a medical licensure test is used to demonstrate situations in which complex, integrated content must be balanced at the total test level for validity reasons, but items assigned to reportable subscore categories may be used under a multidimensional item response theory adaptive paradigm to improve subscore reliability. (SLD)
Descriptors: Adaptive Testing, Certification, Computer Assisted Testing, Licensing Examinations (Professions)

And Others; Mann, Irene T. – Applied Psychological Measurement, 1979
Several methodological problems (particularly the assumed bipolarity of scales, instructions regarding use of the midpoint, and concept-scale interaction) which may contribute to a lack of precision in the semantic differential technique were investigated. Results generally supported the use of the semantic differential. (Author/JKS)
Descriptors: Analysis of Variance, Computer Assisted Testing, Higher Education, Rating Scales

Brennan, Robert L.; Lockwood, Robert E. – Applied Psychological Measurement, 1980
Generalizability theory is used to characterize and quantify expected variance in cutting scores and to compare the Nedelsky and Angoff procedures for establishing a cutting score. Results suggest that the restricted nature of the Nedelsky (inferred) probability scale may limit its applicability in certain contexts. (Author/BW)
Descriptors: Cutting Scores, Generalization, Statistical Analysis, Test Reliability

Waters, Brian K. – Applied Psychological Measurement, 1977
The validity and utility of the stratified adaptive computerized testing model (stradaptive) developed by Weiss are empirically investigated. The model presents a tailored testing strategy based upon Binet IQ measurement theory and Lord's modern test theory. (Author/RC)
Descriptors: Ability, Adaptive Testing, Computer Oriented Programs, Item Banks

Mishara, Brian L.; Baker, A. Harvey – Applied Psychological Measurement, 1978
The validity of the Kinesthetic Aftereffect (KAE) as a measure of personality has been criticized because of KAE's poor test-retest reliability. However, systematic bias effects render KA E retest sessions invalid and make test-retest reliability an inappropriate measure of KAE's true reliability. (Author/CTM)
Descriptors: Kinesthetic Perception, Perception Tests, Personality Measures, Tactual Perception

Weiss, David J., Ed. – Applied Psychological Measurement, 1987
Issues concerning equating test scores are discussed in an introduction, four papers, and two commentaries. Equating methods research, sampling errors, linear equating, population differences, sources of equating errors, and a circular equating paradigm are considered. (SLD)
Descriptors: Equated Scores, Latent Trait Theory, Maximum Likelihood Statistics, Statistical Analysis

Samejima, Fumiko – Applied Psychological Measurement, 1977
Several important implications in latent trait theory, with implications for individualized or tailored testing, are pointed out. A way of using the information function in tailored testing in connection with the standard error estimation of the ability level using maximum likelihood estimation is suggested. (Author/JKS)
Descriptors: Adaptive Testing, Career Development, Error of Measurement, Item Analysis

Cliff, Norman; And Others – Applied Psychological Measurement, 1979
Monte Carlo research with TAILOR, a program using implied orders as a basis for tailored testing, is reported. TAILOR typically required about half the available items to estimate, for each simulated examinee, the responses on the remainder. (Author/CTM)
Descriptors: Adaptive Testing, Computer Programs, Item Sampling, Nonparametric Statistics

Poizner, Sharon B.; And Others – Applied Psychological Measurement, 1978
Binary, probability, and ordinal scoring procedures for multiple-choice items were examined. In two situations, it was found that both the probability and ordinal scoring systems were more reliable than the binary scoring method. (Author/CTM)
Descriptors: Confidence Testing, Guessing (Tests), Higher Education, Multiple Choice Tests

Lunneborg, Clifford E. – Applied Psychological Measurement, 1977
Three studies are described in which choice reaction time (RT) was related to such psychometric ability measures as verbal comprehension, numerical reasoning, hidden figures, and progressive matrices tests. Fairly consistent negative correlations were found between these tests and choice RT when high school samples were used. (Author/CTM)
Descriptors: Cognitive Ability, Cognitive Processes, High Schools, Higher Education

Hambleton, Ronald K., Ed. – Applied Psychological Measurement, 1980
This special issue covers recent technical developments in the field of criterion-referenced testing. An introduction, six papers, and two commentaries dealing with test development, test score uses, and evaluation of scores review relevant literature, offer new models and/or results, and suggest directions for additional research. (SLD)
Descriptors: Criterion Referenced Tests, Mastery Tests, Measurement Techniques, Standard Setting (Scoring)
Previous Page | Next Page ยป
Pages: 1 | 2