Publication Date
In 2025 | 1 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 6 |
Since 2016 (last 10 years) | 8 |
Since 2006 (last 20 years) | 13 |
Descriptor
Comparative Analysis | 20 |
Test Format | 20 |
Test Length | 20 |
Test Items | 11 |
Computer Assisted Testing | 9 |
Simulation | 7 |
Item Response Theory | 6 |
Test Reliability | 6 |
Difficulty Level | 5 |
Sample Size | 5 |
Scores | 5 |
More ▼ |
Source
Author
Ann Arthur | 1 |
Benton, Tom | 1 |
Chen Qiu | 1 |
Chi-Yu Huang | 1 |
Christiansen, Neil D. | 1 |
Coats, Pamela K. | 1 |
Dogan, Nuri | 1 |
Dongmei Li | 1 |
Eignor, Daniel R. | 1 |
Eisenstein, Norman | 1 |
Engelhart, Charles I. | 1 |
More ▼ |
Publication Type
Reports - Research | 14 |
Journal Articles | 12 |
Dissertations/Theses -… | 4 |
Reports - Evaluative | 2 |
Speeches/Meeting Papers | 2 |
Education Level
Elementary Education | 1 |
Grade 6 | 1 |
Higher Education | 1 |
Intermediate Grades | 1 |
Middle Schools | 1 |
Postsecondary Education | 1 |
Audience
Location
United Kingdom | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Kaufman Brief Intelligence… | 2 |
Wechsler Adult Intelligence… | 2 |
ACT Assessment | 1 |
ACTFL Oral Proficiency… | 1 |
Advanced Placement… | 1 |
Marlowe Crowne Social… | 1 |
Minnesota Multiphasic… | 1 |
Wechsler Intelligence Scale… | 1 |
What Works Clearinghouse Rating
Shaojie Wang; Won-Chan Lee; Minqiang Zhang; Lixin Yuan – Applied Measurement in Education, 2024
To reduce the impact of parameter estimation errors on IRT linking results, recent work introduced two information-weighted characteristic curve methods for dichotomous items. These two methods showed outstanding performance in both simulation and pseudo-form pseudo-group analysis. The current study expands upon the concept of information…
Descriptors: Item Response Theory, Test Format, Test Length, Error of Measurement
Dongmei Li; Shalini Kapoor; Ann Arthur; Chi-Yu Huang; YoungWoo Cho; Chen Qiu; Hongling Wang – ACT Education Corp., 2025
Starting in April 2025, ACT will introduce enhanced forms of the ACT® test for national online testing, with a full rollout to all paper and online test takers in national, state and district, and international test administrations by Spring 2026. ACT introduced major updates by changing the test lengths and testing times, providing more time per…
Descriptors: College Entrance Examinations, Testing, Change, Scoring
Erdem-Kara, Basak; Dogan, Nuri – International Journal of Assessment Tools in Education, 2022
Recently, adaptive test approaches have become a viable alternative to traditional fixed-item tests. The main advantage of adaptive tests is that they reach desired measurement precision with fewer items. However, fewer items mean that each item has a more significant effect on ability estimation and therefore those tests are open to more…
Descriptors: Item Analysis, Computer Assisted Testing, Test Items, Test Construction
Kárász, Judit T.; Széll, Krisztián; Takács, Szabolcs – Quality Assurance in Education: An International Perspective, 2023
Purpose: Based on the general formula, which depends on the length and difficulty of the test, the number of respondents and the number of ability levels, this study aims to provide a closed formula for the adaptive tests with medium difficulty (probability of solution is p = 1/2) to determine the accuracy of the parameters for each item and in…
Descriptors: Test Length, Probability, Comparative Analysis, Difficulty Level
Benton, Tom – Research Matters, 2021
Computer adaptive testing is intended to make assessment more reliable by tailoring the difficulty of the questions a student has to answer to their level of ability. Most commonly, this benefit is used to justify the length of tests being shortened whilst retaining the reliability of a longer, non-adaptive test. Improvements due to adaptive…
Descriptors: Risk, Item Response Theory, Computer Assisted Testing, Difficulty Level
Ozdemir, Burhanettin; Gelbal, Selahattin – Education and Information Technologies, 2022
The computerized adaptive tests (CAT) apply an adaptive process in which the items are tailored to individuals' ability scores. The multidimensional CAT (MCAT) designs differ in terms of different item selection, ability estimation, and termination methods being used. This study aims at investigating the performance of the MCAT designs used to…
Descriptors: Scores, Computer Assisted Testing, Test Items, Language Proficiency
Lance M. Kruse – ProQuest LLC, 2019
This study explores six item-reduction methodologies used to shorten an existing complex problem-solving non-objective test by evaluating how each shortened form performs across three sources of validity evidence (i.e., test content, internal structure, and relationships with other variables). Two concerns prompted the development of the present…
Descriptors: Educational Assessment, Comparative Analysis, Test Format, Test Length
Isbell, Dan; Winke, Paula – Language Testing, 2019
The American Council on the Teaching of Foreign Languages (ACTFL) oral proficiency interview -- computer (OPIc) testing system represents an ambitious effort in language assessment: Assessing oral proficiency in over a dozen languages, on the same scale, from virtually anywhere at any time. Especially for users in contexts where multiple foreign…
Descriptors: Oral Language, Language Tests, Language Proficiency, Second Language Learning
Kabasakal, Kübra Atalay; Kelecioglu, Hülya – Educational Sciences: Theory and Practice, 2015
This study examines the effect of differential item functioning (DIF) items on test equating through multilevel item response models (MIRMs) and traditional IRMs. The performances of three different equating models were investigated under 24 different simulation conditions, and the variables whose effects were examined included sample size, test…
Descriptors: Test Bias, Equated Scores, Item Response Theory, Simulation
Wu, Yi-Fang – ProQuest LLC, 2015
Item response theory (IRT) uses a family of statistical models for estimating stable characteristics of items and examinees and defining how these characteristics interact in describing item and test performance. With a focus on the three-parameter logistic IRT (Birnbaum, 1968; Lord, 1980) model, the current study examines the accuracy and…
Descriptors: Item Response Theory, Test Items, Accuracy, Computation
Wang, Wei – ProQuest LLC, 2013
Mixed-format tests containing both multiple-choice (MC) items and constructed-response (CR) items are now widely used in many testing programs. Mixed-format tests often are considered to be superior to tests containing only MC items although the use of multiple item formats leads to measurement challenges in the context of equating conducted under…
Descriptors: Equated Scores, Test Format, Test Items, Test Length
Sunnassee, Devdass – ProQuest LLC, 2011
Small sample equating remains a largely unexplored area of research. This study attempts to fill in some of the research gaps via a large-scale, IRT-based simulation study that evaluates the performance of seven small-sample equating methods under various test characteristic and sampling conditions. The equating methods considered are typically…
Descriptors: Test Length, Test Format, Sample Size, Simulation
Rotou, Ourania; Patsula, Liane; Steffen, Manfred; Rizavi, Saba – ETS Research Report Series, 2007
Traditionally, the fixed-length linear paper-and-pencil (P&P) mode of administration has been the standard method of test delivery. With the advancement of technology, however, the popularity of administering tests using adaptive methods like computerized adaptive testing (CAT) and multistage testing (MST) has grown in the field of measurement…
Descriptors: Comparative Analysis, Test Format, Computer Assisted Testing, Models

Silverstein, A. B. – Perceptual and Motor Skills, 1983
Formulas for estimating the validity of random short forms were applied to the standardization data for the Wechsler Adult Intelligence Scale-Revised, the Minnesota Multiphasic Personality Inventory, and the Marlowe-Crowne Social Desirability Scale. These formulas demonstrated how much "better than random" the best short forms of these…
Descriptors: Comparative Analysis, Intelligence Tests, Measures (Individuals), Test Format

Eisenstein, Norman; Engelhart, Charles I. – Psychological Assessment, 1997
The Kaufman Brief Intelligence Test (K-BIT) (A. S. Kaufman and N. L. Kaufman, 1990) was compared with short forms of the Wechsler Adult Intelligence Scale--Revised (WAIS-R) using results from 64 referrals to a neuropsychology service. Advantages of each test are noted and their use discussed. (SLD)
Descriptors: Adults, Comparative Analysis, Intelligence Tests, Neuropsychology
Previous Page | Next Page »
Pages: 1 | 2