Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 9 |
Descriptor
Test Length | 54 |
Testing Problems | 54 |
Test Construction | 20 |
Test Items | 18 |
Test Validity | 17 |
Test Format | 14 |
Test Reliability | 13 |
Item Analysis | 10 |
Multiple Choice Tests | 10 |
Computer Assisted Testing | 9 |
Mathematical Models | 9 |
More ▼ |
Source
Author
Wainer, Howard | 2 |
Wilcox, Rand R. | 2 |
van der Linden, Wim J. | 2 |
Alonzo, Julie | 1 |
Axelrod, Bradley N. | 1 |
Bay, Luz | 1 |
Bergstrom, Betty | 1 |
Boyd, Thomas A. | 1 |
Brennan, Robert L. | 1 |
Budescu, David | 1 |
Camilli, Gregory | 1 |
More ▼ |
Publication Type
Education Level
Elementary Education | 2 |
Early Childhood Education | 1 |
Elementary Secondary Education | 1 |
Grade 2 | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
Higher Education | 1 |
Intermediate Grades | 1 |
Postsecondary Education | 1 |
Primary Education | 1 |
Audience
Researchers | 6 |
Practitioners | 1 |
Location
Japan | 1 |
New Jersey | 1 |
Rhode Island | 1 |
United Kingdom | 1 |
Vermont | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Uysal, Ibrahim; Sahin-Kürsad, Merve; Kiliç, Abdullah Faruk – Participatory Educational Research, 2022
The aim of the study was to examine the common items in the mixed format (e.g., multiple-choices and essay items) contain parameter drifts in the test equating processes performed with the common item nonequivalent groups design. In this study, which was carried out using Monte Carlo simulation with a fully crossed design, the factors of test…
Descriptors: Test Items, Test Format, Item Response Theory, Equated Scores
Sinharay, Sandip – Applied Measurement in Education, 2017
Karabatsos compared the power of 36 person-fit statistics using receiver operating characteristics curves and found the "H[superscript T]" statistic to be the most powerful in identifying aberrant examinees. He found three statistics, "C", "MCI", and "U3", to be the next most powerful. These four statistics,…
Descriptors: Nonparametric Statistics, Goodness of Fit, Simulation, Comparative Analysis
Isbell, Dan; Winke, Paula – Language Testing, 2019
The American Council on the Teaching of Foreign Languages (ACTFL) oral proficiency interview -- computer (OPIc) testing system represents an ambitious effort in language assessment: Assessing oral proficiency in over a dozen languages, on the same scale, from virtually anywhere at any time. Especially for users in contexts where multiple foreign…
Descriptors: Oral Language, Language Tests, Language Proficiency, Second Language Learning
Kahn, Josh; Nese, Joseph T.; Alonzo, Julie – Behavioral Research and Teaching, 2016
There is strong theoretical support for oral reading fluency (ORF) as an essential building block of reading proficiency. The current and standard ORF assessment procedure requires that students read aloud a grade-level passage (˜ 250 words) in a one-to-one administration, with the number of words read correctly in 60 seconds constituting their…
Descriptors: Teacher Surveys, Oral Reading, Reading Tests, Computer Assisted Testing
Watanabe, Yoshinori – Language Testing, 2013
This article describes the National Center Test for University Admissions, a unified national test in Japan, which is taken by 500,000 students every year. It states that implementation of the Center Test began in 1990, with the English component consisting only of the written section until 2005, when the listening section was first implemented…
Descriptors: College Admission, Foreign Countries, College Entrance Examinations, English (Second Language)
Camilli, Gregory – Educational Research and Evaluation, 2013
In the attempt to identify or prevent unfair tests, both quantitative analyses and logical evaluation are often used. For the most part, fairness evaluation is a pragmatic attempt at determining whether procedural or substantive due process has been accorded to either a group of test takers or an individual. In both the individual and comparative…
Descriptors: Alternative Assessment, Test Bias, Test Content, Test Format
Cui, Ying; Leighton, Jacqueline P. – Journal of Educational Measurement, 2009
In this article, we introduce a person-fit statistic called the hierarchy consistency index (HCI) to help detect misfitting item response vectors for tests developed and analyzed based on a cognitive model. The HCI ranges from -1.0 to 1.0, with values close to -1.0 indicating that students respond unexpectedly or differently from the responses…
Descriptors: Test Length, Simulation, Correlation, Research Methodology
Wollack, James A. – Applied Measurement in Education, 2006
Many of the currently available statistical indexes to detect answer copying lack sufficient power at small [alpha] levels or when the amount of copying is relatively small. Furthermore, there is no one index that is uniformly best. Depending on the type or amount of copying, certain indexes are better than others. The purpose of this article was…
Descriptors: Statistical Analysis, Item Analysis, Test Length, Sample Size
Rhode Island Department of Elementary and Secondary Education, 2007
This handbook will assist principals and school testing coordinators in implementing the spring 2007 administration of the Developmental Reading Assessment (DRA). Information regarding administration timeline, reporting, process, online tools and contact personnel is discussed. Contents include: (1) Scheduling; (2) Identify Primary Test…
Descriptors: Testing Accommodations, Alternative Assessment, Educational Testing, Guidance Programs
Bay, Luz – 1995
An index is proposed to detect cheating on multiple-choice examinations, and its use is evaluated through simulations. The proposed index is based on the compound binomial distribution. In total, 360 simulated data sets reflecting 12 different cheating (copying) situations were obtained and used for the study of the sensitivity of the index in…
Descriptors: Cheating, Class Size, Identification, Multiple Choice Tests

Conger, Anthony J. – Educational and Psychological Measurement, 1983
A paradoxical phenomenon of decreases in reliability as the number of elements averaged over increases is shown to be possible in multifacet reliability procedures (intraclass correlations or generalizability coefficients). Conditions governing this phenomenon are presented along with implications and cautions. (Author)
Descriptors: Generalizability Theory, Test Construction, Test Items, Test Length

Modjeski, Richard B.; Michael, William B. – Educational and Psychological Measurement, 1978
The General Education Performance Index (GEPI) is a comparatively short test covering the same content as the General Educational Development Test (GED), which takes ten hours to administer. Correlations of the subtests of the GEPI with the GED ranged from .28 to .57. (JKS)
Descriptors: Correlation, Equivalency Tests, Military Personnel, Statistical Data
Brennan, Robert L. – 1990
In 1955, R. Levine introduced two linear equating procedures for the common-item non-equivalent populations design. His procedures make the same assumptions about true scores; they differ in terms of the nature of the equating function used. In this paper, two parameterizations of a classical congeneric model are introduced to model the variables…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Research Design

Gallucci, Nicholas T. – Educational and Psychological Measurement, 1986
This study evaluated the degree to which 102 undergraduate participants objected to questions on the Minnesota Multiphasic Personality Inventory (MMPI) which referred to sex, religion, bladder and bowel functions, family relationships, and unusual thinking in comparision to degree of objection to length of the MMPI and repetition of questions.…
Descriptors: College Students, Higher Education, Personality Measures, Psychological Evaluation

Sindhu, R. S.; Sharma, Reeta – Science Education International, 1999
Finds that the time required to attempt all the test items of each question paper in a four-paper sample was inversely proportional to the percentage of students who attempted all the test items of that paper. Extrapolates results to give guidelines for determining the feasibility of newly-developed exam papers. (WRM)
Descriptors: Science Tests, Secondary Education, Test Construction, Test Length