Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 3 |
| Since 2007 (last 20 years) | 6 |
Descriptor
| Data Collection | 14 |
| Test Items | 14 |
| Scoring | 9 |
| Scaling | 5 |
| Elementary Secondary Education | 4 |
| Test Construction | 4 |
| Computer Assisted Testing | 3 |
| National Surveys | 3 |
| Research Design | 3 |
| Scores | 3 |
| Academic Achievement | 2 |
| More ▼ | |
Source
| Journal of Educational… | 2 |
| Applied Psychological… | 1 |
| Educational Evaluation and… | 1 |
| Educational Testing Service | 1 |
| European Journal of… | 1 |
| International Association for… | 1 |
| Research & Practice in… | 1 |
Author
Publication Type
Education Level
| Elementary Secondary Education | 1 |
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 4 |
| Comprehensive Tests of Basic… | 1 |
| National Teacher Examinations | 1 |
| Trends in International… | 1 |
What Works Clearinghouse Rating
Baldwin, Peter; Clauser, Brian E. – Journal of Educational Measurement, 2022
While score comparability across test forms typically relies on common (or randomly equivalent) examinees or items, innovations in item formats, test delivery, and efforts to extend the range of score interpretation may require a special data collection before examinees or items can be used in this way--or may be incompatible with common examinee…
Descriptors: Scoring, Testing, Test Items, Test Format
Setiawan, Risky – European Journal of Educational Research, 2019
The purposes of this research are: 1) to compare two equalizing tests conducted with Hebara and Stocking Lord method; 2) to describe the characteristics of each equalizing test method using windows' IRTEQ program. This research employs a participatory approach as the data are collected through questionnaires based on the National Examination…
Descriptors: Equated Scores, Evaluation Methods, Evaluation Criteria, Test Items
Mullis, Ina V. S., Ed.; Martin, Michael O., Ed.; von Davier, Matthias, Ed. – International Association for the Evaluation of Educational Achievement, 2021
TIMSS (Trends in International Mathematics and Science Study) is a long-standing international assessment of mathematics and science at the fourth and eighth grades that has been collecting trend data every four years since 1995. About 70 countries use TIMSS trend data for monitoring the effectiveness of their education systems in a global…
Descriptors: Achievement Tests, International Assessment, Science Achievement, Mathematics Achievement
Zhu, Mengxiao; Shu, Zhan; von Davier, Alina A. – Journal of Educational Measurement, 2016
New technology enables interactive and adaptive scenario-based tasks (SBTs) to be adopted in educational measurement. At the same time, it is a challenging problem to build appropriate psychometric models to analyze data collected from these tasks, due to the complexity of the data. This study focuses on process data collected from SBTs. We…
Descriptors: Measurement, Data Collection, National Competency Tests, Scoring Rubrics
Livingston, Samuel A. – Educational Testing Service, 2014
This booklet grew out of a half-day class on equating that author Samuel Livingston teaches for new statistical staff at Educational Testing Service (ETS). The class is a nonmathematical introduction to the topic, emphasizing conceptual understanding and practical applications. The class consists of illustrated lectures, interspersed with…
Descriptors: Equated Scores, Scoring, Self Evaluation (Individuals), Scores
Fulcher, Keston H.; Orem, Chris D. – Research & Practice in Assessment, 2010
Higher education experts tout learning outcomes assessment as a vehicle for program improvement. To this end the authors share a rubric designed explicitly to evaluate the quality of assessment and how it leads to program improvement. The rubric contains six general assessment areas, which are further broken down into 14 elements. Embedded within…
Descriptors: Higher Education, Scoring Rubrics, Educational Quality, Program Improvement
Luecht, Richard M. – 2001
The Microsoft Certification Program (MCP) includes many new computer-based item types, based on complex cases involving the Windows 2000 (registered) operating system. This Innovative Item Technology (IIT) has presented challenges beyond traditional psychometric considerations such as capturing and storing the relevant response data from…
Descriptors: Certification, Coding, Computer Assisted Testing, Data Collection
Anderson, Richard Ivan – 1980
Features of a probabilistic testing system that has been implemented on the "cerl" PLATO computer system are described. The key feature of the system is the manner in which an examinee responds to each test item; the examinee distributes probabilities among the alternatives of each item by positioning a small square on or within an…
Descriptors: Computer Assisted Testing, Data Collection, Feedback, Probability
Peer reviewedCliff, Norman; And Others – Applied Psychological Measurement, 1988
A method for ordering persons and items when all responses are ordinal was developed and applied to several sets of questionnaire data (from undergraduates) and one set of archeological data. The method provides a possible nonparametric treatment of data usually treated by more traditional psychometric methods. (SLD)
Descriptors: Archaeology, Data Analysis, Data Collection, Higher Education
Yen, Wendy M. – 1982
The three-parameter logistic model discussed was used by CTB/McGraw-Hill in the development of the Comprehensive Tests of Basic Skills, Form U (CTBS/U) and the Test of Cognitive Skills (TCS), published in the fall of 1981. The development, standardization, and scoring of the tests are described, particularly as these procedures were influenced by…
Descriptors: Achievement Tests, Bayesian Statistics, Cognitive Processes, Data Collection
Peer reviewedKoretz, Daniel – Educational Evaluation and Policy Analysis, 1995
Studies of the mathematics assessments of the National Assessment of Educational Progress (NAEP) are summarized. One study found that omit rates for NAEP test items were higher for African Americans and Hispanics than for whites. The other found that descriptions and examples for the 1992 mathematics achievement levels were misleading. (SLD)
Descriptors: Black Students, Data Collection, Elementary Secondary Education, Hispanic Americans
Beaton, Albert E.; And Others – 1988
This report supplies details of the design and data analysis of the 1986 National Assessment of Educational Progress (NAEP) to allow the reader to judge the utility of the design, data quality, reasonableness of assumptions, appropriateness of data analyses, and generalizability of inferences made from the data. After an introduction by A. E.…
Descriptors: Data Collection, Data Processing, Databases, Field Tests
Bowman, Harry L.; And Others – 1989
A study conducted for the Tennessee State Department of Education by Memphis State University addressed two objectives with respect to the Professional School Service Personnel endorsements for initial licensure for non-teaching professional personnel: to determine the validity of the National Teacher Examination (NTE) Core Battery tests for…
Descriptors: Communication Skills, Data Collection, Education Work Relationship, Elementary Secondary Education
Johnson, Eugene G.; And Others – 1994
The 1992 National Assessment of Educational Progress (NAEP) monitored the performance of students in American schools in reading, mathematics, science, and writing. The sample consisted of more than 145,000 public and private school students in grades 4, 8, 11, and 12. This technical report provides details of instrument development, sample…
Descriptors: Academic Achievement, Data Analysis, Data Collection, Educational Assessment

Direct link
