Publication Date
In 2025 | 0 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 5 |
Since 2016 (last 10 years) | 18 |
Since 2006 (last 20 years) | 47 |
Descriptor
Test Theory | 48 |
Foreign Countries | 24 |
Item Response Theory | 24 |
Test Items | 16 |
Psychometrics | 15 |
Grade 8 | 14 |
Test Construction | 10 |
Difficulty Level | 9 |
Factor Analysis | 9 |
Grade 7 | 9 |
Statistical Analysis | 9 |
More ▼ |
Source
Author
Publication Type
Education Level
Secondary Education | 48 |
High Schools | 18 |
Junior High Schools | 18 |
Middle Schools | 18 |
Elementary Education | 15 |
Grade 8 | 14 |
Grade 7 | 9 |
Grade 6 | 7 |
Elementary Secondary Education | 6 |
Grade 4 | 6 |
Intermediate Grades | 6 |
More ▼ |
Audience
Location
Indonesia | 2 |
Nigeria | 2 |
Tennessee | 2 |
Texas | 2 |
Turkey | 2 |
United Kingdom (England) | 2 |
United States | 2 |
Australia | 1 |
Belgium | 1 |
China | 1 |
Colorado | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Kartianom Kartianom; Heri Retnawati; Kana Hidayati – Journal of Pedagogical Research, 2024
Conducting a fair test is important for educational research. Unfair assessments can lead to gender disparities in academic achievement, ultimately resulting in disparities in opportunities, wages, and career choice. Differential Item Function [DIF] analysis is presented to provide evidence of whether the test is truly fair, where it does not harm…
Descriptors: Foreign Countries, Test Bias, Item Response Theory, Test Theory
Diana Muela-Bermejo; Irene Mendoza-Cercadillo; Lucía Hernández-Heras – Journal of Adolescent & Adult Literacy, 2024
This study involves translating, cross-culturally adapting, and validating the "Literary Response Questionnaire" (LRQ) for 413 Spanish adolescents. It explores the evolution of literary education in Spain and its alignment with the Reading Responses paradigm. The LRQ, adapted across various locations, is validated in Spanish through…
Descriptors: Reader Response, Adolescents, Questionnaires, Translation
Huebner, Alan; Skar, Gustaf B. – Practical Assessment, Research & Evaluation, 2021
Writing assessments often consist of students responding to multiple prompts, which are judged by more than one rater. To establish the reliability of these assessments, there exist different methods to disentangle variation due to prompts and raters, including classical test theory, Many Facet Rasch Measurement (MFRM), and Generalizability Theory…
Descriptors: Error of Measurement, Test Theory, Generalizability Theory, Item Response Theory
Parker, Mark A. J.; Hedgeland, Holly; Jordan, Sally E.; Braithwaite, Nicholas St. J. – European Journal of Science and Mathematics Education, 2023
The study covers the development and testing of the alternative mechanics survey (AMS), a modified force concept inventory (FCI), which used automatically marked free-response questions. Data were collected over a period of three academic years from 611 participants who were taking physics classes at high school and university level. A total of…
Descriptors: Test Construction, Scientific Concepts, Physics, Test Reliability
LeBeau, Brandon; Assouline, Susan G.; Mahatmya, Duhita; Lupkowski-Shoplik, Ann – Gifted Child Quarterly, 2020
This study investigated the application of item response theory (IRT) to expand the range of ability estimates for gifted (hereinafter referred to as high-achieving) students' performance on an above-level test. Using a sample of fourth- to sixth-grade high-achieving students (N = 1,893), we conducted a study to compare estimates from two…
Descriptors: Item Response Theory, Test Theory, Academically Gifted, High Achievement
Ayanwale, Musa Adekunle; Adeleke, Joshua Oluwatoyin; Mamadelo, Titilayo Iyabode – Journal of the International Society for Teacher Education, 2019
A scoring framework that does not reflect true performance of an examinee would ultimately result in an abnormal score. This study assessed invariance person estimates of 2017 Nigerian National Examinations Council Basic Education Certificate Examination Mathematics Multiple Choice using classical test theory (CTT) and item response theory (IRT)…
Descriptors: Test Theory, Item Response Theory, Scoring, National Competency Tests
Chirkina, T. A.; Khavenson, T. E. – Russian Education & Society, 2018
School climate is one of the significant factors determining educational achievement. However, the lack of instruments to measure it has complicated the study of this concept in Russia. We review the history of the study of the concept of "school climate," and we discuss approaches to how it can be defined. We describe the most widely…
Descriptors: Educational Environment, Definitions, Measurement, Questionnaires
Chin, Huan; Chew, Cheng Meng; Lim, Hooi Lian; Thien, Lei Mee – International Journal of Science and Mathematics Education, 2022
Cognitive Diagnostic Assessment (CDA) is an alternative assessment which can give a clear picture of pupils' learning process and cognitive structures to education stakeholders so that appropriate instructional strategies can be designed to tailored pupils' needs. Coincide with this function, the Ordered Multiple-Choice (OMC) items were…
Descriptors: Mathematics Instruction, Mathematics Tests, Multiple Choice Tests, Diagnostic Tests
Goldstein, Harvey – Assessment in Education: Principles, Policy & Practice, 2017
The author's commentary focuses more on the quantitative discussion about educational assessment of the original article than on the idea of the assessment for learning, which did not raise any substantial issues. He starts by offering some general comments on the paper. He feels the authors made a number of assumptions about quantitative…
Descriptors: Educational Assessment, Statistical Analysis, International Assessment, Learning Theories
Ayva Yörü, Fatma Gökçen; Atar, Hakan Yavuz – Journal of Pedagogical Research, 2019
The aim of this study is to examine whether the items in the mathematics subtest of the Centralized High School Entrance Placement Test [HSEPT] administered in 2012 by the Ministry of National Education in Turkey show DIF according to gender and type of school. For this purpose, SIBTEST, Breslow-Day, Lord's [chi-squared] and Raju's area…
Descriptors: Test Bias, Mathematics Tests, Test Items, Gender Differences
Lombardi, Allison R.; Izzo, Margo V.; Rifenbark, Graham G.; Murray, Alexa; Buck, Andrew; Johnson, Victor – Career Development and Transition for Exceptional Individuals, 2017
Information technology (IT) literacy skills are increasingly important for all adolescents to learn, as the majority of post-school pursuits will require at least some amount of computer skills. For adolescents with disabilities, this urgency is perhaps more pronounced, as this subpopulation typically experiences more dismal post-school outcomes…
Descriptors: Computer Literacy, Measures (Individuals), Psychometrics, Information Technology
Vangrieken, Katrien; Boon, Anne; Dochy, Filip; Kyndt, Eva – Frontline Learning Research, 2017
The current gap between traditional team research and research focusing on non-strict teams or groups such as teacher teams hampers boundary-crossing investigations of and theorising on teamwork and collaboration. The main aim of this study includes bridging this gap by proposing a continuum-based team concept, describing the distinction between…
Descriptors: Teamwork, Teacher Researchers, Teacher Collaboration, Questionnaires
Çokluk, Ömay; Gül, Emrah; Dogan-Gül, Çilem – Educational Sciences: Theory and Practice, 2016
The study aims to examine whether differential item function is displayed in three different test forms that have item orders of random and sequential versions (easy-to-hard and hard-to-easy), based on Classical Test Theory (CTT) and Item Response Theory (IRT) methods and bearing item difficulty levels in mind. In the correlational research, the…
Descriptors: Test Bias, Test Items, Difficulty Level, Test Theory
Muhson, Ali; Lestari, Barkah; Supriyanto; Baroroh, Kiromim – International Journal of Instruction, 2017
Item analysis has essential roles in the learning assessment. The item analysis program is designed to measure student achievement and instructional effectiveness. This study was aimed to develop item-analysis program and verify its feasibility. This study uses a Research and Development (R & D) model. The procedure includes designing and…
Descriptors: Item Analysis, Questionnaires, Interviews, Test Items
Sadler, Philip M.; Sonnert, Gerhard; Coyle, Harold P.; Miller, Kelly A. – Educational Assessment, 2016
The psychometrically sound development of assessment instruments requires pilot testing of candidate items as a first step in gauging their quality, typically a time-consuming and costly effort. Crowdsourcing offers the opportunity for gathering data much more quickly and inexpensively than from most targeted populations. In a simulation of a…
Descriptors: Test Items, Test Construction, Psychometrics, Biological Sciences