Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 6 |
Since 2016 (last 10 years) | 17 |
Since 2006 (last 20 years) | 32 |
Descriptor
Test Items | 216 |
Test Validity | 162 |
Test Construction | 108 |
Test Reliability | 68 |
Higher Education | 51 |
Item Analysis | 47 |
Difficulty Level | 42 |
Item Response Theory | 26 |
Test Format | 26 |
Construct Validity | 25 |
Achievement Tests | 24 |
More ▼ |
Source
Author
Publication Type
Education Level
Audience
Researchers | 25 |
Practitioners | 1 |
Teachers | 1 |
Location
Netherlands | 4 |
Canada | 3 |
Tennessee | 3 |
Turkey | 3 |
Alabama | 2 |
Arkansas | 2 |
Australia | 2 |
Florida | 2 |
Taiwan | 2 |
Arizona | 1 |
Georgia | 1 |
More ▼ |
Laws, Policies, & Programs
Comprehensive Education… | 2 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Martha L. Epstein; Hamza Malik; Kun Wang; Chandra Hawley Orrill – Grantee Submission, 2022
Response Process Validity (RPV) reflects the degree to which items are interpreted as intended by item developers. In this study, teacher responses to constructed response (CR) items to assess pedagogical content knowledge (PCK) of middle school mathematics teachers were evaluated to determine what types of teacher responses signaled weak RPV. We…
Descriptors: Teacher Response, Test Items, Pedagogical Content Knowledge, Mathematics Teachers
Karen Leary Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Assessment is a topic of concern to all stakeholders in our educational system. Pattern Based Questions are an assessment tool which is an alternative to the standardized assessment tool, and they are based on generative learning pedagogy, which shows promise in engaging all learners and usefulness in teaching and learning but validity has not yet…
Descriptors: Undergraduate Students, College Mathematics, Mathematics Skills, Thinking Skills
Yazicilar Nalbantoglu, Ümran; Bümen, Nilay T.; Uslu, Öner – Teacher Development, 2022
Curriculum adaptation portrays the way teachers 'tweak' the curriculum. The findings of related studies are based solely on qualitative data, and there is no instrument that can determine curriculum adaptation patterns. Based on this gap, the aim of this study is to develop the Curriculum Adaptation Patterns Scale (CAPS). After examining the…
Descriptors: Curriculum Development, Test Construction, Factor Analysis, Test Reliability
Akhtar, Hanif – International Association for Development of the Information Society, 2022
When examinees perceive a test as low stakes, it is logical to assume that some of them will not put out their maximum effort. This condition makes the validity of the test results more complicated. Although many studies have investigated motivational fluctuation across tests during a testing session, only a small number of studies have…
Descriptors: Intelligence Tests, Student Motivation, Test Validity, Student Attitudes
Baskonus, Turan; Soyer, Fikret – International Journal of Psychology and Educational Studies, 2020
The aim of this study is to develop a scale that measures the attitudes of physical education and sports teachers towards measurement and evaluation. In this study, scale development principles and steps of DeVellis (2017) were used. Initially, a literature review was conducted, 19 physical education and sports teachers were interviewed in written…
Descriptors: Test Construction, Teacher Attitudes, Physical Education Teachers, Test Items
Aleyna Altan; Zehra Taspinar Sener – Online Submission, 2023
This research aimed to develop a valid and reliable test to be used to detect sixth grade students' misconceptions and errors regarding the subject of fractions. A misconception diagnostic test has been developed that includes the concept of fractions, different representations of fractions, ordering and comparing fractions, equivalence of…
Descriptors: Diagnostic Tests, Mathematics Tests, Fractions, Misconceptions
Benton, Tom – Cambridge Assessment, 2018
One of the questions with the longest history in educational assessment is whether it is possible to increase the reliability of a test simply by altering the way in which scores on individual test items are combined to make the overall test score. Most usually, the score available on each item is communicated to the candidate within a question…
Descriptors: Test Items, Scoring, Predictive Validity, Test Reliability
Zhang, Tan; Chen, Ang – AERA Online Paper Repository, 2016
Based on the Job Demands-Resources model, the study developed and validated an instrument that measures physical education teachers' job demands/resources perception. Expert review established content validity with the average item rating of 3.6/5.0. Construct validity and reliability were determined with a teacher sample (n=193). Exploratory…
Descriptors: Physical Education Teachers, Teaching Load, Resources, Measures (Individuals)
Validating a Claim-Evidence-Science Idea-Reasoning (CESR) Framework for Use in NGSS Assessment Tasks
Hardcastle, Joseph M.; Herrmann Abell, Cari F.; DeBoer, George E. – Grantee Submission, 2021
We developed assessment tasks aligned to the Next Generation Science Standards (NGSS) that require students to use argumentation and explanation practices along with disciplinary core ideas and crosscutting concepts to make sense of energy-related phenomena. Scoring rubrics were created to evaluate students' ability to make accurate claims, cite…
Descriptors: Academic Standards, Energy, Scientific Concepts, Persuasive Discourse
Mix, Daniel F.; Tao, Shuqin – AERA Online Paper Repository, 2017
Purposes: This study uses think-alouds and cognitive interviews to provide validity evidence for an online formative assessment--i-Ready Standards Mastery (iSM) mini-assessments--which involves a heavy use of innovative items. iSM mini-assessments are intended to help teachers determine student understanding of each of the on-grade-level Common…
Descriptors: Formative Evaluation, Computer Assisted Testing, Test Validity, Student Evaluation
Lee, Hollylynne; Bradshaw, Laine; Famularo, Lisa; Masters, Jessica; Azevedo, Roger; Johnson, Sheri; Schellman, Madeline; Elrod, Emily; Sanei, Hamid – Grantee Submission, 2019
The research shared in this conference paper report illustrates how an iterative process to item development that involves expert review and cognitive lab interviews with students can be used to collect evidence of validity for assessment items. Analysis of students' reasoning was also used to expand a model for identifying conceptions and…
Descriptors: Middle School Students, Interviews, Misconceptions, Test Items
Perry, Lindsey – AERA Online Paper Repository, 2017
Before an assessment is used to make decisions, the validity of the intended interpretation must be evaluated. The purpose of this paper is to describe how the argument-based approach and an interpretation/use argument (IUA) (Kane, 2013) were used to validate the interpretations made from the new Early Grade Mathematics Assessment (EGMA)…
Descriptors: Student Evaluation, Mathematics Tests, Test Interpretation, Inferences
Kutlu, Omer; Yavuz, Hatice Cigdem – International Journal of Assessment Tools in Education, 2019
Studies based on response processes of individuals can provide information that supports the assessment and increases the validity of the items in the scale or tests. The purpose of this study is to present the extent to which the student response processes are effective in identifying and developing the characteristics of the items in an…
Descriptors: Test Validity, Test Items, Achievement Tests, Grade 4
Continual Improvement of a Student Evaluation of Teaching over Seven Semesters at a State University
Rates, Christopher; Liu, Xiufeng; Vanzile-Tamzen, Carol; Morreale, Cathleen – AERA Online Paper Repository, 2017
In the fall of 2014, the University at Buffalo created a new universal Student Evaluation of Teaching (SET). The purpose of the present study was to establish the construct validity of SET items. Rasch analyses of data from 7 semesters (N=203,194 students) revealed problems with item fit indices and threshold distances. Changes to items and…
Descriptors: Student Evaluation of Teacher Performance, State Universities, College Students, Teacher Effectiveness
Evaluating the Design and Development of an Adaptive E-Tutorial Module: A Rasch-Measurement Approach
Barefah, Allaa; McKay, Elspeth – International Association for Development of the Information Society, 2016
Courseware designers aim to innovate information communications technology (ICT) tools to increase learning experiences, spending many hours developing eLearning programmes. This effort gives rise to a dynamic technological pedagogical environment. However, it is difficult to recognise whether these online programmes reflect an instructional…
Descriptors: Electronic Learning, Courseware, Instructional Design, Quasiexperimental Design