NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational Measurement:…25
Audience
Location
United States1
What Works Clearinghouse Rating
Showing 1 to 15 of 25 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Xuelan Qiu; Jimmy de la Torre; You-Gan Wang; Jinran Wu – Educational Measurement: Issues and Practice, 2024
Multidimensional forced-choice (MFC) items have been found to be useful to reduce response biases in personality assessments. However, conventional scoring methods for the MFC items result in ipsative data, hindering the wider applications of the MFC format. In the last decade, a number of item response theory (IRT) models have been developed,…
Descriptors: Item Response Theory, Personality Traits, Personality Measures, Personality Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Walker, Michael E. – Educational Measurement: Issues and Practice, 2022
Test equating requires collecting data to link the scores from different forms of a test. Problems arise when equating samples are not equivalent and the test forms to be linked share no common items by which to measure or adjust for the group nonequivalence. Using data from five operational test forms, we created five pairs of research forms for…
Descriptors: Ability, Tests, Equated Scores, Testing Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Angela Johnson; Elizabeth Barker; Marcos Viveros Cespedes – Educational Measurement: Issues and Practice, 2024
Educators and researchers strive to build policies and practices on data and evidence, especially on academic achievement scores. When assessment scores are inaccurate for specific student populations or when scores are inappropriately used, even data-driven decisions will be misinformed. To maximize the impact of the research-practice-policy…
Descriptors: Equal Education, Inclusion, Evaluation Methods, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Pommerich, Mary – Educational Measurement: Issues and Practice, 2012
Neil Dorans has made a career of advocating for the examinee. He continues to do so in his NCME career award address, providing a thought-provoking commentary on some current trends in educational measurement that could potentially affect the integrity of test scores. Concerns expressed in the address call attention to a conundrum that faces…
Descriptors: Testing, Scores, Measurement, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Randall, Jennifer; Sireci, Stephen; Li, Xueming; Kaira, Leah – Educational Measurement: Issues and Practice, 2012
As access and reliance on technology continue to increase, so does the use of computerized testing for admissions, licensure/certification, and accountability exams. Nonetheless, full computer-based test (CBT) implementation can be difficult due to limited resources. As a result, some testing programs offer both CBT and paper-based test (PBT)…
Descriptors: Science Tests, Computer Assisted Testing, Scores, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Mislevy, Robert J. – Educational Measurement: Issues and Practice, 2012
This article presents the author's observations on Neil Dorans's NCME Career Award Address: "The Contestant Perspective on Taking Tests: Emanations from the Statue within." He calls attention to some points that Dr. Dorans made in his address, and offers his thoughts in response.
Descriptors: Testing, Test Reliability, Psychometrics, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Dorans, Neil J. – Educational Measurement: Issues and Practice, 2012
Views on testing--its purpose and uses and how its data are analyzed--are related to one's perspective on test takers. Test takers can be viewed as learners, examinees, or contestants. I briefly discuss the perspective of test takers as learners. I maintain that much of psychometrics views test takers as examinees. I discuss test takers as a…
Descriptors: Testing, Test Theory, Item Response Theory, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip; Dorans, Neil J.; Liang, Longjuan – Educational Measurement: Issues and Practice, 2011
Over the past few decades, those who take tests in the United States have exhibited increasing diversity with respect to native language. Standard psychometric procedures for ensuring item and test fairness that have existed for some time were developed when test-taking groups were predominantly native English speakers. A better understanding of…
Descriptors: Test Bias, Testing Programs, Psychometrics, Language Proficiency
Peer reviewed Peer reviewed
Direct linkDirect link
Kato, Kentaro; Moen, Ross E.; Thurlow, Martha L. – Educational Measurement: Issues and Practice, 2009
Large data sets from a state reading assessment for third and fifth graders were analyzed to examine differential item functioning (DIF), differential distractor functioning (DDF), and differential omission frequency (DOF) between students with particular categories of disabilities (speech/language impairments, learning disabilities, and emotional…
Descriptors: Learning Disabilities, Language Impairments, Behavior Disorders, Affective Behavior
Peer reviewed Peer reviewed
Direct linkDirect link
Penfield, Randall D.; Gattamorta, Karina; Childs, Ruth A. – Educational Measurement: Issues and Practice, 2009
Traditional methods for examining differential item functioning (DIF) in polytomously scored test items yield a single item-level index of DIF and thus provide no information concerning which score levels are implicated in the DIF effect. To address this limitation of DIF methodology, the framework of differential step functioning (DSF) has…
Descriptors: Test Bias, Test Items, Evaluation Methods, Scores
Peer reviewed Peer reviewed
Wigdor, Alexandra K. – Educational Measurement: Issues and Practice, 1982
This synopsis of the Committee on Ability Testing report (see ED 213 770 and ED 213 771) discusses recent controversy over testing limitations, test fairness, admissions testing, test disclosure, and test use in public schools to improve substandard performance. (CM)
Descriptors: Ability, Advisory Committees, Aptitude Tests, College Entrance Examinations
Peer reviewed Peer reviewed
Pyburn, Keith M., Jr. – Educational Measurement: Issues and Practice, 1990
Discrepancies between court and professional association views of licensing examinations are discussed. Litigation associated with Title VII of the Civil Rights Act of 1964 and constitutional law is considered. Courts should examine professional standards and apply those that are consistent with the court's determination of the proper…
Descriptors: Civil Rights Legislation, Constitutional Law, Court Litigation, Court Role
Peer reviewed Peer reviewed
Legg, Sue M.; Buhr, Dianne C. – Educational Measurement: Issues and Practice, 1992
Three computerized adaptive tests (CATs) in mathematics, reading, and writing were administered to 628 community college students to determine whether examinees of different ethnic, gender, ability, and age groups, and computer experience were differentially affected. Some differences exist; however, they do not preclude use of CATs. (SLD)
Descriptors: Ability, Adaptive Testing, Age Differences, College Students
Peer reviewed Peer reviewed
McCarty, Joan First; Cardenas, Jose – Educational Measurement: Issues and Practice, 1986
Test scores are being used for a variety of inappropriate purposes, often to the detriment of minorities. This article recounts recent evidence on misuse of testing, offers strategies for preventing abuse, and urges professionals to take on active roles in preventing misuse of their wares. (LMO)
Descriptors: Court Litigation, Culture Fair Tests, Elementary Secondary Education, Minimum Competency Testing
Peer reviewed Peer reviewed
Hills, John R. – Educational Measurement: Issues and Practice, 1989
Test bias detection methods based on item response theory (IRT) are reviewed. Five such methods are commonly used: (1) equality of item parameters; (2) area between item characteristic curves; (3) sums of squares; (4) pseudo-IRT; and (5) one-parameter-IRT. A table compares these and six newer or less tested methods. (SLD)
Descriptors: Item Analysis, Test Bias, Test Items, Testing Programs
Previous Page | Next Page ยป
Pages: 1  |  2