NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 25 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kim, Sooyeon; Walker, Michael – ETS Research Report Series, 2021
In this investigation, we used real data to assess potential differential effects associated with taking a test in a test center (TC) versus testing at home using remote proctoring (RP). We used a pseudo-equivalent groups (PEG) approach to examine group equivalence at the item level and the total score level. If our assumption holds that the PEG…
Descriptors: Testing, Distance Education, Comparative Analysis, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mendoza, Arturo; Martínez, Joaquín – International Journal of Language Testing, 2023
Language placement tests (LPTs) are used to assess students' proficiency in a progressive manner in the target language. Based on their performance, students are assigned to stepped language courses. These tests are usually considered low stakes because they do not have significant consequences in students' lives, which is perhaps the reason why…
Descriptors: Language Tests, English (Second Language), Second Language Learning, Second Language Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational Measurement, 2017
Person-fit assessment (PFA) is concerned with uncovering atypical test performance as reflected in the pattern of scores on individual items on a test. Existing person-fit statistics (PFSs) include both parametric and nonparametric statistics. Comparison of PFSs has been a popular research topic in PFA, but almost all comparisons have employed…
Descriptors: Goodness of Fit, Testing, Test Items, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chen, Haiwen H.; von Davier, Matthias; Yamamoto, Kentaro; Kong, Nan – ETS Research Report Series, 2015
One major issue with large-scale assessments is that the respondents might give no responses to many items, resulting in less accurate estimations of both assessed abilities and item parameters. This report studies how the types of items affect the item-level nonresponse rates and how different methods of treating item-level nonresponses have an…
Descriptors: Achievement Tests, Foreign Countries, International Assessment, Secondary School Students
McQuillan, Mark; Phelps, Richard P.; Stotsky, Sandra – Pioneer Institute for Public Policy Research, 2015
In July 2010, the Massachusetts Board of Elementary and Secondary Education (BESE) voted to adopt Common Core's standards in English language arts (ELA) and mathematics in place of the state's own standards in these two subjects. The vote was based largely on recommendations by Commissioner of Education Mitchell Chester and then Secretary of…
Descriptors: Reading Tests, Writing Tests, Achievement Tests, Common Core State Standards
Lee, Jo Ann; And Others – 1984
The difficulty of test items administered by paper and pencil were compared with the difficulty of the same items administered by computer. The study was conducted to determine if an interaction exists between mode of test administration and ability. An arithmetic reasoning test was constructed for this study. All examinees had taken the Armed…
Descriptors: Adults, Comparative Analysis, Computer Assisted Testing, Difficulty Level
McArthur, David L. – 1981
Item bias, when present in a multiple-choice test, can be detected by appropriate analyses of the persons x items scoring matrix. Five related schemes for the statistical analysis of bias were applied to a widely used, primary skills multiple-choice test which was administered in either its English- or Spanish-language version at each of the two…
Descriptors: Comparative Analysis, Elementary Education, Multiple Choice Tests, Spanish
Peer reviewed Peer reviewed
Jaeger, Richard M. – Journal of Educational Measurement, 1981
Five indices are discussed that should logically discriminate between situations in which: (1) the linear equating method (LEM) adequately adjusts for difference between score distributions of two approximately parallel test forms; or (2) a method more complex than the linear equating method is needed. (RL)
Descriptors: College Entrance Examinations, Comparative Analysis, Difficulty Level, Equated Scores
Peer reviewed Peer reviewed
Lord, Frederic M. – Educational and Psychological Measurement, 1971
A number of empirical studies are suggested to answer certain questions in connection with flexilevel tests. (MS)
Descriptors: Comparative Analysis, Difficulty Level, Guessing (Tests), Item Analysis
Stocking, Martha L. – 1996
The interest in the application of large-scale computerized adaptive testing has served to focus attention on issues that arise when theoretical advances are made operational. Some of these issues stem less from changes in testing conditions and more from changes in testing paradigms. One such issue is that of the order in which questions are…
Descriptors: Adaptive Testing, Cognitive Processes, Comparative Analysis, Computer Assisted Testing
Wilcox, Rand R. – 1978
Two fundamental problems in mental test theory are to estimate true score and to estimate the amount of error when testing an examinee. In this report, three probability models which characterize a single test item in terms of a population of examinees are described. How these models may be modified to characterize a single examinee in terms of an…
Descriptors: Achievement Tests, Comparative Analysis, Error of Measurement, Mathematical Models
Merz, William R. – 1980
Several methods of assessing test item bias are described, and the concept of fair use of tests is examined. A test item is biased if individuals of equal ability have different probabilities of attaining the item correct. The following seven general procedures used to examine test items for bias are summarized and discussed: (1) analysis of…
Descriptors: Comparative Analysis, Evaluation Methods, Factor Analysis, Mathematical Models
Angoff, William H.; Schrader, William B. – 1982
In a study to determine whether a shift from Formula scoring to Rights scoring can be made without causing a discontinuity in the test scale, the analysis of special administrations of the Scholastic Aptitude Test and Chemistry Achievement Test and the variable section of an operational form of the Graduate Management Admission Test (GMAT) is…
Descriptors: Comparative Analysis, Equated Scores, Guessing (Tests), Higher Education
Kim, Seock-Ho; Cohen, Allan S. – 1996
Applications of item response theory to practical testing problems including equating, differential item functioning, and computerized adaptive testing, require that item parameter estimates be placed onto a common metric. In this study, three methods for developing a common metric under item response theory are compared: (1) linking separate…
Descriptors: Adaptive Testing, Comparative Analysis, Computer Assisted Testing, Difficulty Level
Pommerich, Mary; Burden, Timothy – 2000
A small-scale study was conducted to compare test-taking strategies, problem-solving strategies, and general impressions about the test across computer and paper-and-pencil administration modes. Thirty-six examinees (high school students) participated in the study. Each examinee took a test in one of the content areas of English, Mathematics,…
Descriptors: Adaptive Testing, Attitudes, Comparative Analysis, Computer Assisted Testing
Previous Page | Next Page »
Pages: 1  |  2