NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 8 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Asassfeh, Sahail; Al-Ebous, Hana'; Khwaileh, Faisal; Al-Zoubi, Zohair – Educational Studies, 2014
This study is the first to address student evaluation of faculty members (SFE) from a student perspective at a major Jordanian public university using a comprehensive (71-item) questionnaire administered to 620 undergraduates. Addressed are students' perceptions of the SFE process in terms of: (a) their paper-based vs. online-format preferences;…
Descriptors: Foreign Countries, Student Evaluation of Teacher Performance, Universities, Student Attitudes
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Jungtae; Craig, Daniel A. – Computer Assisted Language Learning, 2012
Videoconferencing offers new opportunities for language testers to assess speaking ability in low-stakes diagnostic tests. To be considered a trusted testing tool in language testing, a test should be examined employing appropriate validation processes [Chapelle, C.A., Jamieson, J., & Hegelheimer, V. (2003). "Validation of a web-based ESL…
Descriptors: Speech Communication, Testing, Language Tests, Construct Validity
Walt, Nancy; Atwood, Kristin; Mann, Alex – Journal of Technology, Learning, and Assessment, 2008
The purpose of this study was to determine whether or not survey medium (electronic versus paper format) has a significant effect on the results achieved. To compare survey media, responses from elementary students to British Columbia's Satisfaction Survey were analyzed. Although this study was not experimental in design, the data set served as a…
Descriptors: Student Attitudes, Factor Analysis, Foreign Countries, Elementary School Students
Kim, Do-Hong; Huynh, Huynh – Journal of Technology, Learning, and Assessment, 2007
This study examined comparability of student scores obtained from computerized and paper-and-pencil formats of the large-scale statewide end-of-course (EOC) examinations in the two subject areas of Algebra and Biology. Evidence in support of comparability of computerized and paper-based tests was sought by examining scale scores, item parameter…
Descriptors: Computer Assisted Testing, Measures (Individuals), Biology, Algebra
Puhan, Gautam; Boughton, Keith; Kim, Sooyeon – Journal of Technology, Learning, and Assessment, 2007
The study evaluated the comparability of two versions of a certification test: a paper-and-pencil test (PPT) and computer-based test (CBT). An effect size measure known as Cohen's d and differential item functioning (DIF) analyses were used as measures of comparability at the test and item levels, respectively. Results indicated that the effect…
Descriptors: Computer Assisted Testing, Effect Size, Test Bias, Mathematics Tests
Wang, Jinhao; Brown, Michelle Stallone – Journal of Technology, Learning, and Assessment, 2007
The current research was conducted to investigate the validity of automated essay scoring (AES) by comparing group mean scores assigned by an AES tool, IntelliMetric [TM] and human raters. Data collection included administering the Texas version of the WriterPlacer "Plus" test and obtaining scores assigned by IntelliMetric [TM] and by…
Descriptors: Test Scoring Machines, Scoring, Comparative Testing, Intermode Differences
Peer reviewed Peer reviewed
Direct linkDirect link
Lai, Ah-Fur; Chen, Deng-Jyi; Chen, Shu-Ling – Journal of Educational Multimedia and Hypermedia, 2008
The IRT (Item Response Theory) has been studied and applied in computer-based test for decades. However, almost of all these existing studies evaluated focus merely on test questions with text-based (or static text/graphic) type of presentation form illustrated exclusively. In this paper, we present our study on test questions using both…
Descriptors: Elementary School Students, Semantics, Difficulty Level, Item Response Theory
Peer reviewed Peer reviewed
Schuldberg, David – Computers in Human Behavior, 1988
Describes study that investigated the effects of computerized test administration on undergraduates' responses to the Minnesota Multiphasic Personality Inventory (MMPI), and discusses methodological considerations important in evaluating the sensitivity of personality inventories in different administration formats. Results analyze the effects of…
Descriptors: Analysis of Variance, Comparative Testing, Computer Assisted Testing, Higher Education