NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Students1
Laws, Policies, & Programs
Every Student Succeeds Act…2
What Works Clearinghouse Rating
Showing 1 to 15 of 20 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Jonas Flodén – British Educational Research Journal, 2025
This study compares how the generative AI (GenAI) large language model (LLM) ChatGPT performs in grading university exams compared to human teachers. Aspects investigated include consistency, large discrepancies and length of answer. Implications for higher education, including the role of teachers and ethics, are also discussed. Three…
Descriptors: College Faculty, Artificial Intelligence, Comparative Testing, Scoring
Aaron McVay – ProQuest LLC, 2021
As assessments move towards computerized testing and making continuous testing available the need for rapid assembly of forms is increasing. The objective of this study was to investigate variability in assembled forms through the lens of first- and second-order equity properties of equating, by examining three factors and their interactions. Two…
Descriptors: Automation, Computer Assisted Testing, Test Items, Reaction Time
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Dandan; Hebert, Michael; Wilson, Joshua – American Educational Research Journal, 2022
We used multivariate generalizability theory to examine the reliability of hand-scoring and automated essay scoring (AES) and to identify how these scoring methods could be used in conjunction to optimize writing assessment. Students (n = 113) included subsamples of struggling writers and non-struggling writers in Grades 3-5 drawn from a larger…
Descriptors: Reliability, Scoring, Essays, Automation
Peer reviewed Peer reviewed
Direct linkDirect link
Kershree Padayachee; M. Matimolane – Teaching in Higher Education, 2025
In the shift to Emergency Remote Teaching and Learning (ERT&L) during the COVID-19 pandemic, remote assessment and feedback became a major source of discontent and challenge for students and staff. This paper is a reflection and analysis of assessment practices during ERT&L, and our theorisation of the possibilities for shifts towards…
Descriptors: Educational Quality, Social Justice, Distance Education, Feedback (Response)
Peer reviewed Peer reviewed
Direct linkDirect link
Lam, Elizabeth A.; Rose, Susan; McMaster, Kristen L. – Journal of Deaf Studies and Deaf Education, 2020
This study compared the reliability and validity of student scores from paper--pencil and e-based assessments using the "maze" and "silent reading fluency" (SRF) tasks. Forty students who were deaf and hard of hearing and reading between the second and fifth grade reading levels and their teachers (n = 21) participated. For…
Descriptors: Deafness, Hearing Impairments, Curriculum Based Assessment, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Dalton, Sarah Grace; Stark, Brielle C.; Fromm, Davida; Apple, Kristen; MacWhinney, Brian; Rensch, Amanda; Rowedder, Madyson – Journal of Speech, Language, and Hearing Research, 2022
Purpose: The aim of this study was to advance the use of structured, monologic discourse analysis by validating an automated scoring procedure for core lexicon (CoreLex) using transcripts. Method: Forty-nine transcripts from persons with aphasia and 48 transcripts from persons with no brain injury were retrieved from the AphasiaBank database. Five…
Descriptors: Validity, Discourse Analysis, Databases, Scoring
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Allehaiby, Wid Hasen; Al-Bahlani, Sara – Arab World English Journal, 2021
One of the main challenges higher educational institutions encounter amid the recent COVID-19 crisis is transferring assessment approaches from the traditional face-to-face form to the online Emergency Remote Teaching approach. A set of language assessment principles, practicality, reliability, validity, authenticity, and washback, which can be…
Descriptors: Barriers, Distance Education, Evaluation Methods, Teaching Methods
Salem, Walaa Abdel Fattah – Online Submission, 2021
The study aim was examining the effect of a program based on Cloud-computing platform on developing assessment strategies for EFL primary school teachers. The study reviewed the literature and previous studies dealing with assessment strategies (self-assessment, peer-assessment, strategic use of Questioning and reflection), professional learning…
Descriptors: Second Language Learning, Second Language Instruction, English (Second Language), Elementary School Teachers
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Darling-Hammond, Linda – Learning Policy Institute, 2017
After passage of the Every Student Succeeds Act (ESSA) in 2015, states assumed greater responsibility for designing their own accountability and assessment systems. ESSA requires states to measure "higher order thinking skills and understanding" and encourages the use of open-ended performance assessments, which are essential for…
Descriptors: Performance Based Assessment, Accountability, Portfolios (Background Materials), Task Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ghilay, Yaron; Ghilay, Ruth – Journal of Educational Technology, 2012
The study examined advantages and disadvantages of computerised assessment compared to traditional evaluation. It was based on two samples of college students (n=54) being examined in computerised tests instead of paper-based exams. Students were asked to answer a questionnaire focused on test effectiveness, experience, flexibility and integrity.…
Descriptors: Student Evaluation, Higher Education, Comparative Analysis, Computer Assisted Testing
Darling-Hammond, Linda – Council of Chief State School Officers, 2017
The Every Student Succeeds Act (ESSA) opened up new possibilities for how student and school success are defined and supported in American public education. States have greater responsibility for designing and building their assessment and accountability systems. These new opportunities to develop performance assessments are critically important…
Descriptors: Performance Based Assessment, Accountability, Portfolios (Background Materials), Task Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Kettler, Ryan J.; Elliott, Stephen N.; Kurz, Alexander; Zigmond, Naomi; Lemons, Christopher J.; Kloo, Amanda; Shrago, Jacqueline; Beddow, Peter A.; Williams, Leila; Bruen, Charles; Lupp, Lynda; Farmer, Jeanie; Mosiman, Melanie – Assessment for Effective Intervention, 2014
Motivated by the multiple-measures clause of recent federal policy regarding student eligibility for alternate assessments based on modified academic achievement standards (AA-MASs), this study examined how scores or combinations of scores from a diverse set of assessments predicted students' end-of-year proficiency status on statewide achievement…
Descriptors: Eligibility, Alternative Assessment, Academic Achievement, Predictive Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Jordan, Nancy C.; Glutting, Joseph; Ramineni, Chaitanya; Watkins, Marley W. – School Psychology Review, 2010
Using a longitudinal design, children were given a brief number sense screener (NSB) screener (N = 204) over six time points, from the beginning of kindergarten to the middle of first grade. The NSB is based on research showing the importance of number competence (number, number relations, and number operations) for success in mathematics.…
Descriptors: Mathematics Achievement, Reliability, Achievement Tests, Kindergarten
Peer reviewed Peer reviewed
Direct linkDirect link
Chang, Shu-Ren; Plake, Barbara S.; Kramer, Gene A.; Lien, Shu-Mei – Educational and Psychological Measurement, 2011
This study examined the amount of time that different ability-level examinees spend on questions they answer correctly or incorrectly across different pretest item blocks presented on a fixed-length, time-restricted computerized adaptive testing (CAT). Results indicate that different ability-level examinees require different amounts of time to…
Descriptors: Evidence, Test Items, Reaction Time, Adaptive Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Arce-Ferrer, Alvaro J.; Guzman, Elvira Martinez – Educational and Psychological Measurement, 2009
This study investigates the effect of mode of administration of the Raven Standard Progressive Matrices test on distribution, accuracy, and meaning of raw scores. A random sample of high school students take counterbalanced paper-and-pencil and computer-based administrations of the test and answer a questionnaire surveying preferences for…
Descriptors: Factor Analysis, Raw Scores, Statistical Analysis, Computer Assisted Testing
Previous Page | Next Page »
Pages: 1  |  2