Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 6 |
Since 2006 (last 20 years) | 19 |
Descriptor
Computer Assisted Testing | 22 |
Correlation | 22 |
Validity | 22 |
Scoring | 9 |
Comparative Analysis | 6 |
Foreign Countries | 6 |
Language Tests | 6 |
Reliability | 6 |
Scores | 6 |
Second Language Learning | 5 |
English (Second Language) | 4 |
More ▼ |
Source
Author
Attali, Yigal | 2 |
Apple, Kristen | 1 |
Ardison, Sharon | 1 |
Baldwin, Peter | 1 |
Beaty, Roger E. | 1 |
Berisha, Visar | 1 |
Bisanz, Jeffrey | 1 |
Blackwell, Lisa | 1 |
Breyer, F. Jay | 1 |
Chen, Li-Ju | 1 |
Chen, Yan-Lin | 1 |
More ▼ |
Publication Type
Journal Articles | 20 |
Reports - Research | 17 |
Reports - Evaluative | 4 |
Dissertations/Theses -… | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 5 |
Postsecondary Education | 4 |
Elementary Education | 2 |
Elementary Secondary Education | 2 |
Secondary Education | 2 |
Adult Education | 1 |
Grade 10 | 1 |
Grade 2 | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
High Schools | 1 |
More ▼ |
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Beaty, Roger E.; Johnson, Dan R.; Zeitlen, Daniel C.; Forthmann, Boris – Creativity Research Journal, 2022
Semantic distance is increasingly used for automated scoring of originality on divergent thinking tasks, such as the Alternate Uses Task (AUT). Despite some psychometric support for semantic distance -- including positive correlations with human creativity ratings -- additional work is needed to optimize its reliability and validity, including…
Descriptors: Semantics, Scoring, Creative Thinking, Creativity
LaVoie, Noelle; Parker, James; Legree, Peter J.; Ardison, Sharon; Kilcullen, Robert N. – Educational and Psychological Measurement, 2020
Automated scoring based on Latent Semantic Analysis (LSA) has been successfully used to score essays and constrained short answer responses. Scoring tests that capture open-ended, short answer responses poses some challenges for machine learning approaches. We used LSA techniques to score short answer responses to the Consequences Test, a measure…
Descriptors: Semantics, Evaluators, Essays, Scoring
Dalton, Sarah Grace; Stark, Brielle C.; Fromm, Davida; Apple, Kristen; MacWhinney, Brian; Rensch, Amanda; Rowedder, Madyson – Journal of Speech, Language, and Hearing Research, 2022
Purpose: The aim of this study was to advance the use of structured, monologic discourse analysis by validating an automated scoring procedure for core lexicon (CoreLex) using transcripts. Method: Forty-nine transcripts from persons with aphasia and 48 transcripts from persons with no brain injury were retrieved from the AphasiaBank database. Five…
Descriptors: Validity, Discourse Analysis, Databases, Scoring
Jiao, Yishan; LaCross, Amy; Berisha, Visar; Liss, Julie – Journal of Speech, Language, and Hearing Research, 2019
Purpose: Subjective speech intelligibility assessment is often preferred over more objective approaches that rely on transcript scoring. This is, in part, because of the intensive manual labor associated with extracting objective metrics from transcribed speech. In this study, we propose an automated approach for scoring transcripts that provides…
Descriptors: Suprasegmentals, Phonemes, Error Patterns, Scoring
Porter, Tenelle; Molina, Diego Catalán; Blackwell, Lisa; Roberts, Sylvia; Quirk, Abigail; Duckworth, Angela L.; Trzesniewski, Kali – Journal of Learning Analytics, 2020
Mastery behaviours -- seeking out challenging tasks and continuing to work on them despite difficulties -- are integral to achievement but difficult to measure with precision. The current study reports on the development and validation of the computer-based persistence, effort, resilience, and challenge-seeking (PERC) task in two demographically…
Descriptors: Mastery Learning, Resilience (Psychology), Difficulty Level, Computer Assisted Instruction
Ihme, Jan Marten; Senkbeil, Martin; Goldhammer, Frank; Gerick, Julia – European Educational Research Journal, 2017
The combination of different item formats is found quite often in large scale assessments, and analyses on the dimensionality often indicate multi-dimensionality of tests regarding the task format. In ICILS 2013, three different item types (information-based response tasks, simulation tasks, and authoring tasks) were used to measure computer and…
Descriptors: Foreign Countries, Computer Literacy, Information Literacy, International Assessment
Ventura, Matthew; Shute, Valerie; Zhao, Weinan – Computers & Education, 2013
An online performance-based measure of persistence was developed using anagrams and riddles. Persistence was measured by recording the time spent on unsolved anagrams and riddles. Time spent on unsolved problems was correlated to a self-report measure of persistence. Additionally, frequent video game players spent longer times on unsolved problems…
Descriptors: Correlation, Video Games, Computer Assisted Testing, Test Construction
Breyer, F. Jay; Attali, Yigal; Williamson, David M.; Ridolfi-McCulla, Laura; Ramineni, Chaitanya; Duchnowski, Matthew; Harris, April – ETS Research Report Series, 2014
In this research, we investigated the feasibility of implementing the "e-rater"® scoring engine as a check score in place of all-human scoring for the "Graduate Record Examinations"® ("GRE"®) revised General Test (rGRE) Analytical Writing measure. This report provides the scientific basis for the use of e-rater as a…
Descriptors: Computer Software, Computer Assisted Testing, Scoring, College Entrance Examinations
Watchorn, Rebecca P. D.; Bisanz, Jeffrey; Fast, Lisa; LeFevre, Jo-Anne; Skwarchuk, Sheri-Lynn; Smith-Chant, Brenda L. – Journal of Cognition and Development, 2014
The principle of "inversion," that a + b - b "must" equal a, is a fundamental property of arithmetic, but many children fail to apply it in symbolic contexts through 10 years of age. We explore three hypotheses relating to the use of inversion that stem from a model proposed by Siegler and Araya (2005). Hypothesis 1 is that…
Descriptors: Mathematics Skills, Skill Development, Computation, Attention Control
Suzuki, Yuichi; DeKeyser, Robert – Language Learning, 2015
The present study challenges the validity of elicited imitation (EI) as a measure for implicit knowledge, investigating to what extent online error detection and subsequent sentence repetition draw on implicit knowledge. To assess online detection during listening, a word monitoring component was built into an EI task. Advanced-level Japanese L2…
Descriptors: Comparative Analysis, Validity, Second Language Learning, Correlation
Harik, Polina; Baldwin, Peter; Clauser, Brian – Applied Psychological Measurement, 2013
Growing reliance on complex constructed response items has generated considerable interest in automated scoring solutions. Many of these solutions are described in the literature; however, relatively few studies have been published that "compare" automated scoring strategies. Here, comparisons are made among five strategies for…
Descriptors: Computer Assisted Testing, Automation, Scoring, Comparative Analysis
Weigle, Sara Cushing – Language Testing, 2010
Automated scoring has the potential to dramatically reduce the time and costs associated with the assessment of complex skills such as writing, but its use must be validated against a variety of criteria for it to be accepted by test users and stakeholders. This study approaches validity by comparing human and automated scores on responses to…
Descriptors: Correlation, Validity, Writing Ability, English (Second Language)
Davis, Lawrence Edward – ProQuest LLC, 2012
Speaking performance tests typically employ raters to produce scores; accordingly, variability in raters' scoring decisions has important consequences for test reliability and validity. One such source of variability is the rater's level of expertise in scoring. Therefore, it is important to understand how raters' performance is influenced by…
Descriptors: Evaluators, Expertise, Scores, Second Language Learning
Cox, Troy L.; Davies, Randall S. – CALICO Journal, 2012
This study examined the use of automatic speech recognition (ASR) scored elicited oral response (EOR) tests to assess the speaking ability of English language learners. It also examined the relationship between ASR-scored EOR and other language proficiency measures and the ability of the ASR to rate speakers without bias to gender or native…
Descriptors: English Language Learners, Computer Assisted Testing, Computer Software, Computer Software Evaluation
Yen, Yung-Chin; Ho, Rong-Guey; Chen, Li-Ju; Chou, Kun-Yi; Chen, Yan-Lin – Educational Technology & Society, 2010
The purpose of this study was to examine whether the efficiency, precision, and validity of computerized adaptive testing (CAT) could be improved by assessing confidence differences in knowledge that examinees possessed. We proposed a novel polytomous CAT model called the confidence-weighting computerized adaptive testing (CWCAT), which combined a…
Descriptors: Foreign Countries, English (Second Language), Second Language Learning, Item Response Theory
Previous Page | Next Page »
Pages: 1 | 2