NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 121 to 135 of 9,400 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hojung Kim; Changkyung Song; Jiyoung Kim; Hyeyun Jeong; Jisoo Park – Language Testing in Asia, 2024
This study presents a modified version of the Korean Elicited Imitation (EI) test, designed to resemble natural spoken language, and validates its reliability as a measure of proficiency. The study assesses the correlation between average test scores and Test of Proficiency in Korean (TOPIK) levels, examining score distributions among beginner,…
Descriptors: Korean, Test Validity, Test Reliability, Imitation
Peer reviewed Peer reviewed
Direct linkDirect link
Fu Chen; Ying Cui; Alina Lutsyk-King; Yizhu Gao; Xiaoxiao Liu; Maria Cutumisu; Jacqueline P. Leighton – Education and Information Technologies, 2024
Post-secondary data literacy education is critical to students' academic and career success. However, the literature has not adequately addressed the conceptualization and assessment of data literacy for post-secondary students. In this study, we introduced a novel digital performance-based assessment for teaching and evaluating post-secondary…
Descriptors: Performance Based Assessment, College Students, Information Literacy, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Jianbin Fu; Patrick C. Kyllonen; Xuan Tan – Measurement: Interdisciplinary Research and Perspectives, 2024
Users of forced-choice questionnaires (FCQs) to measure personality commonly assume statement parameter invariance across contexts -- between Likert and forced-choice (FC) items and between different FC items that share a common statement. In this paper, an empirical study was designed to check these two assumptions for an FCQ assessment measuring…
Descriptors: Measurement Techniques, Questionnaires, Personality Measures, Interpersonal Competence
Peer reviewed Peer reviewed
Direct linkDirect link
Pan, Yiqin; Wollack, James A. – Educational Measurement: Issues and Practice, 2023
Pan and Wollack (PW) proposed a machine learning method to detect compromised items. We extend the work of PW to an approach detecting compromised items and examinees with item preknowledge simultaneously and draw on ideas in ensemble learning to relax several limitations in the work of PW. The suggested approach also provides a confidence score,…
Descriptors: Artificial Intelligence, Prior Learning, Item Analysis, Test Content
Peer reviewed Peer reviewed
Direct linkDirect link
Pierce, Corey D.; Epstein, Michael H.; Wood, Matthew D. – Journal of Emotional and Behavioral Disorders, 2023
Strength-based assessment has achieved acceptance from educational, mental health, and social service professionals as a means to measuring emotional and behavioral strengths of children. Several standardized, norm-referenced tests have been developed to assess these strengths; however, the primary mode of assessment is via informal interviews of…
Descriptors: Behavior Rating Scales, Content Validity, Psychometrics, Mental Health
Peer reviewed Peer reviewed
Direct linkDirect link
Welzel, Christian; Brunkert, Lennart; Kruse, Stefan; Inglehart, Ronald F. – Sociological Methods & Research, 2023
Scholars study representative international surveys to understand cross-cultural differences in mentality patterns, which are measured via complex multi-item constructs. Methodologists in this field insist with increasing vigor that detecting "non-invariance" in how a construct's items associate with each other in different national…
Descriptors: Cross Cultural Studies, Social Science Research, Factor Analysis, Measurement Techniques
Peer reviewed Peer reviewed
Direct linkDirect link
Sharma, Harsh; Mathur, Rohan; Chintala, Tejas; Dhanalakshmi, Samiappan; Senthil, Ramalingam – Education and Information Technologies, 2023
Examination assessments undertaken by educational institutions are pivotal since it is one of the fundamental steps to determining students' understanding and achievements for a distinct subject or course. Questions must be framed on the topics to meet the learning objectives and assess the student's capability in a particular subject. The…
Descriptors: Taxonomy, Student Evaluation, Test Items, Questioning Techniques
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J.; Belov, Dmitry I. – Journal of Educational Measurement, 2023
A test of item compromise is presented which combines the test takers' responses and response times (RTs) into a statistic defined as the number of correct responses on the item for test takers with RTs flagged as suspicious. The test has null and alternative distributions belonging to the well-known family of compound binomial distributions, is…
Descriptors: Item Response Theory, Reaction Time, Test Items, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Abdolvahab Khademi; Craig S. Wells; Maria Elena Oliveri; Ester Villalonga-Olives – SAGE Open, 2023
The most common effect size when using a multiple-group confirmatory factor analysis approach to measurement invariance is [delta]CFI and [delta]TLI with a cutoff value of 0.01. However, this recommended cutoff value may not be ubiquitously appropriate and may be of limited application for some tests (e.g., measures using dichotomous items or…
Descriptors: Factor Analysis, Factor Structure, Error of Measurement, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Kaldes, Gal; Tighe, Elizabeth; He, Qiwei – AERA Online Paper Repository, 2023
This study used PIAAC process data to examine time-related allocation patterns (time for the first action, total time, last action) of low-skilled, relative to higher-skilled, adults on digital literacy items. Results suggest that less-skilled (Level 2) and higher skilled adults (Levels 3-5) exhibited similar time allocation patterns; however,…
Descriptors: Time Management, Literacy Education, Adult Literacy, Adult Education
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Pentecost, Thomas C.; Raker, Jeffery R.; Murphy, Kristen L. – Practical Assessment, Research & Evaluation, 2023
Using multiple versions of an assessment has the potential to introduce item environment effects. These types of effects result in version dependent item characteristics (i.e., difficulty and discrimination). Methods to detect such effects and resulting implications are important for all levels of assessment where multiple forms of an assessment…
Descriptors: Item Response Theory, Test Items, Test Format, Science Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mahmut Sami Koyuncu; Mehmet Sata – International Journal of Assessment Tools in Education, 2023
The main aim of this study was to introduce the ConQuest program, which is used in the analysis of multivariate and multidimensional data structures, and to show its applications on example data structures. To achieve this goal, a basic research approach was applied. Thus, how to use the ConQuest program and how to prepare the data set for…
Descriptors: Data Analysis, Computer Oriented Programs, Models, Test Items
Paige Haley – ProQuest LLC, 2023
As the research on feigning has grown, the number and quality of performance validity tests (PVTs) has increased as well. However, while several PVTs have been developed from assessments commonly used as part of neuropsychological batteries, there has been less exploration for PVTs scored from items in cognitive screeners. The Montreal Cognitive…
Descriptors: Cognitive Measurement, Performance, Test Validity, Psychological Testing
Matthew John Davidson – ProQuest LLC, 2022
Digitally-based assessments create opportunities for collecting moment to moment information about how students are responding to assessment items. This information, called log or process data, has long been regarded as a vast and valuable source of data about student performance. Despite repeated assurances of its vastness and value, process data…
Descriptors: Data Use, Psychometrics, Item Response Theory, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Joseph A. Rios; Jiayi Deng – Educational and Psychological Measurement, 2025
To mitigate the potential damaging consequences of rapid guessing (RG), a form of noneffortful responding, researchers have proposed a number of scoring approaches. The present simulation study examines the robustness of the most popular of these approaches, the unidimensional effort-moderated (EM) scoring procedure, to multidimensional RG (i.e.,…
Descriptors: Scoring, Guessing (Tests), Reaction Time, Item Response Theory
Pages: 1  |  ...  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  12  |  13  |  ...  |  627