Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Brownell, Sara E.; Hekmat-Scafe, Daria S.; Singla, Veena; Seawell, Patricia Chandler; Imam, Jamie F. Conklin; Eddy, Sarah L.; Stearns, Tim; Cyert, Martha S. – CBE - Life Sciences Education, 2015
We present an innovative course-based undergraduate research experience curriculum focused on the characterization of single point mutations in p53, a tumor suppressor gene that is mutated in more than 50% of human cancers. This course is required of all introductory biology students, so all biology majors engage in a research project as part of…
Descriptors: Research Projects, Science Instruction, Science Education, College Science
Groß Ophoff, Jana; Schladitz, Sandra; Leuders, Juliane; Leuders, Timo; Wirtz, Markus A. – Peabody Journal of Education, 2015
The ability to purposefully access, reflect, and use evidence from educational research (Educational Research Literacy) is expected of future professionals in educational practice. Based on the presented conceptual framework, a test instrument was developed to assess the different competency aspects: Information Literacy, Statistical Literacy, and…
Descriptors: Educational Research, Research Methodology, Literacy, Educational Development
Ferjencík, Ján; Slavkovská, Miriam; Kresila, Juraj – Journal of Pedagogy, 2015
The paper reports on the adaptation of a D-KEFS test battery for Slovakia. Drawing on concrete examples, it describes and illustrates the key issues relating to the transfer of test items from one socio-cultural environment to another. The standardisation sample of the population of Slovak pupils in the fourth year of primary school included 250…
Descriptors: Executive Function, Foreign Countries, Test Construction, Test Items
Baghaei, Purya – International Education Studies, 2011
This study addresses the issue of the optimal number of gaps in C-Test passages. An English C-Test battery containing four passages each having 40 blanks was given to 104 undergraduate students of English. The data were entered into SPSS spreadsheet. Out of the complete data with 160 blanks seven additional datasets were constructed. In the first…
Descriptors: Language Tests, English (Second Language), Undergraduate Students, Test Items
Svetina, Dubravka – ProQuest LLC, 2011
The purpose of this study was to investigate the effect of complex structure on dimensionality assessment in compensatory and noncompensatory multidimensional item response models (MIRT) of assessment data using dimensionality assessment procedures based on conditional covariances (i.e., DETECT) and a factor analytical approach (i.e., NOHARM). …
Descriptors: Item Response Theory, Factor Analysis, Correlation, Sample Size
Thompson, Tony – International Journal for Mathematics Teaching and Learning, 2011
This research provides insight into one US state's effort to incorporate higher-order thinking on its Algebra I End-of-Course tests. To facilitate the inclusion of higher-order thinking, the state used "Dimensions of Thinking" (Marzano et al., 1988) and "Bloom's Taxonomy" (Bloom et al., 1956). An analysis of Algebra I test…
Descriptors: Test Items, Familiarity, Classification, Algebra
Black, Beth; Suto, Irenka; Bramley, Tom – Assessment in Education: Principles, Policy & Practice, 2011
In this paper we develop an evidence-based framework for considering many of the factors affecting marker agreement in GCSEs and A levels. A logical analysis of the demands of the marking task suggests a core grouping comprising: (i) question features; (ii) mark scheme features; and (iii) examinee response features. The framework synthesises…
Descriptors: Interrater Reliability, Grading, Scoring, High Stakes Tests
Hohensinn, Christine; Kubinger, Klaus D. – Educational and Psychological Measurement, 2011
In aptitude and achievement tests, different response formats are usually used. A fundamental distinction must be made between the class of multiple-choice formats and the constructed response formats. Previous studies have examined the impact of different response formats applying traditional statistical approaches, but these influences can also…
Descriptors: Item Response Theory, Multiple Choice Tests, Responses, Test Format
Facon, Bruno; Magis, David; Belmont, John M. – Research in Developmental Disabilities: A Multidisciplinary Journal, 2011
The matching of groups is a traditional way to control for confounding variables in developmental disabilities research. The equivalency of means across groups is routinely checked for these variables, but not the homogeneity of their variances or the shapes of their distributions. In the present paper, it is argued that group matching can go…
Descriptors: Test Bias, Developmental Disabilities, Scores, Statistical Analysis
Hohensinn, Christine; Kubinger, Klaus D.; Reif, Manuel; Schleicher, Eva; Khorramdel, Lale – Educational Research and Evaluation, 2011
For large-scale assessments, usually booklet designs administering the same item at different positions within a booklet are used. Therefore, the occurrence of position effects influencing the difficulty of the item is a crucial issue. Not taking learning or fatigue effects into account would result in a bias of estimated item difficulty. The…
Descriptors: Measurement, Test Construction, Test Items, Difficulty Level
Viglione, Donald J.; Perry, William; Giromini, Luciano; Meyer, Gregory J. – International Journal of Testing, 2011
We used multiple regression to calculate a new Ego Impairment Index (EII-3). The aim was to incorporate changes in the component variables and distribution of the number of responses as found in the new Rorschach Performance Assessment System, while sustaining the validity and reliability of previous EIIs. The EII-3 formula was derived from a…
Descriptors: Test Items, Self Concept, Validity, Evaluation
Aryadoust, Vahid; Goh, Christine C. M.; Kim, Lee Ong – Language Assessment Quarterly, 2011
Differential item functioning (DIF) analysis is a way of determining whether test items function differently across subgroups of test takers after controlling for ability level. DIF results are used to evaluate tests' validity arguments. This study uses Rasch measurement to examine the Michigan English Language Assessment Battery listening test…
Descriptors: Test Items, Listening Comprehension Tests, Test Bias, Test Validity
Acar, Tulin – Educational Sciences: Theory and Practice, 2011
The purpose of this study is to examine the number of DIF items detected by HGLM at different sample sizes. Eight different sized data files have been composed. The population of the study is 798307 students who had taken the 2006 OKS Examination. 10727 students of 798307 are chosen by random sampling method as the sample of the study. Turkish,…
Descriptors: Test Bias, Sample Size, Regression (Statistics), Models
Babcock, Ben – Applied Psychological Measurement, 2011
Relatively little research has been conducted with the noncompensatory class of multidimensional item response theory (MIRT) models. A Monte Carlo simulation study was conducted exploring the estimation of a two-parameter noncompensatory item response theory (IRT) model. The estimation method used was a Metropolis-Hastings within Gibbs algorithm…
Descriptors: Item Response Theory, Sampling, Computation, Statistical Analysis
Livingston, Samuel A.; Kim, Sooyeon – ETS Research Report Series, 2010
A series of resampling studies investigated the accuracy of equating by four different methods in a random groups equating design with samples of 400, 200, 100, and 50 test takers taking each form. Six pairs of forms were constructed. Each pair was constructed by assigning items from an existing test taken by 9,000 or more test takers. The…
Descriptors: Equated Scores, Accuracy, Sample Size, Sampling

Peer reviewed
Direct link
