NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
National Defense Education Act1
What Works Clearinghouse Rating
Showing 1 to 15 of 28 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
B. Goecke; S. Weiss; B. Barbot – Journal of Creative Behavior, 2025
The present paper questions the content validity of the eight creativity-related self-report scales available in PISA 2022's context questionnaire and provides a set of considerations for researchers interested in using these indexes. Specifically, we point out some threats to the content validity of these scales (e.g., "creative thinking…
Descriptors: Creativity, Creativity Tests, Questionnaires, Content Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Jessica B. Koslouski; Sandra M. Chafouleas; Amy Briesch; Jacqueline M. Caemmerer; Brittany Melo – School Mental Health, 2024
We are developing the Equitable Screening to Support Youth (ESSY) Whole Child Screener to address concerns prevalent in existing school-based screenings that impede goals to advance educational equity using universal screeners. Traditional assessment development does not include end users in the early development phases, instead relying on a…
Descriptors: Screening Tests, Psychometrics, Validity, Child Development
Wenyue Ma – ProQuest LLC, 2023
Foreign language placement testing, an important component in university foreign language programs, has received considerable, but not copious, attention over the years in second language (L2) testing research (Norris, 2004), and it has been mostly concentrated on L2 English. In contrast to validation research on L2 English placement testing, the…
Descriptors: Second Language Learning, Chinese, Student Placement, Placement Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Jessica B. Koslouski; Sandra M. Chafouleas; Amy Briesch; Jacqueline M. Caemmerer; Brittany Melo – Grantee Submission, 2024
We are developing the Equitable Screening to Support Youth (ESSY) Whole Child Screener to address concerns prevalent in existing school-based screenings that impede goals to advance educational equity using universal screeners. Traditional assessment development does not include end users in the early development phases, instead relying on a…
Descriptors: Screening Tests, Usability, Decision Making, Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Traynor, Anne – Educational Assessment, 2017
Variation in test performance among examinees from different regions or national jurisdictions is often partially attributed to differences in the degree of content correspondence between local school or training program curricula, and the test of interest. This posited relationship between test-curriculum correspondence, or "alignment,"…
Descriptors: Test Items, Test Construction, Alignment (Education), Curriculum
Peer reviewed Peer reviewed
Direct linkDirect link
Jordan, Sally – Computers & Education, 2012
Students were observed directly, in a usability laboratory, and indirectly, by means of an extensive evaluation of responses, as they attempted interactive computer-marked assessment questions that required free-text responses of up to 20 words and as they amended their responses after receiving feedback. This provided more general insight into…
Descriptors: Learner Engagement, Feedback (Response), Evaluation, Test Interpretation
Peer reviewed Peer reviewed
Direct linkDirect link
Cabrera, Nolan L.; Cabrera, George A. – Educational Horizons, 2011
Just like all the high-stakes tests that determine students' futures nowadays, The Chorizo Test is a standardized test rooted in the culture of the test makers. It was originally created to be used with students in teacher training programs to sensitize them to the pitfalls inherent in standardized pencil-and-paper tests, such as linguistic bias…
Descriptors: Test Use, Standardized Tests, Social Sciences, High Stakes Tests
Peer reviewed Peer reviewed
Crites, John O.; Savickas, Mark L. – Journal of Career Assessment, 1996
The Career Maturity Inventory was revised in 1995 using previously unpublished longitudinal data for item selection. The new inventory has 25 attitude and 25 competence items, each yielding a score that measures degree of career maturity of conative and cognitive variables, respectively. (SK)
Descriptors: Career Development, Measures (Individuals), Scoring, Test Interpretation
Peer reviewed Peer reviewed
Harnisch, Delwyn L. – Journal of Educational Measurement, 1983
The Student-Problem (S-P) methodology is described using an example of 24 students on a test of 44 items. Information based on the students' test score and the modified caution index is put to diagnostic use. A modification of the S-P methodology is applied to domain-referenced testing. (Author/CM)
Descriptors: Academic Achievement, Educational Practices, Item Analysis, Responses
Zimmerman, Irla L.; Woo-Sam, James M. – 1982
Two kinds of WISC-R short forms, item reduction and subtest reduction, are reviewed in terms of their ability to meet these criteria of adequacy: a significant correlation between the full scale IQ and the short form IQ, a non-significant difference between the full and short form mean IQ, a low percentage of IQ classification changes resulting…
Descriptors: Intelligence Tests, Test Interpretation, Test Items, Test Reliability
Peer reviewed Peer reviewed
Millman, Jason – Educational Measurement: Issues and Practice, 1994
The unfulfilled promise of criterion-referenced measurement is that it would permit valid inferences about what a student could and could not do. To come closest to achieving all that criterion-referenced testing originally promised, tests of higher item density, with more items per amount of domain, are required. (SLD)
Descriptors: Criterion Referenced Tests, Educational History, Inferences, Norm Referenced Tests
Peer reviewed Peer reviewed
Ackerman, Terry A. – Applied Measurement in Education, 1994
When item response data do not satisfy the unidimensionality assumption, multidimensional item response theory (MIRT) should be used to model the item-examinee interaction. This article presents and discusses MIRT analyses designed to give better insight into what individual items are measuring. (SLD)
Descriptors: Evaluation Methods, Item Response Theory, Measurement Techniques, Models
Peer reviewed Peer reviewed
Wilson, Sandra Meachan; Hiscox, Michael D. – Educational Measurement: Issues and Practice, 1984
This article presents a model that can be used by local school districts for reanalyzing standardized test results to obtain a more valid assessment of local learning objectives can be used to identify strengths/weaknesses of existing programs as well as individual students. (EGS)
Descriptors: Educational Objectives, Item Analysis, Models, School Districts
Peer reviewed Peer reviewed
Hills, John R. – Educational Measurement: Issues and Practice, 1993
A scenario and accompanying questions and answers are posed to help educators examine possible problems in interpreting a student's test score profile. Profiles developed and used soundly are very helpful, but possible pitfalls in test interpretation must be recognized. (SLD)
Descriptors: Academic Achievement, Educational Assessment, Elementary Secondary Education, Performance
American Coll. Testing Program, Iowa City, IA. – 1995
A study was designed to provide recommendations regarding the use of the achievement levels set in 1992 for reporting National Assessment of Educational Progress (NAEP) reading results in 1994 and in future NAEP reading assessments. Two procedures were used: the Item Difficulty Categorization (IDC) procedure involved an evaluation of the…
Descriptors: Elementary Secondary Education, Grade 12, Grade 4, Grade 8
Previous Page | Next Page ยป
Pages: 1  |  2