Chapter 15- Assessments and Grading Flashcards
standardized tests
tests given under uniform conditions and scored according to uniform procedures, teachers do not have much say in selecting these tests
classroom assessment
selected and created by teachers and can take many different forms- unit tests, essays, portfolios, projects, performances, oral presentations, critical because teaching involves making many kinds of judgements
measurement
an evaluation expressed in quantitative (number) terms, tells how much, how often, or how well by providing scores, ranks, or ratings, compare one students performance on a task to a standard or the performance of other students
assessment
procedures used to obtain information about student performance, broader than testing and measurement, includes all kinds of ways to sample and observe students’ skills, knowledge, and abilities, formal or informal
formative assessment
ungraded testing used before or during instruction to aid in planning and diagnosis, purpose is to guide teacher in planning and improving instruction and to help students improve learning, helps form instruction and provides feedback
pretest
type of formative test for assessing students knowledge, readiness, and abilities, identifies areas of weakness, not graded
summative assessment
testing that follows instruction and assesses achievement, purpose is to inform the teacher and the student about the level of accomplishment attained, provides a summary for accomplishment
norm-referenced testing
testing in which scores are compared with the average performance of others, people who have taken the test provide a backdrop for determining the meaning of an individual’s score
norm group
large sample of students serving as a comparison group for scoring tests, can determine if the test score is above, below, or around the average for the group the person belongs to, four types: class/school, school district, national samples, international samples, formed so that a variety of demographic samples are included, tend to encourage competition and do not say whether a student is ready to move on to the next level
criterion-referenced testing
testing in which scores are compared to a set performance standard, benchmark, or minimum passing score, measures the accomplishments of very specific objectives, results tell exactly what the student can or can’t do, best for teaching basic skills, standard can be arbitrary and based on the teachers experience
reliability
consistency of test results, scores are reliable if a test gives consistent and stable reading of a persons abilities from one occasion to the next, measuring reliability by giving the test on two separate occasions indicates stability (test-retest reliability), if a group of people take two equivalent forms of a test and the scores on both tests are comparable, this has alternate-form reliability, refers to the internal consistency or precision of the test (split-half reliability: compare performance on half of the test questions with performance on the other half)
standard error of measurement
hypothetical estimate of variation in scores if testing were repeated, the more reliable a test is, the less error there will be in the score we observe, a reliable test is defined as one with a small standard error of measurement
confidence interval
range of scores within which an individual’s particular score is likely to fall, calculated using the standard error of measurement and identify a range of scores above the actual test score and below it, width of an interval represents how much a student’s score might vary due to errors of measurement
true score
the score the student would get if the measurement were completely accurate and error-free, confidence interval may not include the true score
validity
degree to which a test measures what it is intended to measure, to have validity, the decisions and inferences based on the test must be supported by evidence, judged in relation to a particular use or purpose, content related evidence: test measures the skills covered in the course so test questions are about the important topics, criterion-related evidence: scores correlate with academic performance in school, construct-related evidence: demonstrated when the results of a test correlate with the results of another well-established, valid measure of that same construct, test must be reliable before it can be valid, and reliability does not guarantee validity
assessment bias
qualities of an assessment instrument that offend or unfairly penalize a group of students because of the students’ gender, SES, race, ethnicity, etc, can arise from many factors such as content, language, or examples that might distort the performance of a group, the questions asked may centre on experiences and facts more familiar to students from the dominant culture than to students from minority groups
culture-fair or culture-free testing
a test without cultural bias, unsuccessful, minority students scored the same or worse, cannot separate culture from cognition: every students learning is embedded in their own culture, and every test question emerges some kind of cultural knowledge
objective testing
multiple choice, matching, true/false, short answer, fill-in tests, scoring answers does not require interpretation, variety in question types lower’s students anxieties because the entire grade does not depend on one type of question that a particular student may find difficult