Final Flashcards

1
Q

item-level analysis

A

Selected-response items
Constructed-response items
Performance tasks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

test-level analysis

A

Scores from tests and quizzes
Homework marks
Performance assessment marks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

item difficulty

A

how many students are able to answer the question correctly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

easy items

A

p higher than 0.9

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

p value (selected response)

A

proportion of items that are correct. Number of correct answers/number of students taking the test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

difficult items

A

p lower than 0.5

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

moderate items

A

p between 0.5 and 0.9

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

p value (constructed response)

A

sum of marks on the item/ (number of students) x (Max possible mark on the item)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

high discrimination

A

over 0.2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

item discrimination

A

Whether an item can distinguish low achieving students and high achieving students

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Discrimination equation

A

(number of correct answers/upper group) - (Number of correct answers/lower group)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

distractor functioning

A

distractors are incorrect response options.
Developing high-quality (e.g., plausible) distractors can be a challenging task

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

low discrimination

A

under 0.2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

reliability

A

indicates whether an assessment can give consistent
results for students

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

summative reliability

A

between 0.8 and 0.9

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

marks

A

scores on individual summative assessments

12
Q

grades

A

aggregates of marks, a score, or letter that is representative
of a students work over a period of time

13
Q

numeric grading

A

Easier record keeping, more precise measurement

14
Q

letter grading

A

Interpretable right off the box, Mostly standardized. Comparable.

15
Q

grading systems

A

typically consists of the following steps:
1. Analysis of test items and scores
2. Assigning a grade
3. Reporting student performance

16
Q

absolute grading

A

Also known as criterion-referenced
grading
Focus:
Each student’s performance
vs.
Some specified standard

17
Q

relative grading

A

Also known as norm-referenced
grading
Focus:
Each student’s performance
vs.
Performance of his/her peers

18
Q

report cards

A

should provide straightforward
information about what students:
* know
* can demonstrate relative to the graded curriculum
* need to do to achieve/maintain mastery

19
Q

SLA’s and PATs

A
  • determine if students are learning what they are expected to learn
  • provide information to Albertans regarding how well students have achieved
    provincial standards at given points in their schooling
  • assist schools, school authorities, and the province in monitoring and improving
    student learning.
20
Diplomas
* certify the level of individual student achievement in selected Grade 12 courses * ensure that province-wide standards of achievement are maintained * report individual and group results.
21
SLA's
Grade 3. improve student learning, help parents and teacher identify areas of growth
22
PATs
Grades 6 and 9. * Informing students, parents, principals, and public about student achievement levels * Helping schools and authorities in monitoring and improving student learning
23
percent correct scores
the score you get on exam based on number of questions you got right
24
scaled scores
E.g., Your scaled score is 345 out of 500 and your achievement level is “Proficient”.
25
percentile ranks
E.g., You are at the 80th percentile (i.e., you passed 80% of the examinees)
26
standardization
1.requires all students to answer the same questions, or a selection of questions from a common bank of questions 2.requires all the students to take the test in the same way 3.is scored in a “standard” or consistent manner
27
teacher bias
Differences in grading practices among teachers can result in variations in SAMs, leading to discrepancies compared to standardized DIPs.
28
assessment format
Variation in the format of assessments (e.g., essays, multiple-choice questions, practical exams) between SAMs and DIPs may affect comparability due to differences in scoring criteria and emphasis on different skills.
29
curriculum alignment
Misalignment between classroom instruction and the content assessed in diploma exams can lead to differences in student performance and, consequently, discrepancies between SAMs and DIPs.
30
student motivation
Variations in student motivation levels during classroom assessments versus high-stakes diploma exams can impact performance and the comparability of grades.
31
assessment weighting
Differences in the weighting assigned to various assessment components (e.g., classwork, homework, projects) in SAMs versus DIPs can affect overall grades and comparability.
32
computer based testing
delivers a digital version of the test via computers (or a similar tool, such as tablets).
33
computer adaptive testing
After every item, the program calculates student performance. * Items are continuously presented until the maximum number of items/time is met.
34
electronic scoring of written responses
A scoring program is trained by “reading” a large representative sample of student responses to a common question. ➢The program learns what characteristics of a student response predict the score provided by the human raters.