Writing And Evaluating Test Items Flashcards

1
Q

6 guidelines for item writing

A

Define clearly what you want to measure
Generate an item pool
Avoid long items
Reading difficulty appropriate for those who will complete scale
Avoid double barreled items that convey two or more ideas
Can setup mixing positively and negatively worded items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

6 item formats

A
Dichotomous
Polychotomous
Likert
Category
Checklist
Q sorts
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Dichotomous format pros and cons

A
Positives
Easy to Mark
Absolute judgement
Simplicity ease of administration quick scoring
Disadvantages
Encourage memorization
No room for Shades of Grey
Can get 50% by chance
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Polychotomous format pros and cons

example: multiple choice

A

More than two alternatives
Positives
Ease of scoring quick to do decrease is the probability of guess ING
Distrators important
Three alternative multiple choice items as valid and reliabe as five alternative multiple choice
if distracors too easy reliability and validity decreases
Guessing possible for a for choice item 25% probability of getting correct by chance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Likert format pros and cons

A

Positives
Allows researchers to determine how much people endorse items
Can use Factor analysis and find groups of items that go together
Familiar and easy to use
Negative
Ordinal level of data shouldn’t use parametric statistics to analyse it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Category format pro and cons

A

10 point rating system can have more or less than 10 points
Negative
People change ratings depending on context
Positive
Clearly labelled options increase reliability and validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Category format pros and cons

A

10 point rating system can have more or less than 10 points
Negative
People change ratings depending on context
Positive
Clearl

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Checklists definition

A

forced choice on lists of items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

q Sort explanation

A
many cards related to criterion
Sort cards into piles
forced choice 
pile one not at all like me
pile ten very like me
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Item Analysis (3 Methods)

A

Item Difficulty
Discriminability
Item Characteristics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Item Difficulty

definition
formula
guideline

A

percentage of people who get item correct converted to a decimal
optimum difficulty = (100% - chance performance level) divide by two.
many tests need a variety of difficulty
most tests: 0.30 - .70

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Item Discriminability definition and methods

A

have people who have done well on item done well in whole test
Extreme Group
Point Biserial correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Extreme Group Method of Item Discriminability

A

compares people who have done well with people who did poorly.
the proportion of each group who got each item correct
difference between proportions in the discrimination index/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Item Discriminability Point Biserial Method

A

Correlation between performance on item and performance on whole test
dichotomous x continuous variable correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Pictures of item characteristcs

A

item characteristic curve

good curve increases as a function of the total test score.
poor curve is flat
A curve that dips at the end has a fault with the alternation designated correct. Good students know it isn’t completely correct.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Item Response theory advantages

A

provides information on item functioning, value of specific items. reliability of scale
test takers scores in defined by the level of difficulty of item they can get correct.
can easily adapt for computer administration
means test takers don’t spend time on items that are too easy or too hard
reduces bias against slow test takers
identifies test takers with unusual response patterns
cope with items in different formats

17
Q

Criterion Referenced test definition

A

compares performance with a clearly defined criterion

18
Q

Constructing a criterion referenced test

A

specify objectives to learn
develop test items
give to students who should meet the objective (done the learning) and one who shouldn’t.
Frequency polygon should be a v.
the bottom of the v is the antimode (least frequent score)
The antimode is the cut score for meeting the objective.

19
Q

Limitations of item Analysis

A

They enable test constructor to separate students but don’t help students learn.
Don’t always identify weaknesses or knowledge gaps.

20
Q

Example of using Q Sort from Carl Rogers evaluating self-concept
method
meaning of large discrepency

A

person recieves set of cards with appropriate statements. person sorts cards from most to least descriptive .
first sort: who they are
second sort: ideal self
large discrepencies reflect poor adjustment and low self-esteem

21
Q

Information given on items in item response theory

A

provides information on item functioning, value of specific items. reliability of scale
Identifies unusual response patterns

22
Q

Steps to test construction

A

Cougers do dudes during Dawn

Conduct review
Describe use and interpretation
Decide who and why
Need for measures of dissimulation

23
Q

Measures of dissimulation

A

Mmpi
Social desirability
Fake bad
Response set

24
Q

Test manual needs:

Normals do outrageous orgasms in london

A
Norms, sampling
Development of construction, items, norms, scoring
Reliability
Validity
Special Out groups 
Parts of test