Psyc Testing and Assessment-Chapter 4 (Reliability) Flashcards

1
Q

What are two sources of Errors in test construction?

A
  • Random Error

- Systematic Error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Random Error?

A

Situational factors:

  • ->Noise(ie. room should be queit and confidential if oral testing is being conducted.
  • -> Lighting
  • ->Fire drill
  • ->sense of urgency (bathroom)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Systematic Error?

A

Measure was inaccurate.
–>for example, the ruler was off by one-tenth of an inch so all the measurements previously taken with that ruler were systematically off.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are two sources of Errors in Testing

A

1) Test Construction

2) Test Administration

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Sources of Errors in Test Construction

A

Domain Sampling Model-small test as an estimate of true ability may not represent the writer’s true knowledge/ability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Sources of Errors in Test Administration (3)

A

1) Test environment-room temperature, lighting, ventilation, pencil or desk surface
2) Test Takers-sleep, effects of drugs or medication, emotional problems, physical discomfort, answered “b” instead of “d”
3) Examiner/Test Administrator-physical appearance of test administrator/examiner (maybe they are intimidating), presence or absence, any clues given

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Sources of Errors in Test Scoring and Interpretation (3)

A

1) growing reliance on objective, computer scorable items (instead of testing critical thinking ability)
2) Intelligence test (lacks creativity aspect)
3) Personality test-open ended questions (hard to interpret)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Methodological Error

A

Interviewers may not have been trained properly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Test-Retest Method

A

Test is given once to a participant, and then they are given the same test at a later date to compare results.

  • Easy correlation b/w scores from T1 and T2
  • MEASURES TRAITS WELL (as traits do not change)
  • DOES NOT measure constantly changing characteristics

Higher the reliability with a correlation the better
-correlation is maximum +1 (for positive correlation) and -1 (for negative correlation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Issues of Test-retest Method (2)

A

1) Carryover effect
2) Practice effects

Interval between testing sessions must be selected and evaluated carefully

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Carryover Effects

A

Participant remembers answers from the first test (issue of test-retest method)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Practice Effects

A

Participants may score better with practice (issue of test-retest method)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Trial Test

A

Some researchers will conduct a trial test to ensure that someone has less anxiety for the test that will be measured.
Trial test ensures that the participant understands the test and what is expected (instructions) to ensure reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Parallel Forms Method

A
  • Two forms of the test (given to everyone)
  • BOTH TESTS have to measure the same level of construct
    - ->can divide the questions into two sets = many items that reflect the same construct

-This can be burdensome as two versions of the test must be created.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Split-Half Method

A

Test is given and divided into halves that are scored separately and results are compared (correlation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Issue of the Split-Half Method

A
  • Test scores gain reliability as the number of items increases;
    • ->Therefore two half tests would be less reliable than the whole test (not the best test unless you have enough items that can be reliably added/divided)
  • The first and second half of the test may not be divided evenly when it comes to difficulty/intensity
17
Q

What is Sampling error (4)

A

1) Item Sampling
2) Time sampling
3) Interater/interobserver
4) Internal consistency

18
Q

Item Sampling:

A

the same construct or attribute may be assessed using a wide pool of items.
–>this sort of error is assessed using alternate forms, or parallel forms reliability; the correlation between two forms of a test is created by randomly sampling a large pool of items believed to assess a particular construct

19
Q

Time sampling:

A

behavior is only watched during a certain time-may not be accurate display
–>Same test given at different points in time may produce different scores, even if given to the same test takers.

20
Q

Interater/interobserver

A

Agreement Reliability-do these observations match?

  • ->80-85% is agreement is best to ensure reliability
  • ->Kappa statistic can adjust the index of agreement
21
Q

Reliability

A

refers to several methods that are used to assess different sources of error.
–>All potential sources of error should be taken into consideration when evaluating reliability

22
Q

Internal Consistency

A

Refers to the intercorrelations among items within the same test.
–>internal consistency is evaluated using split-half reliability, the KR 20 method, or coefficient alpha.

23
Q

How reliable is reliable (measure)?

A
  • You want to see a score of .70%-80% (or higher) for most purposes.
  • Reiko suggests that 80-85% is best
  • Cronbach’s alpha on SPSS will let us know which items should be removed to make the other items/test more reliable.
  • Clinical setting should have 95%
24
Q

How to correct low reliability?

A
  • Increase number of items
  • Factor and item analysis (helps to determine if some of those items belong to the category they are in or if they belong elsewhere)
    • ->items in the wrong category may be affecting your reliability.
    • ->Valid = Reliable
    • ->Reliable does not equal Valid
25
Q

Kudar Richardson 20

A

Forumla for calculating the reliability of a test in which the items are dichotomous (scored usually as 0 or 1 for right or wrong)

26
Q

Cronbach’s Coefficient Alpha

A
  • can confirm that a test has substantial reliability; cannot determine if a test is unreliable
  • marks only the lower bound for reliability (actual reliability may be high)
  • used when two halves of a test have UNEQUAL VARIANCES
  • high alpha: can assume reliability of the test is acceptable b/c the lowest boundary of reliability is still high; reliability will not drop below alpha