Intro to Psych Assessment Flashcards

1
Q

What does the MMPI Measure

A

Mental Health Needs - eg. psychological disorders

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the Mean and SD of the MMPI

A

M = 50, SD = 10 (scores as a t-test)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the most popularly used IQ test?

A

Stanford Binet V

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the Mean and SD of the Stanford Binet Inventory?

A

M = 100, SD = 15

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does the NEO-PIR measure

A

Personality variables (big 5)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How often are many psychological tests updated

A

Approx. every 10 years

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why are tests updated regularly

A

Because cultural and social values change and the tests need to be updated to reflect these

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is meant by Reliability?

A

Consistency of measurement of a test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is Psychological Research?

A

Psych Research aims to make generalisations about a population from a sample of people.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is Psychological Assessment?

A

A special kind of Psych research which seeks to make generalisations about specific individuals with a sample n=1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is test-retest reliability?

A

When scores from the same test administered at different times are highly correlated at diff. time points (ie. r = close to 1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the importance of a CI as it relates to test scores?

A

Because it provides a range of values between which you’d expect the test score to fall.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Why is consistency (reliability) important?

A

Because we want to know that our assessment of needs at one point is similar to our assessment of your needs at a future time . The test should not just refer to your needs in the moment, but generalise to your needs across time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is Psychological Testing?

A

The process of administering one or more psychological test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is validity?

A

Assesses the usefulness of inferences made from test scores.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How are testing and theory interconnected?

A

They are two aspects of the same thing, with good theory we can develop good tests and vice versa.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Psychometric theory is cumulative

A

There has been no serious rebuttal to psychometric theory.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Why is variance in psychological testing a good thing?

A

Because a test with good reliability that produces a wider spread of scores allows us to identify individuals more easily than a test with lower variability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

When is variance in psychological testing a bad thing?

A

When it is caused by error/noise (measurement error) rather than by actual individual differences.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What two types of variance exist in psychological testing?

A
  1. Individual variability (good)

2. Error/random variability aka measurement error (bad)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is a good way of measuring Reliability.

A

The extent to which a test correlates with itself. High correlation = good reliability

Test-Retest reliability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What are the two components of an observed score?

A

X (observed score) = T (true score) + E (error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What are the assumptions of general model of reliability?

A
  1. Mean error of measurement = 0 (because it is random)
  2. True scores and errors are uncorrelated r the = 0
  3. Errors on different measurements are uncorrelated r e1e2 = 0
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What are the formulas for test scores from classical test theory?

A

X=T+E
and
σ^2X = σ^2T + σ^2E (variance version)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

What is the formula for the reliability coefficient?

A

r tt = σ^2T / σ^2X (= true score variance/observed score variance)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

What statistical technique is commonly used to determine reliability?

A

Correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

If r tt = 0.9, what percentage of variance in observed scores is attributed to the true scores?

A

= 90%

Note that this gives the same value as R squared would in a traditional correlation/regression. But test/retest reliability interprets this differently.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

What values can r tt range between?

A

0 and 1 (0 = no reliability, 1 = perfect test/restest reliability)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

How is measurement error determined using r tt?

A

1 - value of r tt = proportion of variance caused by measurement error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

What is the main aim of test developers?

A

Reducing measurement error!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

What is the most common/simple way of measuring reliability?

A

The test-retest reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

What is the reliability of gender questionnaires?

A

r tt = 0.95

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Average reliability for individual items in a test

A

r tt = 0.25 (so having many items in a test important as a means of reducing measurement error).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

What is used to quantify the CI around a persons test score?

A

The standard error of measurement.

Which describes the distribution of scores for an individual if they were to be tested an infinite number of times.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

What is the formula for standard error of measurement?

A

σmeas = σx√(1-rtt)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

What is the SEM (standard error of measurement) when r tt = 0

A

SEM = σx√(1-0) = σx

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

As reliability decreases, variability ________?

A

As reliability decreases, variability increases

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

What is the predicted true score?

A

A score that the CI is based around. It is corrected for the unreliability in the observed true score (so always closer to population mean) .

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

What happens to predicted true score when observed score is below the mean?

A

The predicted true score goes up (towards the population mean)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

What happens to the predicted true score when observed score is above the mean?

A

The predicted true score goes down (towards the mean).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

How does the reliability coefficient (r tt) relate to the predicted true score?

A

The predicted true score will move towards the mean by the value of the error term e.g. if rtt is 0.9, the PTS will move to/away from the mean by 0.1; if the rtt is 0.6, the PTS will move to/away from the mean by 0.4

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

How long is the interval between retesting for test-retest reliability?

A

No more than 6 months. Often over a few weeks or days.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

What are the limitations of test-retest reliability testing?

A

Because re-testing may be influenced by practice effects, and tests that require problem solving will advantage performance on the re-test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

What kinds of tests are best suited to test-retest reliability?

A

Sensory Discrimination and Motor Tests that are not effected by repetition.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

What is test-retest reliability?

A

When same test is given at two different time points and the correlation of the scores determined to see how reliable it is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

What is the reliability coefficient?

A

r tt (the correlation between test scores at two different time points)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

What is Alternate-Form Reliability?

A

When the same person is tested at two time points using two different, but equivalent tests. Reliability coefficient is obtained by correlating score on first and second tests.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

Why is Alternate-form testing often better than test-retest reliability?

A

Because it avoids practice effects and cheating.

49
Q

What are the limitations of Alternate Form Testing?

A
  1. Can still be influenced by practice effects if they are large (reduces but doesn’t eliminate their influence).
  2. Alternate forms are not available for all tests (or can be difficult to produce).
50
Q

What effect will practice effects have on reliability coefficient?

A

r tt will reduce as practice effect increases (but not effected by small practice effects)

51
Q

How does the length of the test effect its’ reliability?

A

The longer a test (more items in it), the more reliable it is.

52
Q

What does the Spearman-Brown formula estimate?

A

The effect of lengthening or shortening a test on its reliability (the reliability coefficient)

53
Q

What is split-half reliability?

A

When the scores from a test are split in half and the two ‘halves’ correlated with each other.

Gets around the problem of not having an alternate-form test, by using two halves of one test in this way.

54
Q

How can tests be divided for split-half reliability?

A
  1. Odd or Even Numbers
  2. First half and second half, but this is not recommended as can be infl. by fatigue effects and inconsistencies with difficulty of questions.
55
Q

What is the estimated reliability coefficient?

A

r nn = number of items in new test/ number of items in old test.

Product of the Spearman-Brown Prophecy Formula.

56
Q

What happens to the product of the Spearman-Brown Prophecy formula as r tt increases?

A

It produces a negatively accelerating growth curve or a “diminishing returns” function.

57
Q

What is internal/interim consistency?

A

The average correlation between each item and all the other items in the test.

58
Q

What are two ways of measuring internal consistency?

A
  1. KR-20 (Kuder-Richardson): dichotomous items ie. yes/no

2. Cronbach’s Alpha (Coefficient Alpha): for items with interval level scores i.e. high/medium/low

59
Q

What is a speed test?

A

One in which individual differences depend on speed of performance. Items are uniform and easy but not enough time provided for anyone to answer all questions.

60
Q

What is a power test?

A

One in which individual differences depend on ability to solve difficult problems. Enough time to complete test given but some items too difficult for anyone to solve. Items start easy and finish hard.

61
Q

What are the 5 methods for testing reliabilty?

A
  1. Test-Retest Reliabilty
  2. Alternate-Form Reliabilty
  3. Split-Half Reliabilty
  4. Kuder-Richardson and Coefficient Alpha
  5. Scorer (Inter-Rater) reliability
62
Q

What is the Standard Error Measurement?

A

The standard deviation for a distribution of individual test scores.

63
Q

What is used to establish a confidence interval around expected scores on a test?

A

The Standard Error of Measurement (SEM) and the Predicted True Score (PTS).

CIs are centred around the PTS, not the observed scores and are therefore asymmetrical to the observed score.

64
Q

How many applications of a test are required when assessing internal consistency (KR20 or Coefficient Alpha)?

A

One

65
Q

What do KR20 and Coefficient Alpha Mesure?

A

Internal Consistency: the correlation between one test item and all of the other items

66
Q

In a test with 20 items, how many correlations is KR20 or Coefficient Alpha made up of?

A

20

67
Q

What is confirmatory factor analysis?

A

Factor analysis that allows a theoretical model to be specified and tested (hypothesis testing)

68
Q

What is exploratory factor analysis?

A

An ‘old’ type of factor analysis that did not allow for null hypothesis testing.

69
Q

What is face validity?

A

What a test appears to measure to the person taking the test

70
Q

What is content validity?

A

how adequately a test samples behaviour representative of the universe of behaviour that the test was designed to sample (eg. exam that covers all of course material)

71
Q

What is criterion related validity?

A

how adequately a test score can be used to infer an individual’s most probable standing on some measure of interest (the criterion)

72
Q

What are two forms of criterion-related validity?

A
  1. Concurrent Validity

2. Predictive Validity

73
Q

WHat is concurrent validity?

A

degree to which a test score is related to some criterion measure obtained at the same time (eg. correlation between test scores and expert rater on someone’s psychological wellbeing)

74
Q

What is predictive validity?

A

an index of the degree to which a test score predicts some criterion measure (eg. do SAT scores predict performance at University?)

75
Q

WHat is a criterion?

A

the standard agains which a test or test score is evaluated

76
Q

Characteristics of a criterion

A
  1. Relevant
  2. Valid
  3. Uncontaminated
77
Q

What is the validity coefficient

A

correlation between test scores and scores on a criterion measure

78
Q

What value should the validity coefficient have?

A

It should not be defined, but should be high enough to result in the identification and

79
Q

What do KR-20 and Cronbach’s Alpha set?

A

An upper limit for the reliability of a test.

80
Q

How do the internal consistency coefficient relate to the test-retest reliability coefficient?

A

The internal consistency coefficient can be much higher than the test-retest reliability coefficient.

81
Q

Is the internal consistency coefficient the same as the test-retest reliability coefficient?

A

No, and they can be different, so just because internal consis. is good, doesn’t mean rtf will be good.

82
Q

What is test standardisation

A

When a test is administered and controlled in the same way each time it is administered.

83
Q

What are the disadvantages of using computerised tests?

A

Many test takers prefer the ‘human’ element in the assessment process.

84
Q

What is convergent validity?

A

Different tests measuring same constructs should correlate highly.

85
Q

What is divergent/discriminant validity?

A

When tests measuring different constructs do NOT correlate

86
Q

What can cause low correlation between two tests?

A

Unreliability in test scores can attenuate (lower/press down) the correlation between the two tests.

87
Q

What is disattenuation?

A

A formula used to remove the influence of the unreliability of test scores on the correlation between two tests.

88
Q

What is the disattenuation formula?

A

r xy / square root of (rxx x ryy)

89
Q

What is the effect of test unreliability on correlation between two tests?

A

The more unreliability, the more the correlation is likely to be attenuated by the unreliability (ie. the correlation will be lower than it should be because of the influence of the unreliability)

90
Q

How can you tell if there is good convergent validity between tests?

A

The disattenuated correlation between the two is high (rxy = close to 1)

91
Q

What is ‘n’ used in the spearman -brown prophecy formula?

A

n = the number of times the test is being lengthened by

n = number of items in ‘new’ version of test/ number of items in ‘old’ test’

eg. when increasing number of test items from 25 to 100 n = 100/25 = 4

92
Q

How many items are usually in a good IQ test?

A

500 - 1000

93
Q

How long would it take to administer a good IQ test

A

Approx 1 hour

94
Q

What is the minimum reliability of a test for individual assessment?

A

0.9

95
Q

What is the minimum reliability for a test of group research?

A

0.7

96
Q

Can you use tests with reliability lower than 0.9?

A

Yes, so long a you know how to interpret the SEM and use this to create CIs (though the CI will be less precise)

97
Q

How does the reliability of a test affect its CI

A

The more reliable, the more precise the CI

98
Q

What is the CHC Model

A

The Horn-Cattel Model of assessing cognitive abilities.

99
Q

How many cognitive abilities are listed at level 2 of the CHC model

A

10

100
Q

80 - 90% of variance in cognitive abilities can be assessed using how many of the CHC categories?

A

5 (out of 10)

101
Q

What are the 5 main categories of cognitive abilities in the CHC?

A
  1. Fluid intelligence (GF)
  2. Crystallised Intelligence (GC)
  3. Short Term Memory (GSm)
  4. Long Term Retrieval (Glr)
  5. Processing Speed (Gs)
102
Q

What are the 5 lesser categories in the CHC Model?

A
  1. Quantitative knowledge
  2. Visual processing
  3. Auditory processing
  4. Correct Decision Speed
  5. Reading/Writing
103
Q

What is fluid intelligence (Gf)?

A

Problem solving abilities (executive function), ability to derive creative solutions from available information.

104
Q

What is Crystallised intelligence (Gc)?

A

Acquired knowledge that grows over time, such as educational/cultural knowledge.

105
Q

What is short-term memory (Gsm)?

A

Working memory, ability to hold information in mind to guide our behaviour over short intervals.

106
Q

What is long term retreival (Glr)?

A

Ability to learn from experience and use info from long term storage. Based on evolutionary old parts of brain such as hippocampus and limbic system

107
Q

What is processing speed (Gs)?

A

The speed at which we can solve problems.

108
Q

How many of the CHC categories are typically used in a Weschler test?

A

4 (does not include long term retrieval - this is an ad on)

109
Q

What is the cut score on a BDI_II test?

A

A score of 20

110
Q

What effect does the prevalence (base rate) have on sensitivity and specificity?

A

In cases of extreme frequency (v. high or low base rates), the sensitivity and specificity of the test can be inaccurate.

111
Q

What is positive predictive power?

A

The probability that someone with a positive test result actually has the condition.

= a/ a + b (row ratio)

Ratio of true positives/all people with a positive test result.

112
Q

WHat is negative predictive power?

A

The probability that someone with a negative test result actually doesn’t have the condition.

= d/ d+c (row ratio)

Ratio of true negatives/ all people with a negative test result

113
Q

What is sensitivity?

A

The proportion of people who have a condition that are correctly identified by the test (column ratio = a/ a +c)

Ratio of true positives/all people with the condition

114
Q

What is specificity?

A

The proportion of people without the target condition for whom the test is also negative (column ratio = d/ d +b)

Ratio of true negatives/all people without the condition.

115
Q

At what level should PPP be before a decision not to test is made?

A

PPP must be greater than 0.5 (50%), if lower, shouldn’t be using a test to diagnose.

116
Q

When prevalence is low, what happens to NPP and PPP?

A

The PPP becomes very low, while the NPP stays ok (so most people who are diagnosed with condition, don’t have it)

117
Q

What are common ways of misinterpreting NPP and PPP?

A

They are not fixed values, but depend on the prevalence in the test setting. Especially problematic when prevalence in a study setting is around 50%, while it would normally be much lower in a professional/population settings (eg. the child abuser example from Cohen & Swerdlick - as base rate decreased, false positives increased and false negatives decreased)

118
Q

For two tests both with same sensitivity and specificity, what happens as base rate decreases?

A

The number of false positives increases.

PPP of the test decreases (so likelihood of having the condition if you’ve been diagnosed is not high)

119
Q

What is the standard error of measurement equal to when rtt = 0?

A

When rtt = 0, the SEM is equal to the standard deviation of the test (ie. 15 if it was an IQ test)