Scale Development Flashcards

1
Q

Learning objectives

A

Overview:
- Measurement scales and latent variables
- Reliability
- Validity
- Guidelines for developing a scale
- Using SPSS to develop a scale
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _
Learning objectives:
1. Understand and be able to explain… measurement scales and latent variables.
2. Understand and be able to explain… the difference between the reliability and validity of a scale.
3. Understand and be able to explain… different ways to assess the reliability and validity of a scale.
4. Understand and be able to explain… the different steps to follow to develop a reliable and valid scale.
5. Be able to… conduct, report and interpret reliability and validity analyses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The issue – how to measure a psychological variable using a questionnaire?

A

Simply putting together questions to measure a construct for which there is no pre-existing scale leaves the nagging doubt that the items may not be reliable or validly measuring what we aim to investigate.

We can’t directly measure psychological phenomena such as attitudes, so we used questionnaire items.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Scales and latent variables

A

Y ——–> X1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Reliability

A

A reliable scale is a scale in which variation in scale scores can be attributed to a latent variable that exerts a causal influence over all the items

i.e. if people respond similarly to all the items, the scale is reliably measuring the effect of the latent variable, all the items are being causally influenced in a similar way by the latent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Measures of reliability

List them (3)

A
  1. Split-Half Reliability
  2. Internal Consistency
  3. Test-Retest Reliability
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Measures of reliability

  1. Split-Half Reliability
A
  1. Split-Half Reliability
    e. g. where scores on the first half of the scale are correlated with scores on the second half of the scale.
    e. g. where scores on the odd items are correlated with scores on the even items.

If the scale is reliable, the two halves of the scale should correlate strongly.

Avoids fatigue effects, can see if this is the case easily

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Measures of reliability

  1. Internal consistency
A
  1. Internal consistency

Coefficient Alpha (α) indicates the proportion of the variance in the scale scores that is attributable to the true score (score that would be obtained given no error) - how much variance is a result of the latent variable alone.

Coefficient alpha is based on correlations between each scale item score and the total score.

It provides a more accurate measure of internal reliability than a split-half correlation

Scores - goes from 0->1, closer to 1 the better. DeVellis recommends that an α of .60 or below is unacceptable, .80-.90 is seen as very good. Above .90 means the scale can likely be shortened.

Diadvantages of internal consistency:

  • Coefficient alpha is dependent not only on the magnitude of the correlations among items, but also on the number of items in the scale. Scales with more items have higher alpha coefficients, the effect of unreliable items is diminished.
  • Coefficient alpha is not a measure of dimensionality, nor a test of unidimensionality. A high alpha coefficient cannot necessarily be taken to reflect unidimensionality (might be measuring more than one latent variable, we wouldn’t know)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Measures of reliability

  1. Test-Retest Reliability
A
  1. Test-Retest Reliability

Scores on the test at one point in time are correlated with scores on the test at another point in time.

If the scale has high test-retest reliability, a strong correlation will be found between the two administrations of scale.

***However, low test-retest reliability may reflect temporal instability in the underlying latent variable (e.g. exam stress 4 and 2 weeks before an exam).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Validity

A

A valid scale is a scale which measures the latent variable of interest (its measuring what we want it to)

You might have a reliable scale but not a valid one i.e. it is accurately measuring a construct but not the one we think it is

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Measures of validity

List them (3)

A
  1. Content validity
  2. Construct validity
  3. Predictive validity
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Measures of validity

  1. Content validity
A
  1. Content validity

To what extent does the content of the scale items reflect the latent variable of interest?

e. g. conduct interviews with members of the target population to generate items for the scale
e. g. ask experts to rate the extent to which each item reflects the proposed latent variable of interest

…Do the items look like they’re measuring the right construct?

Increasing face validity –> interviews with target pop or ask experts to rate the items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q
  1. Construct validity
A

Does the scale relate to other constructs in line with theoretical expectations?

e. g. administer the scale to a sample of participants along with another related measure. A strong correlation would indicate construct validity.
e. g. a measure of personal strengths would be expected to correlate with a measure of self-esteem.

We expect the scale to correlate reasonably well with some related constructs e.g. depression with anxiety.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q
  1. Predictive validity
A

Does the scale have an association with some other external criterion?

e. g. correlate scores on the scale with an external criterion/measure.
e. g. a measure of medication beliefs would be expected to correlate with an objective measure of medication adherence.

Association with some other gold standard? Behaviour/ other scale?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Guidelines for Scale Development (DeVellis, 1991)

A

Guidelines for Scale Development (DeVellis, 1991)

  1. Define the latent variable of interest.
  2. Generate an item pool
    - brainstorm, construct interviews, pluck from other measures
  3. Review items for content.
    - reliable? valid? reduce number
  4. Administer items to a development sample.
    - choose best items
  5. Evaluate the items…
    - reverse coding (pos-neg coding, same direction of scores must mean same thing)
    - mean and SD (mean in middle? adequate variation?)
    - skewness and kurtosis (normal distribution?)
  6. Compute coefficient alpha.
    - reliability
  7. Validate the scale.
    - e.g. running other studies etc
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Skewness and Kurtosis

A

Skewness is where the hump of the bell curve is off to either side.

Kurtosis is where the hump of the bell curve is either super thin or really flat and wide/ non-existent

Both represent abnormal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Investigating Skewness and Kurtosis

A

Both skewness and kurtosis can be represented by Z scores (standardised scores)

this is where the skewness/ kurtosis value is divided by the standard error of either measure e.g. Z score for kurtosis would be kurtosis/ S.E of kurtosis

If z > 3.08, p < .001,
which indicates highly significant levels of skewness/kurtosis.

we want skewness and kurtosis z scores to be BELOW 3.08 and to sig to a level of p

17
Q

Example - mindfulness scale, 6 items

A

a) Distribution - Skewness and Kurtosis:

  • Look at descriptives for each item
  • means look ok? one a bit low
  • S.Ds look ok?
  • skewness, take biggest stat and divide by standard error, bigger than 3.08? needs removing, work way down list until they’re ok
  • look at kurtosis
  • could also look at histograms

b) Reliability - Crohbach’s alpha

  • anything over .6 is technically acceptable, want at least over .7
  • look at correlations between items and rest of scale, would the alpha go up if removed? if yes, remove item

c) Validity - Construct validity
- does the measure correlate with with a related theoretical construct? if yes its valid…

Validity - Predictive validity

  • does the measure correlate with an external criterion/ gold standard? if yes its valid…
18
Q

SPSS: Recoding variables

A

Recode things into different variable name - so you don’t forget you’ve recoded it e.g. item1 => item1r

So,

  • change name
  • click old and new values
  • reverse numbers so lowest value now = highest etc for items that need recoding
  • add and continue
19
Q

SPSS: Evaluating items

A
  • Descriptives
  • select items
  • options (mean, SD, min + max, kurt. + skew)
  • display order - variable list
20
Q

SPSS: Coefficient Alpha (Reliability Analysis)

A
  • Analyse scale
  • reliability analysis
  • drag items
  • statistics -> scale items if deleted
21
Q

SPSS: Combining items into scale

A
  • Compute variable
  • add all items / by no. of items
  • gives composite score of new scale

or, go to functions and select mean (good if you have missing data)