Week 7: Tests Flashcards

1
Q

What are psychometrics?

A

The branch of psychology concerned with testing and measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is classical test theory?

A

Observed score = true score + error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the goal of testing?

A

To produce useful tests and reduce the capacity for error as much as possible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

List some examples of test construction that could influence error

A
  • poorly worded questionnaires
  • extreme statements
  • confusing words and expression
  • culturally confusing terms
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are some factors that could influence errors in test administration?

A
  • deviation from instructions
  • testing environment
  • test administrator
  • participant derived error
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Examples of participant derived error

A
  • hostile or uncooperative
  • donkey vote
  • misunderstanding of instructions
  • temporary illness or condition that might impact test scores
  • contrast your motivations in collecting data with participant motivations in taking part in research
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the types of validity

A
  • face
  • construct
  • content
  • criterion
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Face validity

A

Does the test meet participants expectations and appear to measure what it is supposed to?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Content validity

A

Does the test assess the whole content area of what it’s meant to be assessing?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

List some useful non-specific bases to cover in generating content

A
  • emotions
  • thoughts
  • behaviours
  • interpersonal relationships
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are four ways of categorising constructs?

A
  • component analysis
  • factor analysis
  • cluster analysis
  • taxometric analysis
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Describe component analysis

A

About taking a bunch of things and reducing them into one thing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Describe factor analysis

A

Finding the underlying constructs of items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Describe cluster analysis

A

Used for grouping participants based on current data or predicted outcomes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the aim of cluster analysis?

A

To create groups that are homogenous as possible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Describe taxometric analysis

A
  • attempts to answer continuum vs category questions
17
Q

What is null hypothesis significance testing?

A

Process of logic examining the probability that a selection of data indicating an effect is drawn from the same sampling distribution as a selection of data not demonstrating an effect (the hypothesis that there is no effect occurring)

18
Q

When do we reject the null?

A

When the probability is small (e.g.

19
Q

What are some short comings of significance?

A
  • higher sample size likely to bias towards lower p value

- lower p value tied to stronger effect, so doesn’t necessarily detect small but consistent effects

20
Q

Discuss how to cope with assumption violations

A
  • run the original tests and qualify this in the discussion
  • modify the distribution using a transformation
  • identify tests that have different assumptions the might be more appropriate (e.g. non parametric)
21
Q

What is the ‘optimal’ solution for assumption violations?

A
  • run the statistics in a raw form and one where you compensate for violated assumptions
  • only report a modified data if the pattern of results substantially differs
22
Q

Why is significance controversial?

A
  • significant data is reported and biased in publication
23
Q

How do computerised tasks typically operate?

A

Examining number of errors (accuracy) or reaction time (speed)

24
Q

List some methodological decisions for testing

A
  • number of items
  • evenness of content distribution
  • format of response scale
  • mid point option in response scale
  • use of reverse coding
  • instructions framing questionnaire
  • time period responses is asked to reflect on
25
Q

What is ipsatisation?

A

When we normalise each participant’s responses so that they reflect their distance from an individual midpoint

26
Q

What is the issue with ipsatisation?

A

It controls for differences in how the response scale is perceived, but can be viewed as removing useful variance, as perhaps the way that the response scale is perceived is an important part of the trait in the first place

27
Q

What are some options to adapt a test to your needs?

A
  • deleting items
  • adding items
  • changing response styles
  • making slight alterations to wording
28
Q

Test-restest reliability

A
  • how are scores expected to change over time?

- need to consider whether change in these scores is legitimate or due to error

29
Q

Inter-rater reliability

A
  • concordance in agreement between raters

- can be affected by a defective scale, or poorly trained raters

30
Q

Split-half reliability

A
  • test is split in two, scores for each half compared

- consistency suggests reliability

31
Q

Parallel-forms reliability

A

Administering diffrent versions of an assessment tool to the same people - both versions must reflect the same construct and content

32
Q

Internal consistency

A
  • cronbach’s alpha

- the degree to which a set of items measures a single, unidimensional construct

33
Q

Convergent construct validity

A
  • is the test outcome correlated with the variables that we could expect?
  • can be a negative or positive correlation
  • can be an identical construct
  • can be different constructs
34
Q

Discriminant construct validity

A
  • does a measure lack correlation with variables that we would not expect it to correlate with?
35
Q

Criterion related validity

A
  • should be an observable, behavioural or group based characteristic
36
Q

Concurrent

A

When you assess it at the same time as the measure you are validating

37
Q

Predictive

A

When you assess the criterion later on

38
Q

Why does psych have a higher error rate than other sciences?

A

Because we look at abstract variables and infer meaning

39
Q

What are some ways to improve computerised testing?

A
  • utilise metrics
  • engage/motivate the participant
  • ask them about the quality when they’re done