Validity Flashcards

1
Q

What is the difference between reliability and validity?

A

Reliability tells us about the consistency of the measurement while validity tells us about the accuracy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What type of validity aims to show that the test appears to be testing what it set out to test?

A

Face validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is predictive and concurrent validity aiming to measure?

A

How well a test estimates scores on a criterion.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the difference between predictive and concurrent validity?

A

Predictive aims to predict how a person will perform on a future criterion, while concurrent aims to predict how they will go on a test performed at the same time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is content validity aiming to measure?

A

How well a test covers a representative sample of the domain being measured.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is meant by construct and how do we observe them?

A

A construct is a hypothetical concept that allows us to make more sense of behaviour. We use operations to observe constructs and tests can be considered operations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

For validity, do we care more about how large the coefficient is or if it is significant?

A

Significance matters more than the size of the coefficient.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What size coefficient do we want for validity tests?

A

.2-.5

.2 is the absolute minimum.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the standard error of estimate telling us

A

The amount of error between our predicted scores and our actual scores. Think of a regression line.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How do we estimate the 95% CI around a predicted score?

A

We use the standard error of estimate.
Y+/-1.96SEest
This is different to an obtained score!!!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the difference between measuring a CI around a predicted score and an observed score?

A

Predicted scores use SEest (because the error is predicted vs obtained)
Obtained scores use SEM (as we want to find out what the true score was)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the difference between SEest and SEM?

A

SEest looks at the error between predicted score and obtained score, while SEM looks at error between obtained score and true score.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

If a prediction criterion is 80 and a person’s score 95% CI ranges from 73-82, would we accept that candidate?

A

Yes, because their true score could be 80 according to the CI.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a cutting point?

A

A cutting point is a score that will separate two groups. People who are above will be accepted, and those who fall below will be rejected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are the factors that affect the validity coefficient?

A

Low sample size
Restrictions of range (selectivity problem)
Non-linear relationship between test scores and criterion
Problems with criterion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does the restriction of range affect validity?

A

restriction of range means we aren’t getting the full range of scores. In essence, we aren’t getting a normative sample. May be testing a group who has already passed the criterion, so upper range restriction. Or self-selection means people won’t complete the test, so skewed with conscientious people.

17
Q

What is the weakest part of criterion-related validity?

A

The criterion!
We can get criterion contamination where the assessor knows the person will do well so they are biased.
OR
Criterion requirements change, so measuring something different to before.

18
Q

Which two types of validity aren’t empirically established?

A

Content and face validity.

19
Q

How we make sure we are achieving high construct validity?

A

Make sure the test is driven by current research and theories.

20
Q

How do we test construct validity?

A

Correlate scores with other constructs or behaviours or groups. Looking for convergent or divergent results.
Age differentiation
Criterion groups (high and low scoring groups)
Relationship to behaviour