2. Test worthiness and Statistics Flashcards

1
Q

What is an important test in creation and measures of worthiness?

A

Correlations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is the range for correlations?

A

-1 - 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what are the ways to observe correlation?

A

Linear - line of best fit

Curvilinear - what is the shape of the line?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does shared variance question?

A

What are the factors that contribute to shared variance?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is another term for shared variance

A

Squared coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is the only way to determine a variance when you have a correlation coefficient?

A

square the variance then you know how much variance there are between two variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what does reliability refer to?

A

Refers to how free the test is from measurement error; are you going to get the same, or a very close score if you sit the test again? It’s about internal consistency, reliability & dependability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what does reliability depend on?

A

construction of test & environment administered in

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

why will there always be error?

A

There’s no perfect test or environment so will always be some error but we want to minimise it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how is reliability usually reported as?

A

correlation coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

why do different types of tests have different levels of reliability?

A

reliability e.g., well constructed achievement tests may have reliability coefficients of .90 but personality tests often much lower (.70) because the concept is abstract & potentially fluctuates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

there are many ways to test reliability. What are some measures of reliability?

A

test-retest, alternative forms , consistency of measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is test-retest reliability?

A

Give test twice to same group, usually a couple of weeks apart. then correlate results

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what does a higher correlation suggest in test-retest reliability?

A

Higher the correlation, more reliable the test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

why does the fluctuations in results of test-retest reliability occur?

A

Results can fluctuate depending on things such as the time between test being taken, forgetting information, might learn more about test contents by studying in the interim, more familiar with test format second time around

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

When is test-retest reliability likely to fluctuate?

A

when testing a stable construct (e.g. IQ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

what is ALTERNATIVE, PARALLEL OR EQUIVALENT FORMS RELIABILITY?

A

Making two or more versions of the same test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

what issues ALTERNATIVE, PARALLEL OR EQUIVALENT FORMS RELIABILITY prevent?

A

`Stops issues like people remembering or studying particular answers between test-retest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

what is the difficulty of ALTERNATIVE, PARALLEL OR EQUIVALENT FORMS RELIABILITY?

A

Hard to make tests equal in terms of content & levels of difficulty, or ensuring administration was exactly the same

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

what must test developers demonstrate in order to have good ALTERNATIVE, PARALLEL OR EQUIVALENT FORMS RELIABILITY?

A

Test developer must demonstrate the versions are truly parallel

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

what is reliability as internal consistency measuring?

A

Measures how test items relate to each other & the test as a whole

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

where does reliability as internal consistency look into?

A

Looks within the test to measure reliability

• E.g., test to measure of anxiety – respondents should answer items that tap aspects of anxiety in a similar way

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

what is them most common form of internal consistency measure?

A

Most common forms of internal consistency (i. e., reliability) are split-half and Cronbach’s alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What is split-half reliability?

A

Use one form of the test administered at the same time. Split the test in two and correlate the scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

what are the issues of the split-half reliability?

A

test may get harder as you go along so first half is not equal to second. May compare odd and even numbered questions but still the halves may not be equal & the test is shorter which can decrease reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

what equation can be used to compensate for shorter tests in split-half reliability?

A

Can use Spearman-Brown equation to compensate for shorter test (rx2/1+r)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

what does CRONBACH’S COEFFICIENT ALPHA AND KUNDER-RICHARDSON attempt to rate?

A

Try to rate internal consistency by estimating reliability of all possible split-half combinations by correlating each item with the total and averaging

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

what is Kuder-Richardson used with?

A

Kuder-Richardson used with forced-choice format tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

what is validity?

A

The extent to which all the available evidence supports that the test is actually measuring what it is intended to measure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

what is validity essential for?

A

It is a central requirement for test without which the test items/tasks would not have meaning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

what are the categories of content validity?

A

face validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

what are the categories of construct validity?

A
  • Criterion-related validity

* Predictive validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

what is content validity testing?

A

The content of a test reflects what the test is aiming to measure. It is sometimes enough for a validity test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

What is construct validity not enough to ascertain test validity for?

A

Not enough to ascertain test validity beyond achievement type tests e.g., for more abstract constructs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

what is face validity?

A

It refers to the look of the test but maybe superficial.
o E. g., the items in the test look as though they ought to measure what you are aiming to measure
Where some tests may look valid and not be or others dont look valid but are

36
Q

what are constructs?

A

Constructs are theoretically driven ways of talking about certain features in the world

37
Q

what does construct validity ask?

A

Construct validity asks how well a test can give a construct meaning
o e.g., anxiety – only exists in so much as the construct represents a set of behaviours, thoughts and feelings

38
Q

what is construct irrelevance?

A

Scores are influenced by something other than what the test is supposed to measure e.g., anxiety or illness impacting exam score.

39
Q

what does construct validity provide?

A

Scientific evidence demonstrating that the construct (model, concept, idea, notion) is actually being measured by the test.

40
Q

when is construct validity most important?

A

Most important when developing tests to measure abstract constructs like depression, anxiety, happiness, love, empathy.

41
Q

what is construct validity measured with?

A

Measured with statistical tools and methods…

42
Q

what does criterion-related validity question?

A

What is the relationship between the criterion (another standard) for the test and the test scores?

43
Q

what does concurrent validity question?

A

It refers to the extent to which the results of a particular test, or measurement, correspond to those of a previously established measurement for the same construct.

44
Q

what does concurrent validity relate to?

A

Relates to what is known at this point in time

45
Q

what is convergent validity?

A

e.g. If you think your test measures post-traumatic growth, you would expect it to be related to other instruments designed to measure positive post-trauma perception (e.g., SRG & Thriving)

46
Q

why dont you want a perfect correlation with other tests when measuring convergent validity?

A

You don’t want a perfect correlation with other tests as that would make yours redundant.

47
Q

how does convergent validity vary?

A

Depending on how close the theoretical connection is between the tests, the coefficient will vary.

48
Q

what does discriminant validity use to determine validity?

A

As with convergent validity, discriminant validity uses other established tests to test construct validity.

49
Q

when using discriminant validity, what are you looking for?

A

This time, you are looking for little or no correlation between your measure (e.g., IES-R & PTGI).

50
Q

what can significantly influence discriminant validity in a clinical context?

A

Knowing what you are dealing with may have significant clinical influence (e.g., Shakespeare-Finch & de Dasell, 2009).

51
Q

when does one turn to predictive validity?

A

If a standard to compare to is not available, interest turns to Predictive validity

52
Q

what does predictive validity test?

A

The relationship between the test scores now and a standard in the future.
o For example, OP scores and first year academic success – depends on age…
o OP for school leavers better predictor than learning strategies which is more predictive for mature age
o Combining evidence often offers more predictive validity e.g., OP. introversion and learning strategies in school leavers.

53
Q

what is cross-cultural fairness>

A

The idea that ethnicity, gender, class, background etc impact on results

54
Q

why were there lots of laws passed in the US about tests being culturally fair?

A

Lots of laws passed in the US about tests being culturally fair so as not to repeat errors of the past which disadvantaged minority groups e.g., employment tests must be able to demonstrate the test is relevant to the job sought

55
Q

what is practicality?

A

Choosing the right test to administer; time, format, cost etc

56
Q

what are the elements of practicality?

A

time, cost, format, readability, ease of administration, scoring and interpretation

57
Q

what is the time aspect of practicality?

A

Time taken to do test needs to reflect person targeted e.g., attention span, age, time available to do the test

58
Q

what is the cost aspect of practicality?

A

Many tests cost & some cost a lot. Balance cost of test with reliability of it and level of need to take it.

59
Q

what is the format aspect of practicality?

A

Types of questions, font size, layout, MC lower anxiety but not always culturally fair (e.g., white males & MC)

60
Q

what is the readability aspect of practicality?

A

Test items be reviewed for readability – our school

61
Q

what is the Ease of administration, scoring & interpretation aspect of practicality?

A

o Understanding the test manuals
o How many people are taking the test & does this impact on ease of administration?
o Level of training needed to administer, score & interpret the results
o How long will scoring take and how long will report take?
o Time needed to explain results to test taker
o Other materials like publisher’s preformatted sheets

62
Q

how does one select and administer a good test?

A

Sourcing tests through articles, books, publisher’s catalogues, test library, online and examining the research about the test such as reliability data and validity

63
Q

what are questions to considered when selecting and administering a good test?

A
  • There are 000s of tests – how to choose?
  • What are the goals of the client or researcher?
  • Which tests can achieve that goal?
64
Q

what do the major test categories include?

A

IQ, aptitude, achievement, behaviour, development, personality, neuropsychological, science, sensory perception, speech, hearing…

65
Q

what are raw scores?

A

Raw scores have not been manipulated in any way

66
Q

what is the meaning of raw scores?

A

They mean nothing without putting them in context so might compare the score an individual gets with the ‘normed score’ (Norm data usually generated from hundreds and hundreds of samples)

67
Q

what are ways to interpret raw scores?

A

o How did the client fair next to others from the same type of group who have taken the test before?
o Compare people from 2 groups e.g., using percentiles
o Compare results for one person on 2 or more tests – sometimes discrepancies between these scores indicate an impairment of sorts

68
Q

what does frequency distribution question?

A

What was the score and how often did it occur?

69
Q

How does one determine frequency distribution?

A

Listing scores in numerical order you can easily see if this person scored higher or lower than most on the distribution. Might list scores or groups of scores. Histograms & frequency polygons etc assist in getting an overview of the data. We can learn a lot about the data from its shape

70
Q

what is leptikurtic

A

tall distribution

71
Q

what is platykurtic

A

flat distribution

72
Q

what is positively skewed distribution

A

tail to the right

73
Q

what is a negatively skewed distribution

A

tail to the left

74
Q

what is commonly used to express standard scores?

A

percentiles

75
Q

what are standard scores?

A

z scores

76
Q

what is the mean of a standard score?

A

0

77
Q

what is the standard deviation of a standard score?

A

1

78
Q

What is the Z score?

A

the score, less the mean, divided by the standard deviation and is therefore sensitive to all components of the variance equations including sample size.

79
Q

when a t scores normally used?

A

personality testing

80
Q

what is the mean in t scores?

A

50

81
Q

what is the sd in t scores?

A

10

82
Q

what is the equation for t scores?

A

T = z(SD)+M

83
Q

what are quantitative approaches?

A
o	Deductive
o	Positivist 
o	Realist
o	Objective
o	Reductionist
o	Generalisation
o	Numbers
84
Q

what are the qualitative approaches?

A
o	Inductive
o	Interpretive
o	Constructivist
o	Subjective
o	Holisitc
o	Uniqueness
o	Words
85
Q

what can qualitative methods be used to support?

A

quantitative measures and inform quantitative studies