Midterm 2 Flashcards

1
Q

Reliability - Definition

A

We can trust to measure each person in approximately the same way each time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a reliability coefficient + example?

A

A correlation between two sets of scores (statistical evaluation of reliability), f.ex. test-Retest, alternate forms, internal consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the practice effect?

A

In relation to Test-Retest: the person scores quicker and more efficiently the second time (when close to real trial)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the order effect?

A

Changes in test scores might occur, resulting from the order in which the tests were taken (f.ex. social desirability the 2nd time)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is scorer reliability/interscorer agreement?

A

The amount of consistency among scorer’s judgement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is intrascorer reliability?

A

Clinicians being consistent in way he or she assigned scores from test to test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the confidence interval?

A

Range of scores that we feel confident will include the test taker’s true score (95% sure f.ex.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is frequency distribution?

A

Orderly arrangement of a group of numbers (represented by histogram f.ex.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does a test-retest method assume?

A

That a trait stays stable over time (test taker’s skills and qualities), but mood, fatigue, problems can change.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the disadvantages of the test-retest method?

A

People can check answers, practice effects, and it lowers correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the disadvantages of the alternate form method?

A

It is hard to construct, time consuming

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are three examples of internal consistency measures?

A

Split-half method, Kuder Richardson Method KR 20, Cronbach’s Alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are advantages and disadvantages of the Split half methods?

A

ADV: time efficient (one sitting), DISADV: not applicable in heterogenous tests, reliability only based on 50% of test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is KR-20 + DISADV?

A

Measures consistency of objective right/wrong tests. DISADV: does not work when items = unequal difficulty, and does not work for non-objective tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is Cronbachs’ Alpha?

A

Uses all possible half test version s for Likert-type (non-objective tests) to measure reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Validity - Definition

A

Are the inferences I am going to draw from a person’s score on a test appropriate? (does it measure what it’s supposed to measure)

17
Q

What is validity as a process?

A
  1. Built from the start/test development (existing lit)
  2. Evidence based on content (contact experts, ppl in the situation)
  3. Evidence based on internal structure (similar variables relation)
    4.Evidence based on consequences of testing
18
Q

Factor analysis - Definition

A

A statistical method that allows us to show that scale has number of dimensions that you hypothesized (items in related dimensions correlate with each other)

19
Q

What are the level of measurements?

A
  1. Nominal (categories)
  2. Ordinal (rank-order, distances vary)
  3. Interval (rank-order, equal distance, no zero point, always agree sometimes agree f.ex.)
  4. Ratio (rank order, distance same, has zero point, how many children f.ex.)
20
Q

What is a norm group?

A

We can compare converted scores with those from what we call norm group (previously tested)

21
Q

Variance - definition

A

Measures if individual scores tend to be similar or different from mean.

22
Q

What is the most common correlation coefficient & explain

A

Pearson product-moment correlation = Measures linear association between 2 variables measured on interval/ratio sclaes

23
Q

What is reliability according to the classical test theory?

A

The true-score variance divided by total observed score variance.

24
Q

What are parallel forms?

A

Refers to two tests that have certain identical and hard to achieve statistical properties.

25
Q

What is Cohen’s Kappa?

A

For nominal/ordinal data, it is an index of how consistent scorers are (interrater agreement)

26
Q

What is the standard error of measurement?

A

SEM = index of amount of inconsistency or error expected in observed test score (how much it differs from T)

27
Q

What factors influence reliability?

A

Test itself, test administrators, test scoring, test taker.

28
Q

What is the generalisability theory for reliability?

A

Theory of how well and when we can generalize estimation of reliability from on test to another, or same test in different circumstances.

29
Q

What are the 5 sources of evidence for validity?

A

Evidence based on test content, response processes, internal structure, relations with other variables , consequences of testing.

30
Q

How can we demonstrate evidence of validity based on test content DURING test development?

A

By defining the testing universe (body of knowledge), developing test specifications, carefully constructing test questions.

31
Q

How can we demonstrate evidence of validity based on test content AFTER test development?

A

By examining the extent to which experts agree on relevance of test item content (f.ex. calculating content validity ratio, 1.00- -1.00)