Reliability Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

Reliability definition?

A

the extent to which a measurement of behav is consistent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

internal reliability definition?

A

consistency within a study or measure (tool and/or researcher)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

across researchers def?

A

more than 1 researcher - is there consistency in how they are interpreting behav or carrying out the study

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

inter-rater reliability def ?

A
  • internal reliability
  • 2+ observers watching same behaviour should have same data
  • both can agree on
  • mainly used in observations
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

across measures reliability def?

A
  • MORE THAN 1 MEASURE - if there is more than 1 test (such as used in a repeated measures design) then is there consistency of challenge in each test?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

split-half reliability def ?

A
  • used to test IR
  • answers to test are halved and compared - should be similar
  • questionnaire/ test of behav
  • ppts/ measures ect..
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

External reliability definition def?

A

consistency across time (different occassions)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

across time or different occasions def?

A
  • Behaviour is only measured once - the test may not measure behav consistenly at other times e.g. a personality test may only indicate ppt’s personality traits at the time the study was carried out
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

test - retest reliability def?

A
  • assess ER
  • test scorers compared, should be same
  • retested again (after a while)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Dealing w issues of IR - more than one researcher?

A
More than 1 researcher: 
      - clearly operationalise variable so less room for 
        interpretation 
      - standardise procedures
      - detailed training of researchers
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Dealing w issues of IR - multiple measures?

A

multiple measures:

- standardize measures - make sure they are the same

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Dealing w issues of IR - within a measure?

A

within a measure:

- standardise each question or item

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

method: Lab environment, reliability:?

A

High R - high control of the envir - control over EVs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Method: field experiment, Reliability:?

A

Low R - experiment in real life conditions - low control over EVs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Method: quasi experiment - Natural, R:?

A

Low R - naturally occurring IV -low control over EVs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Questionnaires - Reliability?

A

High R - ppts can repeat questions ro check for consistent results

17
Q

Interview - unstructured: R?

A

Low reliability as each ppt is asked diiff questions - diff to repeat

18
Q

Interview structured: R?

A

High R - each ppt is asked the same questions so it can be repeated to check for consistent results

19
Q

Observation - controlled: R?

A

High R - behav is observed in a controlled envir - control over EVs

20
Q

Observation - naturalistic: R?

A

Low R - behav is observed in real life envir - low control over EVs

21
Q

Case study: R?

A

Low R - unique cases - impossible to repeat in same conditions

22
Q

Content analysis: R?

A

High R - behaviour categories - can be repeated/ checked by other researchers

23
Q

Assessing IR: Inter- rater reliability?

A
  • all of the researchers carry out a pilot on the same ppts, if they are 80% or higher in agreement then they are reliable
  • Pilot the different measures on the same ppts w/o changing the conditions and then compare their scores to ensure consistency
24
Q

Assessing IR: Split half?

A

Measure behaviour using 1st half of the test and then the second half of the test using same ppts, if scores are similar, then the test is consistent throughout

25
Q

Assessing ER?

A
  • Test - retest

- The same ppts repeat the same test at another time e.g. 4 weeks later, to show a consistency over time

26
Q

How to deal with ER?

A

Ensure tests ect. Are designed in relation to measuring behaviours, attitudes ect beyond one setting or time

27
Q

What is assessing reliability?

A

Measures used to check the consistency of a set of results

28
Q

DEFINITION FOR EXAM - inter rater reliability?

A
  • Method of assessing IR
  • Where 2 or more psychologists in the same study produce consistent results by using a standardised procedure , agreed coding system, or correlation of their data.
  • This can be checked for by comparing each psychologists results if there is 80% + agreement then = IR
29
Q

DEFINITION FOR EXAM - split half?

A
  • Method of assessing IR
  • Involves splitting a pp’s test answers in half and seeing whether s/he got the same or similar scores on the 2 halves. If so, IR is high
30
Q

DEFINITION FOR EXAM - test re-test?

A
  • Method of assessing external R
  • Involves testing and retesting the same ppts over time, with the same test, and comparing their scorers. If the scores are the same, the test has ER
31
Q

What are the issues of internal reliability?

A

1- more than 1 researcher
2 - more than 1 measure
3 - within a measure

32
Q

What are the issues of external reliability?

A

Behaviour is only measured once

33
Q

How to deal with issues of IR?

A

1- inter rater reliability

2 - split half

34
Q

How to deal w issues of ER?

A

1- test retest