Week 3 Flashcards
1
Q
Reliability
A
- Synonym for dependability and consistency
- Refers to consistency of measurement in psychometrics
- Not necessarily reflecting good or bad results, just consistent ones
- Test may be reliable in one context but not another
2
Q
Reliability Coefficient
A
- Quantifies reliablity ranging from 0 (not reliable) to 1 (perfectly reliable)
3
Q
Measurement Error
A
- In everyday language Error means some kind of preventable mistake
- In science the meaning is broader relating to measurement imperfection which is inevitable
e.g. 25 could be 24.9978 or 25.1232 - Small fluctuations can be rounded but they are almost never trivial
- Noticeable differences are routinely observed - Think building a steel bridge in hot climate
4
Q
True Scores
A
- Can never be observed directly
- Useful fiction that allows us to understand reliability
- At best true scores are guessed by averaging many measurements
5
Q
Repeated Measurements
A
- Repeated measurement can have problems
- Time between testing has an effect
- some states are in constant flux so they might average differently at different times
6
Q
Carryover Effects
A
- An effect of being tested in one condition on participants’ behavior in later conditions
- The practice effect , where participants perform a task better in later conditions because they have had a chance to practice it.
- The fatigue effect happens when repeated measurement causes results to diminish due to fatigue
7
Q
Construct Score
A
- A theoretical variable we beleive exists such as depression agreeableness or reading ability
- Testing for these is flawed so it can never be a True Score
- Long term averages can still produce close to True Score flaws and all
8
Q
True Scores
A
- We can never observe True Scores directly
- The concept helps us understand reliability
- High reliability does not mean high validity
9
Q
Concept of Reliability
A
- True Score is the long term average of many measurements
- No Carryover Effects
- T = A True Score
- X = Measurement is called Observed Score
- E = Measurement Error
- If the observed score is moslty found by the measurement error the test is unreliable
- Better if the true score is found by the True Score
X = T + E
10
Q
variance (o squared)
A
- Useful to describe test score variability
- Standard Deviation Squared
- Can be broken into components
- True Scores are stable and give consistency to tests
Reliability referesd to the proportion of total variance attributed to T
11
Q
Measurement Error
A
Chapter 5 - Page 297
12
Q
Measuring Psychological Constructs
A
- Constructs can’t be observed
- Can be inferred from what we observe
- Observe behaviour
- Observe responses to self report scales
13
Q
Characteristics of a Typical Scale/Sub-Scale
A
- Statements or questions designed to measure a construct behaviour
- Fixed choide responses - consistent across scale
- Responses are correlated because they asses the same thing
- responses averaged to find an overall score
- Some items are reverse coded
- Strong psychometric properties - reliableitly validity & factor structure
- Normative or Standardised data collected from a wide range of people
14
Q
Commercial Scales
A
- Pay per use
- Published by commercial publishers
- Commonly used for clinical or applied purposes like recruitment or diagnosis
- Sometimes used in research
- Expensive
- Detailed normative data
- MMPI, NEO-PI, Beck Depresion Inventory
15
Q
Non-Commercial Scales
A
- Free to use
- Published in books, journals articles or online
- Often used for research purposes
- Unlikely to be published with Normative Data
- MINI-IPIP, Person Environment Fit Scale
- Typically used by research students
16
Q
Uni Dimensional Scales
A
- Measures only one construct
- All items are intercorrelated
- All items averaged to derive overall score
e.g. Relationship Satisfaction Scale, Beck Depression Inventory
17
Q
Multi Dimensional Scales
A
- Measures Multiple scales
- Each construct has a sub-scale and is a variable
- Each sub-scale is intercorrelated
- Each sub-scale is averaged to derive the score
- Adding up scores for multiple constructs is meaningless
- Sub-scales do not correlate
- Separate reliability and validity data calcullated for each sub-scale.
e.g.: MINI-IPIP (Big 5); Person Environment Fit (three fit dimensions).
18
Q
Define Reliability
A
- How consistent are the tools we use to measure a construct
- Does it produce the same results over time
- Unreliable measures cannot be trusted
19
Q
Four types of Reliability
A
- Internal Consistency
- Test-Retest
- Alternate Forms
- Inter Rater