Exam Revision Flashcards

1
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

There are constraints of getting valuable info for hypothesis from study’s design such as - (2)

A

duration of the study
how many people you can recruit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a sample?

A

A sample is the specific group that you will collect data from.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a population?

A

A population is the entire group that you want to draw conclusions about.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Example of population vs sample (2)

A

Population : Advertisements for IT jobs in the UK
Sample: The top 50 search results for advertisements for IT jobs in the UK on 1 May 2020

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is inferential statistics?

A

Inferential statistics allow you to test a hypothesis or assess whether your data is generalisable to the broader population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why is there a focus to do parametric tests than others in research? - (3)

A
  • they are more rigorous, powerful and sensitive than non-parametric tests to answer your question
  • This means that they have a higher chance of detecting a true effect or difference if it exists.
  • They also allow you to make generalizations and predictions about the population based on the sample data.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

We can obtain multiple outcomes from the

A

same people

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

We can obtain outcomes under

A

different conditions, groups or both

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the 4 types of outcomes we measure? (4)

A
  1. Ratio
  2. Interval
  3. Ordinal
  4. Nominal
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a continous variables? - (2)

A

: there is an infinite number of possible values these variables can take on-

entities get a distinct score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

2 examples of continous variables (2)

A
  • Interval
  • Ratio
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is an interval variable?

A

: Equal intervals on the variable represent equal differences in the property being measured

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Examples of interval variables - (2)

A

e.g. the difference between 600ms and 800ms is equivalent to the difference between 1300ms and 1500ms. (reaction time)

temperature (Farenheit), temperature (Celcius), pH, SAT score (200-800), credit score (300-850)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is ratio variable?

A

The same as an interval variable and also has a clear definition of 0.0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Examples of ratio variable - (3)

A

E.g. Participant height or weight
(can have 0 height or weight)

temp in Kelvin (0.0 Kelvin really does mean “no heat”)

dose amount, reaction rate, flow rate, concentration,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is a categorical variable? (2)

A

A variable that cannot take on all values within the limits of the variable

    • entities are divided into distinct categories
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are 2 examples of categorical variables? (2)

A
  • Nominal
  • Ordinal
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is nominal variable? - (2)

A

a variable with categories that do not have a natural order or ranking

Has two or more categories

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Examples of nominal variable - (2)

A

genotype, blood type, zip code, gender, race, eye color, political party

e.g. whether someone is an omnivore, vegetarian, vegan, or fruitarian.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is ordinal variables?

A

categories have a logical, incremental order

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Examples of ordinal variables - (3)

A

e.g. whether people got a fail, a pass, a merit or a distinction in their exam

socio economic status (“low income”,”middle income”,”high income”),

satisfaction rating [Likert Scale] (“extremely dislike”, “dislike”, “neutral”, “like”, “extremely like”).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Using the term ‘variables’ for continous and categorical variables as - (2)

A

both outcome and predictor are variables

We will see later on that not only the type of outcome but also type of predictor influences our choice of stats test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Likert scale is ordinal variable but sometimes outcomes measured on likert scale are treated as - (3)

A

continuous after inspection of the distribution of the data and may argue the divisons on scale are equal

(i.e., treated as interval if distribution is normal)

gives greater sensitivity in parametric tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

What is measurement error?

A

The discrepancy between the actual value we’re trying to measure, and the number we use to represent that value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

In reducing measurement error in outcomes, the

A

values have to have the same meaning over time and across situations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Validity means that the (2)

A

instrument measures what it set out to measure

refers to the accuracy of a measure (whether the results really do represent what they are supposed to measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Reliability means the

A

ability of the measure to produce the same results under the same conditions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

Test-retest reliability is the ability of a

A

measure to produce consistent results when the same entities are tested at two different points in time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

3 types of variation (3)

A
  • Systematic variation
  • Unsystematic variation
  • Rnadomisation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

What is systematic variation - (2)

A

Differences in performance created by a specific experimental manipulation.

This is what we want

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

What is Unsystematic variation (3)

A

Differences in performance created by unknown factors
.
Age, Gender,** IQ, **Time of day, Measurement error etc.

These differences can be controleld of course (e.g., inclusion/exclusion of pps setting age range of 18-25)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Randomisation (other approaches) minimises - (2)

A

effects of unsystematic variation

does not remove unsystematic variation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

What is the independent variable (Factors)? ( 3)

A
  • The hypothesised cause
  • A predictor variable
  • A manipulated variable (in experiments)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

What is depenedent variable? (measures)- (3)

A
  • The proposed effect , change in DV
  • An outcome variable
  • Measured not manipulated (in experiments)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

In all experiments we have two hypotheses which is (2)

A
  • Null hypothesis
  • Alternative hypothesis
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

What is null hypothesis?

A

that there is no effect of the predictor variable on the outcome variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

What is alternative hypothesis?

A

is that this is an effect of the predictor variable on the outcome variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Null Hypothesis Signifiance Testing computes the probability the (2)

A

the probability of the null hypothesis being true (referred as p-value) by computing a statistic and how likely it is that the statistic has that value by chance alone

Referred as p-value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

The NHST does not compute the probability of the

A

null hypothesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

There can be directional and non-directional hypothesis of

A

an alternate hypothesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

non-directional alternate hypothesis is..

A

The alternative hypothesis is that this is an effect of the group on the outcome variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

Directional alternate hypothesis is…

A

The alternative hypothesis is that this the mean of the outcome variable for group 1 is larger than the mean of group 2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

Example of directional alternate hypothesis

A

There would be far greater engagment in stats lecture if they were held at 4 PM and not 9AM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

For a non-directional hypothesis you will need to divide your alpha value at

A

two ends of the tail of normal distirbution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

The 3 misconceptions of Null Hypothesis Signifiance Testing (NHST) - (3)

A
  1. A significant result means the effect is important
  2. A non-significant result means the null hypothesis is true
  3. A significant result means the null hypothesis is false (just give probability that data occured given null hypothesis, doesn’t say huge evidence that null hypothesis is categorically false)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

P-Hacking and HARKING is another issue with

A

NHST

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

p-Hacking and HARKINGS are the - (2)

A

researchers degrees of freedom

cchange after results are in and some analysis has been done

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

P-hacking refers to a

A

selective reporting of significant results

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

Harking is

A

Hypothesising After the Results are Known

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

P-hacking and HARKING are often used in

A

combination

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

What does EMBERS stand for? (5)

A
  1. Effect Sizes
  2. Meta-analysis
  3. Bayesian Estimation
  4. Registration
  5. Sense
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

EMBERS can reduce issues of

A

NHST

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

Uses of Effect sizes and Types of Effect Size (3)

A
  • There a quite a few measures of effect size
  • Get used to using them and understanding how studies can be compared on the basis of effect size
  • A brief example: Cohen’s d
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

Meaning of Effect Size (2)

A

Effect size is a quantitative measure of the magnitude of the experimental effect.

The larger the effect size the stronger the relationship between two variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

Formula of Cohen’s d

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

What is meta-analysis?

A

Meta-analysis is a study design used to systematically assess previous research studies to derive conclusions about that body of research

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

Meta-analysis brings together.. and assesses (2)

A
  • Bringing together multiple studies to get a more realistic idea of the effect
  • Can assess effect siz that are averaged across studies
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

Funnel plots in meta-analysis can be made to….. values stuides…. (2)

A

investigating publication bias and other bias in meta-analysis

values studies by their sample size and observe bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

Bayesian approaches capture

A

probabilities of the data given the hypothesis and null hypothesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

Bayes factor is now often computed and stated alongside

A

conventional NHST analysis (and effect sizes)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

Registration is where (5)

A
  • Telling people what you are doing before you do it
  • Tell people how you intend to analyze the data
  • Largely limits researcher degrees of freedom (HARKING p-hacking)
  • A peer reviewed registered study can be published whatever the outcome
  • The scientific record is therefore less biased to positive findings
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

Sense is where (4)

A
  • Knowing what you have done in the context of NHST
  • Knowing misconceptions of NHST
  • Understanding the outcomes
  • Adopting measures to reduce researcher degrees of freedom (like preregistration etc..)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

most of the statistical tests in this book rely on
having data measured

A

at interval level

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

To say that data are interval, we must be certain that
equal intervals on the scale represent

A

equal differences in the property being measured.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

The distinction between continous and discrete variables can often be blurred - 2 examples- (2)

A

continuous variables can be measured in discrete terms; we measure age we rarely use nanoseconds but use years (or possibly years and months).
In doing so we turn a continuous variable into a discrete
one

treat discrete variables as if they were continuous, e.g., the number of boyfriends/girlfriends
that you have had is a discrete variable. However, you might read a magazine that says ‘the average number of boyfriends that women in their 20s have has increased from 4.6 to 8.9’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

a device for measuring sperm motility that actually measures sperm count is not

A

valid

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

Criterion validity is whether the

A

instrument is measuring what it claims to measure (does
your lecturers’ helpfulness rating scale actually measure lecturers’ helpfulness?).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

The two sources of variation that is always present in independent and repeated measures design is

A

unsystematic variation and systematic variation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

effect of our experimental manipulation
is likely to be more apparent in a repeated-measures design than in a

A

between-group design,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

effect of experimental manipulation is more apparent in repeated-design than independent since in independent design,

A

differences between the characteristics of the people allocated to each of the groups is likely to create considerable random variation both within
each condition and between them

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
72
Q

This means that, other things being equal, repeated-measures designs have
more power to

A

, repeated-measures designs have
more power to d

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
73
Q

We can use randomization in two different ways depending on
whether we have an

A

independent and repeated design measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
74
Q

Two sources of systematic variation in repeated design measure - (2)

A
  • Practice effects
  • Boredom effects
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
75
Q

What is practice effects?

A

Participants may perform differently in the second condition because
of familiarity with the experimental situation and/or the measures being used.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
76
Q

What is boredom effects?

A

: Participants may perform differently in the second condition because
they are tired or bored from having completed the first condition.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
77
Q

We can ensure no systematic variation between conditions in repeated measure is produced by practice and boredom effects by

A

counterbalancing the order in which a person participates in a condition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
78
Q

Example of counterbalancing

A

we randomly determine whether a participant
completes condition 1 before condition 2, or condition 2 before condition 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
79
Q

*

What distribution is needed for parametric tests?

A

A normal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
80
Q

The normal distribution curve is also referred as the

A

bell curve

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
81
Q

Normal distribution is symmetrical meaning

A

This means that the distribution curve can be divided in the middle to produce two equal halves

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
82
Q

The bell curve can be described using two parameters called (2)

A
  1. Mean (central tendency)
  2. Standard deviation (dispersion)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
83
Q

μ is

A

mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
84
Q

σ is

A

standard deviation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
85
Q

Diagram shows:

A

e.g., If we move 1σ to the right then it contains 34.1% of the valeues

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
86
Q

Many statistical tests (parametric) cannot be used if the data are not

A

normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
87
Q

The mean is the sum of

A

scores divided by the number of scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
88
Q

Mean is a good measure of

A

central tendency for roughly symmetric distributions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
89
Q

The mean can be a misleading measure of central tendency in skewed distributions as

A

it can be greatly influenced by scores in tail e.g., extreme values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
90
Q

Aside from the mean, what are the 2 other measured of central tendency? - (2)

A
  1. Median
  2. Mode
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
91
Q

The median is where (2)

A

the middle score when scores are ordered.

the middle of a distribution: half the scores are above the median and half are below the median.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
92
Q

The median is relatively unaffected by … and can be used with… (2)

A
  • extreme scores or skewed distribution
  • can be used with ordinal, interval and ratio data.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
93
Q

The mode is the most

A

frequently occurring score in a distribution, a score that actually occurred

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
94
Q

The mode is the only measure of central tendency that can be used with

A

with nominal data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
95
Q

The mode is greatly subject to

A

sample fluctuations and is therefore not recommended to be used as the only measure of central tendency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
96
Q

Many distributions have more than one

A

mode

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
97
Q

The mean, median and mode are identical in

A

symmetric distribtions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
98
Q

For positive skewed distribution, the

A

mean is greater than the median, which is greater than the mode

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
99
Q

For negative skewed distribution

A

usually the mode is greater than the median, which is greater than the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
100
Q

Kurtosis in greek means

A

bulge or bend in greek

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
101
Q

What is central tendency?

A

the tendency for the values of a random variable to cluster round its mean, mode, or median.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
102
Q

Diagram of normal kurotsis, positive excess kurotsis (leptokurtic) and negative excess kurotsis (platykurtic)

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
103
Q

What does lepto mean?

A

prefix meaning thin

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
104
Q

What is platy

A

a prefix meaning flat or wide (think Plateau)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
105
Q

Tests of normality (2)

A

Kolmogorov-Smirnov test
Shapiro-Wilks test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
106
Q

Tests of normality is dependent on

A

sample size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
107
Q

If you got a massive sample size then you will find these normality tests often come out as …. even when your data visually can look - (2)

A

significant
normally disttibuted

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
108
Q

If you got a small sample size, then the normality tests may look non-siginificant, even when data is normally distributed, due to

A

lack of power in the test to detect a significant effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
109
Q

There is no hard or fast rule for

A

determining whether data is normally distributed or not

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
110
Q

Plot your data because this helps inform on what decisions you want to make with respect to

A

normality

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
111
Q

Even if normality test is sig and data looks visually normally distributed then still do

A

parametric tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
112
Q

A frequency distribution or a histogram is a plot of how many times

A

each score occurs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
113
Q

2 main ways a distribution can deviate from the normal - (2)

A
  1. Lack of symmetry (called skew)
  2. Pointyness (called kurotsis)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
114
Q

In a normal distribution the values of skew and kurtosis are 0 meaning…

A

tails of the distribution are as they should be

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
115
Q

Is age nominal or continous?

A

Continous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
116
Q

Is gender continous or nominal?

A

Nominal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
117
Q

Is height continous or nominal?

A

Continous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
118
Q

Which of the following best describes a confounding variable?

A. A variable that affects the outcome beingmeasured as well as, or instead of, theindependent variable

B. A variable that is manipulated by theexperimenter

C. A variable that has been measured using an unreliable scale

D.A variable that is made up only of categories

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
119
Q

If a test is valid , what does it mean?

A.The test measures what it claims to measure.

B. The test will give consistent results. (Reliability)

C.The test has internal consistency (measure for correlations between different items on same test = see if it measures same construct)

D. Test measures a useful construct or variable = test can measure something useful but not valid

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
120
Q

A variable that measures the effect that manipulating another variable has is known as:

A. DV

B. A confounding variable

C. Predictor variable

D. IV

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
121
Q

The discrepancy between the numbers used to represent something that we are trying tomeasure and the actual value of what we are measuring is called:

A. Measurement error

B. Reliability

C. The ‘fit’ of the model

D. Variance

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
122
Q

A frequency distribution in which low scores are most frequent (i.e. bars on the graph arehighest on the left hand side) is said to be:

A. Positively skewed

B. Leptokurtic = distribution with positive kurotsis

C. Platykurtic = negative kruotsis

D. Negatively skewed = frequent scores

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
123
Q

Which of the following is designed to compensate for practice effects?

A. Counterbalancing

B. Repeated measures design = practice effects issue in repeated measures

C. Giving a participants a break between tasks = this compenstates for bordeom effects

D. A control condition = provides reference point

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
124
Q

Variation due to variables that has not been measured is

A. Unsystematic variation

B. Homogenous variance = assumption variance each population is equal

C. Systematic variation = due to exp manpulation

D. Residual variance = confirms how well regression line constructed fit to actual data

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
125
Q

Purpose of control condition is to

A. Allow inferences about cause

B. Control for participants’ characteristics = randomisation

C. Show up relationship between predictor variables

D. Rule out tertium quid

A

A Allow inferences of cause

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
126
Q

If the scores on a test have a mean of 26 and a standard deviation of 4, what is the z-score for a score of 18?

A. -2

B. 11

C. 2

D. -1.41

A

A (18-26) = -8/4 = -2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
127
Q

The standard deviation is the square root of the

A. Variance

B. Coefficient of determination = r squared

C. Sum of squares = sum of squared deviances

D. Range = largest = smallest

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
128
Q

Complete the following sentence:A large standard deviation (relative to the value of the mean itself

A. Indicate data points are distant from man (i.e., poor fit of data)

B. Indicate the data points are close to mean

C. Indicate that mean is good fit of data

D. Indicate that you should analyse data with parameteric

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
129
Q

The probability is p = 0.80 that a patient with a certain disease will besuccessfully treated with a new medical treatment. Suppose that thetreatment is used on 40 patients. What is the “expected value” of thenumber of patients who are successfully treated?

A. 32

B. 20

C. 8

D. 40

A

A = 80% of 40 is 32 (0.80 * 40)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
130
Q

Imagine a test for a certain disease.

Suppose the probability of a positive test result is .95if someone has the disease, but the probability is only .08 that someone has the disease if his or her test result was positive.

A patient receives a positive test, and the doctor tellshim that he is very likely to have the disease. The doctor’s response is:

A. confusion of intervse

B. Law of small numbers

C. Gambler’s fallacy

D. Correct, because test is 95% accurate when someone has the disease = incorrect as doctor based assumption on incorrect inverse proability

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
131
Q

Which of these variables would be considered not to have met the assumptions ofparametric tests based on the normal distribution?

(Hint many statistical tests rely on data measured on interval level)

A. gender

B. Reaction time

C. Temp

D. Heart rate

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
132
Q

The test statistics we use to assess a linear model are usually _______ based on thenormal distribution

(Hint: These tests are used when all of the assumptions of a normal distribution havebeen met

A. Parametric

B. Non-parametric

C. Robust

D. Not

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
133
Q

Which of the following is not an assumption of the general linear model?

A. Dependence

B. Addictivity

C. Linearity

D. Normally distributed residuals

A

A = independence is an assumption of parametric and not dependence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
134
Q

Looking at the table below, which of the following statements is the most accurate?

Hint: The further the values of skewness and kurtosis are from zero, the more likely it is that thedata are not normally distributed

A. For the number of hours spent practicsing , there is not an issue of kruotsis

B. For level of msucial skill, data are heavily negatively skewed

C. For number of hours spent practicsing there is an issue of kruotsis

D. For the number of hours spent practicsing, the data is fairly positively skewed

A

A - correct

B. Incorrect as value of skewnessis –0.079, which suggests that the dataare only very slightly negatively skewedbecause the value is close to zero

C. Incorrect as value of kurtosis is0.098, which is fairly close to zero,suggesting that kurtosis was not aproblem for these data

D. Incorrect as value of skewnessfor the number of hours spent practisingis –0.322, suggesting that the data areonly slightly negatively skewed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
135
Q

Diagram of skewness

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
136
Q

In SPSS, output if value of skewness is between -1 and 1 then

A

all good

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
137
Q

In SPSS, output if value is below -1 or above 1 then

A

data is skewed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
138
Q

In SPSS, output if value of skewness is below -1 then

A

negatively skewed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
139
Q

In SPSS, output if value of skewness is above 1 then

A

positively skewed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
140
Q

Diagram of lepto kurotic, platykurtic and mesokurtic( normal)

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
141
Q

What does kurotsis tell you?

A

how much our data lies around the ends/tails of our histogram which helps us to identify when outliers may be present in the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
142
Q

A distribution with positive kurtosis, so much of the data is in the tails, will be

A

pointy or leptokurtic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
143
Q

A distribution with negative kurtosis, so the data lies more in the middle, will be more

A

be more sloped or platykurtic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
144
Q

Kurtosis is the sharpness of the

A

peak of a frequency-distribution curve

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
145
Q

If our Kurtosis value is 0, then the result is a

A

normal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
146
Q

If kurotsis value in SPSS between -2 and 2 then

A

all good! = normal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
147
Q

If kurotsis value in SPSS less than -2 then

A

platykurtic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
148
Q

If kurotsis value is greater than 2 in SPSS then

A

leptokurtic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
149
Q

Are we good for skewness and kurotsis in this output SPSS?

A

Good because both the skewness is between -1 and 1 and kurtosis values are between -2 and 2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
150
Q

Are we good for skewness and kurotsis in this output SPSS?A

A

Bad because although the skewness is between 1 and -1, we have a problem with kurtosis with a value of 2.68 which is larger than 2 and -2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
151
Q

Correlational research doesn’t allow to rule out the presence of a

A

third variable = confounding variable

e.g, we find that drownings and ice cream sales are correlated, we conclude that ice cream sales cause drowning. Are we correct? Maybe due to the weather

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
152
Q

The tertium quid is a variable that you may not have considered that could be

A

influencing your results e.g., ice cream and drowning session

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
153
Q

How to rule out tertium quid? - (2)

A

Use of RCTs.

Randomized Controlled Trials allow to even out the confounding variables between the groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
154
Q

Correlation does not mean

A

causation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
155
Q

To infer causation,

A

we need to actively manipulate the variable we are interested in, and control against a group (condition) where this variable was not manipulated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
156
Q

Correlation does not mean causation as according to Andy

A

causality between two variables cannot be assumed because there may be other measured or unmeasured variables affecting the results”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
157
Q

Aside from checking of kurotsis and skewness assumptions in data also check if it has

A

linearity or less commonly additivity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
158
Q

Additivity refrs to the combined

A

effect of many predictors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
159
Q

What does this diagram show in terms of additivty /linearity? - (5)

A

There is a a linear effect when the data increases at a steady rate like the graph on the left.

Your cost increases steadily as the number of chocolate bars increases.

The graph on the right shows a non-linear effect when there is not this steady increase rather there is a sharp change in your data.

So you might feel ok if you eat a few chocolate bars but after that the risk of you having a stomach ache increases quite rapidly the more chocolates you eat.

This effect is super important to check or your statistical analysis will be wrong even if your other assumptions are correct because a lot of statistical tests are based on linear models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
160
Q

Discrepnacy between measurement and actual value in population is .. and not..

A

measurement error and NOT variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
161
Q

Measurement error can happen across all psychological experiments from.. to ..

A

recording instrument failure to human error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
162
Q

What are the 2 types of measurement errors? - (2)

A
  1. Systematic
  2. Random
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
163
Q

What is systematic measurement error?

A

: predictable, typically constant or proportional to the true value and always affect the results of an experiment in a predictable direction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
164
Q

Example of systematic measurement error

A

for example if I know I am 5ft2 and when I go to get measured I’m told I’m 6ft this is a systematic error and pretty identifiable - these usually happen when there is a problem with your experiment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
165
Q

What is random measurement error?

A

measurable values being inconsistent when repeated measures of a constant attribute or quantity are taken.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
166
Q

Example of random measurement error

A

for example my height is 5ft2 when I measure it in the morning but its 5ft when I measure myself in the evening. This is because my measurements were taken at different times so there would be some variability – for those of you who believe you shrink throughout the day.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
167
Q

What is variance?

A

Average squared deviation of each number from its mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
168
Q

Variability is an inherent part of

A

things being measured and of the measurement process

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
169
Q

Diagram of variance formula

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
170
Q

In central limit theorem - (2)

A

states that the sampling distribution of the mean approaches a normal distribution, as the sample size increases. This fact holds especially true for sample sizes over 30.

Therefore, as a sample size increases, the sample mean and standard deviation will be closer in value to the population mean μ and standard deviation σ .

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
171
Q

What does histogram look at? - (2)

A

Frequency of scores

Look at distribution of data, skewness, kurotsis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
172
Q

What does boxplot look at? - (2)

A

To identify outliers

Shows median rather than mean (good for non-normally distributed data)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
173
Q

What do line graphs are?

A

simply bar charts with lines instead of bars

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
174
Q

Bar charts are a good way to display

A

display means (and
standard errors)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
175
Q

What do scatterplot illustrates? - (2)

A

a relationship between two variables, e.g. correlation or regression

Only use regression lines for regressions!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
176
Q

What are matrix scatterplots? - (2)

A

Particular kind of scatterplot that can be used instead of the 3-D scatterplot

clearer to read

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
177
Q

Using data provided how would you summarise skew?

A. The data has an issue with positive skew
B.The data has an issue with negative skew
C.The data is normally distributed

A

B

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
178
Q

What is the median number of bullets shot at a partner by females?

A

67.00

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
179
Q

What descriptive statistics does the red arrow represents?

A. Inter quartile range

B. Median

C. Mean

D. Range

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
180
Q

What is the mean of males and females SD? - (2)

A

Males M = 27.29

Females SD = 12.20

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
181
Q

What is the respective standard error of mean for femals and males?

A

3.26 & 3.42

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
182
Q

Answering the question: Meets assumption of parametric tests will determine whether our continous data can be tested with

A

with parametric or non-parametric tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
183
Q

A normal distribution is a distribution with the same general shape which is a

A

bell shape

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
184
Q

A normal distribution curve is symmetric around

A

the mean μ

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
185
Q

A normal distribution is defined by two parameters - (2)

A

the mean (μ) and the standard deviation (σ).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
186
Q

Many statistical tests (parametric) cannot be used if the data is not

A

normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
187
Q

What does this diagram show? - (2)

A

μ = 0 is peak of distribution

Block areas under the curve and gives us insight to way data is distributed and certain scores occuring if they belong to normally distribution e.g., 34.1% of values lie one SD below mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
188
Q

A z score in standard normal distribution will reflect the number of

A

SD above or below the mean of a particular score is

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
189
Q

How to calculate a z score?

A

Take a value of participant (e.g., 56 years old) and take away mean of distribution (e.g., mean age of class is 23) divided by SD (class like 2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
190
Q

If a person scored a 70 on a test with a mean of 50 and a standard deviation of 10

Converting the test scores to z scores, an X of 70 would be…

What the result means…. - (2)

A

a z score of 2 means the original score was 2 standard deviations above the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
191
Q

We can convert our z scores to

A

pecentiles

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
192
Q

Example: What is the percentile rank of a person receving a score of 90 on the test? - (3)

Mean - 80
SD = 5

A

First calculating z score: graph shows that most people scored below 90. Since 90 is 2 standard deviations above the mean z = (90 - 80)/5 = 2

Z score to pecentile can be looked at table that z score of 2 is equivalent to the 97.7th percentle:

The proportion of people scoring below 90 is thus .977 and proportion of people scoring above 90 is 2.3% (1-0.977)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
193
Q

What is the sample mean?

A

an unbiased estimate of the population mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
194
Q

How can we know how that our sample mean estimate is representative of the population mean?

A

Via computing standard error of mean - smaller SEM the better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
195
Q

Standard deviation is used as a measure of how

A

representative the mean was of the observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
196
Q

Small standard deviations represented a scenario in which most data points were

A

most data points were close to the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
197
Q

Large standard deviation represented a situation in which data points were

A

widely spread
from the mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
198
Q

How to calculate the standard error of mean?

A

computed by dividing the standard deviation of the sample by the the square root of the number in the sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
199
Q

The larger the sample the smaller the - (2)

A

standard error of the mean

more confident we can be that the sample mean is representative of the population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
200
Q

The central limit therom proposes that

A

as samples get large (usually defined as greater than 30), the sampling distribution has a normal distribution with a mean equal to the population mean, SD = SEM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
201
Q

The standard deviation of sample means is known as the

A

SEM (standard error of the mean)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
202
Q

A different approach to assess accuracy of sample mean as estimate of - population mean, aside from SE, is to - (2)

A

calculate boundaries and range of values within which we believe the true value of the population mean value will fall.

Such boundaries are called confidence intervals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
203
Q

Confidence intervals are created by

A

samples

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
204
Q

A 95% confidence intervals is consructed such that

A

these intervals (created by samples) will contain the population mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
205
Q

95% Confidence interval for 100 samples (CI constructed for each) would mean

A

95 of these samples, the confidence intervals we constructed would contain the true value of the mean in the population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
206
Q

Diagram shows- (4)

A
  • Dots show the means for each sample
  • Lines sticking out representing Ci for the sample means
  • If there was a vertical line down it represents population mean
  • If confidence intervals don’t overlap then it shows significant difference between the sample means
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
207
Q

In fact, for a specific confidence interval, the probability that it contains the population value is either - (2)

A

0 (it does not contain it) or 1 (it does contain it).

You have no way of knowing which it is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
208
Q

if our sample means were normally distributed with a mean of 0 and a
standard error of 1, then the limits of our confidence interval

A

would be –1.96 and +1.96 -

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
209
Q

95% of z scores fall between

A

-1.96 and 1.96

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
210
Q

Confidence intervals can be constructed for any estimated parameter, not just

A

μ - mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
211
Q

. If the mean represents the true mean well, then the confidence interval of that mean should be

A

small

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
212
Q

if the confidence interval is very
wide then the sample mean could be

A

very different from the true mean, indicating that it
is a bad representation of the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
213
Q

Remember that the standard error of the mean gets smaller with the number of observations and thus our confidence interval also gets

A

smaller - make sense as more we measure more certain sample mean close to population mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
214
Q

Calculating Confidence Intervals for sample means - rearranging in z formula

A

LB = Mean - (1.96 * SEM)
UB = Mean + (1.96 * SEM)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
215
Q

The standard deviation of SAT verbal scores in a school system is known to be 100. A researcher wishes to estimate the mean SAT score and compute a 95% confidence interval from a random sample of 10 scores.

The 10 scores are: 320, 380, 400, 420, 500, 520, 600, 660, 720, and 780.

Calculate CI

A

* M - 530
* N = 10
* SEM = 100/ square root of 10 = 31.62

* Value of z for 95% CI is number of SD one must go from mean (in both directions) to contain 0.95 of the scores
* Value of 1.96 was found in z-table
* Since each tail is to contain 0.025 of the scores, you find the values of z for which is 1-0.025 = 0.975 of the socres below
* 95% of z scores lie between -1.96 and +1.96
* Lower limit = 530 - (1.96) (31.62) = 468.02
* Upper limit = 530 + (1.96)(31.62) = 591.98

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
216
Q

Think of test statistic capturing

A

signal/noise

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
217
Q

Hypo

A testStatistic for which the frequency of particular values is known (t, F, chi-square) and thus we can calculate the

A

probability of obtaining a certain value or p value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
218
Q

To test whether the model fits the data or whether our hypothesis is a good explanation of the data, we compare

A

systematic variation against unsystematic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
219
Q

If the probability (p-value) less than or equal to the significance level, then

A

the null hypothesis is rejected; When the null hypothesis is rejected, the outcome is said to be “statistically significant”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
220
Q

If the proabilibty (p-value) is greater than the signifiance leve, the

A

null hypothesis is not rejected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
221
Q

What is a type 1 error in terms of variance? - (2)

A

think the variance accounted for by the model is larger than the one unaccounted for by the model (i.e. there is a statistically significant effect but in reality there isn’t)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
222
Q

Type 1 is a false

A

positive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
223
Q

What is type II error in temrs of variance?

A

think there was too much variance unaccounted for by the model (i.e. there is no statistically significant effect but in reality there is)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
224
Q

Type II error is false

A

negative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
225
Q

Example of Type I and Type II error

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
226
Q

Type I and Type II errors are mistakes we can make when testing the

A

fit of the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
227
Q

Type 1 errors when we believe there is a geniue effect in

A

population, when in fact there isn’t.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
228
Q

Acceptable level of type I error is usually

A

a-level of usually 0.05

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
229
Q

Type II error occurs when we believe there is no effect in the

A

population when, in reality, there is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
230
Q

Acceptable level of Type II error is probability/-p-value is

A

β-level (often 0.2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
231
Q

An effect size is a standardised measure of

A

the size of the an effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
232
Q

Properities of effect size (3)

A

Standardized = comparable across studies

Not (as) reliant on the sample size

Allows people to objectively evaluate the size of observed effect.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
233
Q

Effect Size Measures

r = 0.1, d = 0.2 (small effect):

A

the effect explains 1% of the total variance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
234
Q

Effect size measures

r = 0.3, d = 0.5 (medium effect) means

A

the effect accounts for 9% of the total variance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
235
Q

Effect size measures

r = 0.5, d = 0.8 (large effect)

A

effect accounts for 25% of the variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
236
Q

Beware of the ‘canned’ effect sizes (e.g., r = 0.5, d = 0.8 and rest) since the size of

A

effect should be placed within the research context.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
237
Q

We should aim to achieve a power of

A

.8, or an 80% chance of detecting
an effect if one genuinely exists.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
238
Q

When we fail to reject the null hypothesis, it is either that there truly are no difference to be found,

OR

A

it may be because we do not have enough statistical power

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
239
Q

Power is the probability of

A

correctly rejecting a false H0 OR the ability of the test to find an effect assuming there is one in the population,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
240
Q

Power is calculated by

A

1 - β OR probability of making Type II error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
241
Q

To increase statistical power of study you can increase

A

your sample sizee

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
242
Q

Factors affecting the power of the test: (4):

A
  1. Probability of a type 1 error or a-level [level at which we decide effect is sig - p-value) –> bigger [more lenient] alpha then more power)
  2. True alternate hypothesis H1 [effect size] (degree of overlap, less means more power) - if you find large effect in lit then better chance of detecting something
  3. The sampel size [N]) –> bigger the sample, less the noise and more power
  4. The particular tests to be employed - parametric tests greater power to detect sig effect since more sensitive
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
243
Q

How to calculate the number of pps they need for reasonable chance of correctly rejecting null hypothesis?

A

Sample size calculation at a desired level of power (usually power set to 0.8 in formula)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
244
Q

With power, we can do 2 things - (2)

A
  • Calculate power of test
  • Calculate sample size necessary to detect an decent effect size and achieve a certain level of power based on past research
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
245
Q

Diagram of Type I error, Type II error, power - (4) and making correct decisions

A

Type 1 error p = alpha
Type II error p = beta
Accepting null hypothesis which is correct - p = 1- alpha
Accepting alternate hypo which is correct - p = 1 - beta

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
246
Q

If there is a less degree of overlap in h0 and h1 then

A

bigger difference means higher power and and correctly reject the null hypothesis than distributions that overlap more

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
247
Q

If distribution between h0 and h1 are narrower then

A

This means that the overlap in distributions is smaller and the power is therefore greater, but this time because of a smaller standard error of our estimate of the means.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
248
Q

Most people want to assess how many participants they need to test to have a reasonable chance of correctly rejecting the null hypothesis (the Power). This formula shows - (2)

A

us how.

We usually set the power to 0.8.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
249
Q

What is z scores? - (2)

A

A measure of variability:

The number of standard deviations from the population mean or a particular data point is
Z-scores are a standardised measure, hence they ignore measurement units

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
250
Q

Why should we care about z scores? - (2)

A

Z-scores allow researchers to calculate the probability of a score occurring within a standard normal distribution

Enables us to compare two scores that are from different samples (which may have different means and standard deviations)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
251
Q

Diagram of finding percentile of Trish

Trish takes a test and gets 25

Mean of the class is 20

SD = 4

25-20/4 = 1.25

Z-score = 1.25

A

Let’s say Trish takes a test and scores 25 and the mean is 20 You may calculate the z-score to be 1.25 you would use a z-score table to see what percentile they would be in (marked in red) so to read the table you would go down to the value 1.2 and you would go across to 0.05 which totals to 1.25 and you can see about 89.4% of other students performed worse.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
252
Q

Diagram of z score and percentile

Josh takes a different test and gets 1150
Mean of the class is 1000
SD = 150

1150 – 1000/150 = 1.0
Z score = 1.0

Who performed better Trish or Josh?

Trish had z score of 1.25

A

We would use our table and look down the column to a z-score of 1 and across to the 0.00 column (in purple) and we can see 84.1% of students performed worse than Josh so Trish performed better than Josh.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
253
Q

Diagram of z scores and normal distribution - (3)

A

68% of scores are within 1 SD of the mean,

95% are within 2 SDs and

99.7% are within 3 SDs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
254
Q

Whats standard error?

A

: by taking into account the variability and size of our sample we can estimate how far away from the real population mean our mean is!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
255
Q

If we took infinite samples from the population, 95% of the time the population mean will lie within the

A

the 95% confidence interval range

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
256
Q

What does narrow CI represent?

A

high statistical power

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
257
Q

Wide CIs represent?

A

low statistical power

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
258
Q

Power bring the probability of catching a real effect (as opposed to

A

missing a real effect – Type II error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
259
Q

We can never say the null hypothesis is

A

FALSE (or TRUE).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
260
Q

TheP valueor calculated probability is the estimated probability of us

A

us finding an effect when the null hypothesis (H0) is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
261
Q

p = probability of observing a test statistic at least as a big as the one we have if the

A

H0 is true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
262
Q

Hence, a significant p value (p <.05) tells us that there is a less than 5% chance of getting a test statistic that is

A

larger than the one we have found if there were no effect in the population (e.g. the null hypothesis were true)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
263
Q

Statistical signifiance does not equal importance - (2)

A

p = .049, p = .050 are essentially the same thing- the former is ‘statistically significant’.

Importance is dependent upon the experimental design/aims: e.g., A statistically significant weight increase of 0.1Kg between two adults experimental groups may be less important than the same increase between two groups of babies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
264
Q

Children can learn a second language faster before the age of 7’. Is this statement:

A. One-tailed

B. A non scientific

C. Two-tailed

D. Null hypothesos

A

A as one-tailed is directional and two tailed is non-direcitonal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
265
Q

Which of the following is true about a 95% confidence interval of the mean:

A. 95 out of 100 CIs wll contain population mean

B. 95 out of 100 sample means will fall within the limits of the confidence interval.

C. 95% of population means will fall within the limits of the confidence interval.

D. There is a 0.05 probability that the population mean falls within the limits of the confidence interval.

A

A as If we’d collected 100 samples, calculated the mean and then calculated a confidence interval for that mean, then for 95 of these samples the confidence intervals we constructed would contain the true value of the mean in the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
266
Q

What does a significant test statistic tell us?

A. That the test statistic is larger than we would expect if there were no effect in the population.

B. There is an important effect.

C. The null hypothesis is false.

D. All of the above.

A

A and just because test statistic is sig does not mean its important effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
267
Q

Of what is p the probability?

(Hint: NHST relies on fitting a ‘model’ to the data and then evaluating the probability of this ‘model’ given the assumption that no effect exists.)

A.p is the probability of observing a test statistic at least as big as the one we have if there were no effect in the population (i.e., the null hypothesis were true).

B. p is the probability that the results are due to chance, the probability that the null hypothesis (H0) is true.

C. p is the probability that the results are not due to chance, the probability that the null hypothesis (H0) is false

D. p is the probability that the results would be replicated if the experiment was conducted a second time.

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
268
Q

A Type I error occurs when:

(Hint: When we use test statistics to tell us about the true state of the world, we’re trying to see whether there is an effect in our population.)

A. We conclude that there is an effect in the population when in fact there is not.

B. We conclude that there is not an effect in the population when in fact there is.

C. We conclude that the test statistic is significant when in fact it is not.

D. The data we have typed into SPSS is different from the data collected.

A

A as If we use the conventional criterion then the probability of this error is .05 (or 5%) when there is no effect in the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
269
Q

True or false?

a. Power is the ability of a test to detect an effect given that an effect of a certain size exists in a population.

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
270
Q

True or False?

We can use power to determine how large a sample is required to detect an effect of a certain size.

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
271
Q

True or False?

c. Power is linked to the probability of making a Type II error.

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
272
Q

True or False?

d. The power of a test is the probability that a given test is reliable and valid.

A

FALSE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
273
Q

What is the relationship between sample size and the standard error of the mean?

(Hint: The law of large numbers applies here: the larger the sample is, the better it will reflect that particular population.)

A. The standard error decreases as the sample size increases.

B. The standard error decreases as the sample size decreases.

C. The standard error is unaffected by the sample size.

D. The standard error increases as the sample size increases.

A

A The standard error (which is the standard deviation of the distribution of sample means), defined as σ_Χ ̅ =σ/√N, decreases as the sample size (N) increases and vice versa

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
274
Q

What is the null hypothesis for the following question: Is there a relationship between heart rate and the number of cups of coffee drunk within the last 4 hours?

A. There will be no relationship between heart rate and the number of cups of coffee drunk within the last 4 hours.

B. People who drink more coffee will have significantly higher heart rates.

C. People who drink more cups of coffee will have significantly lower heart rates.

D. There will be a significant relationship between the number of cups of coffee drunk within the last 4 hours and heart rate

A

A The null hypothesis is the opposite of the alternative hypothesis and so usually states that an effect is absent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
275
Q

A Type II error occurs when :

(Hint: This would occur when we obtain a small test statistic (perhaps because there is a lot of natural variation between our samples.)

A. We conclude that there is not an effect in the population when in fact there is.

B. We conclude that there is an effect in the population when in fact there is not.

C. We conclude that the test statistic is significant when in fact it is not.

D. The data we have typed into SPSS is different from the data collected.

A

A A Type II error would occur when we obtain a small test statistic (perhaps because there is a lot of natural variation between our samples)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
276
Q

In general, as the sample size (N) increases:

A. The confidence interval gets narrower.

B. The confidence interval gets wider.

C. The confidence interval is unaffected.

D. The confidence interval becomes less accurate

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
277
Q

Which of the following best describes the relationship between sample size and significance testing?

(Hint: Remember that test statistics are basically a signal-to-noise ratio, so given that large samples have less ‘noise’ they make it easier to find the ‘signal’.)

A. In large samples even small effects can be deemed ‘significant’.

B. In small samples only small effects will be deemed ‘significant’.

C. Large effects tend to be significant only in small samples.

D. Large effects tend to be significant only in large samples.

A

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
278
Q

The assumption of homogeneity of variance is met when:

A. The variances in different groups are approximately equal.

B. The variances in different groups are significantly different.

C. The variance across groups is proportional to the means of those groups.

D. The variance is the same as the interquartile range.

A

A - To make sure our estimates of the parameters that define our model and significance tests are accurate we have to assume homoscedasticity (also known as homogeneity of variance)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
279
Q

Next, the lecturer was interested in seeing whether males and females reacted differently to the different teaching methods.

Produce a clustered bar graph showing the mean scores of teaching method for males and females.

(HINT: place TeachingMethod on the X axis, Exam Score on the Y axis, and Gender in the ‘Cluster on X’ box. Include 95% confidence intervals in the graph).

Which of the following is the most accurate interpretation of the data?

A.Females performed better than males both the reward and indifferent conditions. Regarding the confidence intervals, there was a large degree of overlap between males and females in all conditions of the teaching method.

B.Males performed better than females in the reward condition, and females performed better than males in the indifferent condition. Regarding the confidence intervals, there was no overlap between males and females across any of the conditions of teaching method.

C.Males performed better than females in all conditions. Regarding the confidence intervals, there was a small degree of overlap between males and females for the reward and indifferent conditions, and a large degree of overlap between males and females for the punish condition.

D.Males performed better than females in the reward condition, and females performed better than males in the indifferent condition. Regarding the confidence intervals, there was a small degree of overlap between males and females for the reward and indifferent conditions, and a large degree of overlap between males and females for the punish condition.

A

D

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
280
Q

Produce a line graph showing the change in mean anxiety scores over the three time points.

NOTE: this is a repeated measures (or within subjects) design, ALL participants took part in the same condition.

Which of the following is the correct interpretation of the data?

A.Mean anxiety increased across the three time points.

BMean anxiety scores were reduced across the three time points, and there was a slight acceleration in this reduction between the middle and end of the course.

CMean anxiety scores were reduced across the three time points, though this reduction slowed down between the middle and end of the course.

DMean anxiety scores did not change across the three time points.

A

B

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
281
Q

A general approach in regression is that our outcomes can be predicted by a model and what remains

A

is the error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
282
Q

The i in the general model in regression shows

A

e.g., outcome 1 is equal to model plus error 1 and outcome 2 is equal to model plus error 2 and so on…

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
283
Q

For correlation, the outcome is modelled by

A

scaling (multiplying by a constant) another variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
284
Q

Equation of correlation model

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
285
Q

If you have a 1 continous variable which meets assumtpion of parametric test then you can conduct a

A

pearson correlation or regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
286
Q

Variance is a feature of outcome measurements we have obtained and we want to predict with a model in correlation/regression that…

A

captures the effect of the predictor variables we have manipulated or measured

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
287
Q

Variance of a single variable represents the

A

average amount that the data cary from the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
288
Q

Variance is the standard deviation

A

squared (s squared)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
289
Q

Variance formula - (2)

A

xi minus average of all scores of pp which is squared and divided by total number of participants minus 1

done for each participant (sigma)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
290
Q

Variance is SD squared meaning that it captures the

A

average of the squared difference the outcome values from the mean of all outcomes (explaining what the formula of variance does)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
291
Q

Covariance gathers information on whether

A

one variable covarys with another

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
292
Q

In covariance if we are interested whether 2 variables are related then interested whether changes in one variable are met with changes in other

therefore.. - (2)

A

when one variable deviates from its mean we
would expect the other variable to deviate from its mean in a similar way.

So, if one variable increases then the other, related variable, should also increase or even decrease at the same level.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
293
Q

If one variable covaries with another variable then it means these 2 variables are

A

related

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
294
Q

To get SD from variance then you would

A

square root variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
295
Q

What would you do in covariance formula in proper words? - (5)

A
  1. Calculate the error between the mean and each subject’s score for the first variable (x).
  2. Calculate the error between the mean and their score for the second variable (y).
  3. Multiply these error values.
  4. Add these values and you get the product deviations.
  5. The covariance is the average product deviations
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
296
Q

Example of calculaitng covariance and what does answer tell you?

A

The answer ispositive: that tells us the x and y values tend to risetogether.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
297
Q

What does each element of covariance formula stand for? - (5)

A

X = the value of ‘x’ variable
Y = the value of ‘y’ variable
X(line) = mean of ‘x’ - e.g., green
Y(line) = mean of ‘y’ - e.g., blue
n = the number of items in the data set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
298
Q

covariance will be large when values below

A

the mean for one variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
299
Q

What does a positive covariance indicate?

A

as one variable deviates from the mean, the other
variable deviates in the same direction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
300
Q

What does negative covariance indicate?

A

a negative covariance indicates that as one variable deviates from the mean (e.g. increases), the other deviates from the mean in the opposite direction (e.g. decreases).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
301
Q

What is the problem of covariance as a measure of the relationship between 2 variables? - (5)

A

dependent upon the units /scales of measurement used

So covariance is not a standardised measure

e.g., if 2 variables measured in miles and covariance is 4.25 then if we convert data to kilometres then we have to calculate covariance again and see it increases to 11.

Dependence of scale measurement is a problem as can not compare covariances in an objective way –> can not say whether covariance is large or small to another data unless both data sets measured in same units

So we need to STANDARDISE it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
302
Q

What is the process of standardisaiton?

A

To overcome the problem of dependence on the measurement scale, we need to convert
the covariance into a standard set of units

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
303
Q

How to standardise the covariance?

A

dividing by product of the standard deviations of both variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
304
Q

Formula of standardising covariance

A

Same formula of covariance but multipled of SD of x and SD of y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
305
Q

Formula of Pearson’s correlation coefficient, r

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
306
Q

Example of calculating Pearson’s correlation coefficient, r - (5)

A

standard deviation for the number of adverts watched (sx)
was 1.67,

SD of number of packets of crisps bought (sy) was 2.92.

If we multiply these together we get 1.67 × 2.92 =
4.88.

.Now, all we need to do is take the covariance, which we calculated a few pages ago as being 4.25, and divide by these multiplied standard deviations.

This gives us r = 4.25/
4.88 = .87.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
307
Q

The standardised version of covariance is the

A

correlational coefficient or Pearson’s r

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
308
Q

Pearson’s R is … version of covariance meaning independent of units of measurement

A

standardised

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
309
Q

What does correlation describe? - (2)

A

Describes a relationship between variables

If one variable increases, what happens to the other variable?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
310
Q

Pearson’s correlation coefficient r was also called the

A

product-moment correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
311
Q

Linear relationship and normally disturbed data and interval/ratio and continous data is assumed in

A

Pearson’s r correlation coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
312
Q

Pearson Correlation Coefficient varies between

A

-1 and +1 (direction of relationship)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
313
Q

The larger the R Pearson’s correlation coefficient value, the closer the values will

A

be with each other and the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
314
Q

The smaller R Pearson’s correlation coefficient values indicate

A

there is unexplained variance in the data and results in the data points being more spread out.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
315
Q

What does these two graphs show? - (2)

A
  • example of high negative correlation. The data points are close together and are close to the mean.
  • On the other hand, the graph on the right shows a low positive correlation. The data points are more spread out and deviate more from the mean.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
316
Q

The Pearson Correlation Coefficient measures the strength of a relationhip

A

between one variable and another hence its use in calculating effect size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
317
Q

A Pearson’s correlation coefficient of +1 indicates

A

two variablesare perfectly positively correlated, so as one variable increases, the other increases by a
proportionate amount.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
318
Q

A Pearson’s correlation coefficient of -1 indicates

A

a perfect negative relationship: if one variable increases, the other decreases by a proportionate amount.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
319
Q

Pearson’s r
+/- 0.1 means

A

small effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
320
Q

Pearson’s r
+/- 0.3 means

A

medium effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
321
Q

Pearson’s r
+/- 0.5 means

A

large effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
322
Q

In Pearson’s correlation, we can test the hypothesis that - (2)

A

correlation coefficient is different from zero

(i.e., different from ‘no relationship’)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
323
Q

In Pearson’s correlation coefficient, we can test the hypothesis that the correlation is different from 0

If we find our observed coefficient was very unlikely to happen if there was no effect in population then gain confidence that

A

relationship that
we have observed is statistically meaningful.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
324
Q

. In the case of a correlation
coefficient we can test the hypothesis that the correlation is different from zero (i.e. different
from ‘no relationship’).

There are 2 ways to test this hypothesis

A
  1. Z scores
  2. T-statistic
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
325
Q

Confidence intervals tells us something

A

likely correlation in the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
326
Q

Can calculate confidence intervals of Pearson’s correlation coefficient by transforming formula of CI

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
327
Q

As sample size increases, so the value of r at which a significant result occurs

A

decreases e.g 20 n p is not < 0.05 but at 200 pps it is p < 0.05

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
328
Q

Pearson’s r = 0 means - (2)

A

indicates no linear relationship at all

so if one variable changes, the other stays the same.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
329
Q

Correlation coefficients give no indication of direction of… + example - (2)

A

causality

e.g., although we conclude no of adverts increase nmber of toffees bought we can’t say watching adverts caused us to buy toffees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
330
Q

We have to be caution of causality in terms of Pearson’s correlation r as - (2)

A
  • Third variable problem - causality between variables can not be assumed in any correlation
  • Direction of causality: Correlation coefficients give nothing about which variables causes other to change.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
331
Q

If you got weak correlation between 2 variables = weak effect then take a lot of measurements for that relationship to be

A

significant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
332
Q

R correlation coefficient gives the ratio of

A

covariance to a measure of variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
333
Q

Example of correlations getting stronger

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
334
Q

R squared is known as the

A

coefficient of determination

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
335
Q

of cor

R^2 can be used to explain the

A

proportion of the variance for a dependent variable )outcome) that’s explained by an independent variable . (predictor)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
336
Q

Example of R^2 coefficient of determination - (2)

X = exam anxiety
Y = exam performance

If R^2 = 0.194

A

19.4% of variability in exam performance can be explained by exam anxiety

the variance in y accounted for by x’,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
337
Q

R^2 calculate the amount of shared

A

variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
338
Q

Example of r and R^2

A

Multiply 0.1 * 0.1 for example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
339
Q

R^2 gives you the true strength of.. but without

A

the correlation but without an indication of its direction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
340
Q

What are the three types of correlations? - (3)

A
  1. Bivarate correlations
  2. Partial correlations
  3. Semi-partial or part correlations
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
341
Q

Whats bivarate correlation?

A

relation between 2 variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
342
Q

What is a partial correlation?

A

looks at the relationship between two variables while ‘controlling’ the effect of one or more additional variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
343
Q

The partial correlation partials out the

A

the effect of one or more variables on either X or Y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
344
Q

A partial correlation controls for third variable which is made from - (3)

A
  • A correlation calculates each data points distance from line (residuals)
  • This is the error relative to the model (unexplained variance)
  • A third variable might predict some of that variation in residuals
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
345
Q

The partial correlation compares the unique variation of one variable with the

A

unfiltiered variation of the other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
346
Q

The partial correlation holds the

A

third variable constant (but we don’t manipulate these)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
347
Q

Example of partial correlation- (2)

A

For example, when studying the effect of a diet, the level of exercise might also influence weight loss

We want to know the unique effect of diet, so we need to partial out the effect of exercise

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
348
Q

Example of Venn Diagram of Partial Correlation - (2)

A

Partial Correlation between IV1 and DV = D / D+C

Unique variance accounted for by the predictor (IV1) in the DV, after accounting for variance shared with other variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
349
Q

Example of Partial Correlation - (2)

A

Partial correlation: Purple / Red + Purple

If we were doing just a partial correlation, we would see how much exam anxiety is influencing both exam performance and revision time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
350
Q

Example of partial correlation and semi-partial correlation - (2)

A

The partial correlation that we calculated took
account not only of the effect of revision on exam performance, but also of the effect of revision on anxiety.

If we were to calculate the semi-partial correlation for the same data, then this would control for only the effect of revision on exam performance (the effect of revision
on exam anxiety is ignored).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
351
Q

In partial correlation, the third variable is typically not considered as the primary independent or dependent variable. Instead, it functions as a

A

control variable—a variable whose influence is statistically removed or controlled for when examining the relationship between the two primary variables (IV and DV).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
352
Q

The partial correlation is

The amount of variance the variable explains

A

relative to the amount of variance in the outcome that is left to explain after the contribution of other predictors have been removed from both the predictor and outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
353
Q

These partial correlations can be done when variables are dichotomous (including third variable) e.g., - (2)

A

we could look at the relationship between bladder relaxation (did the person wet themselves or not?) and the number of large tarantulas crawling up your leg controlling for fear of spiders

(the first variable is dichotomous, but the second variable and ‘controlled for’ variables are continuous).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
354
Q

What does this partial correlation output show?

Revision time = partial, controlling for its effect

Exam performance = DV

Exam anxiety = X - (5)

A
  • . First, notice that the partial correlation between exam performance and exam anxiety is −.247, which is considerably less than the correlation when the effect of
    revision time is not controlled for (r = −.441).
  • . Although this correlation is still statistically significant (its p-value is still below .05), the relationship is diminished.
  • value of R2 for the partial correlation is .06, which means that exam anxiety can now account for only 6% of the variance in exam performance.
  • When the effects of revision time were not controlled for, exam anxiety shared 19.4% of the variation in exam scores and so the inclusion of revision time has severely diminished the amount of variation in exam scores shared by anxiety.
  • As such, a truer measure of the role of exam anxiety has been obtained.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
355
Q

Partial correlations are most useful for looking at the unique
relationship between two variables when

A

other variables are ruled out

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
356
Q

In a semi-partial correlation we control for the

A

effect that
the third variable has on only one of the variables in the correlation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
357
Q

The semi partial (part) correlation partials out the - (2)

A

Partials out the effect of one or more variables on either X or Y.

e.g. The amount revision explains exam performance after the contribution of anxiety has been removed from the one variable (usually the predictor- e.g. revision).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
358
Q

The semi-partial correlation compares the

A

unique variation of one variable with the unfiltered variation of the other.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
359
Q

Diagram of venn diagram of semi-partial correlation - (2)

A
  • Semi-Partial Correlation between IV1 and DV = D / D+C+F+G

Unique variance accounted for by the predictor (IV1) in the DV, after accounting for variance shared with other variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
360
Q

Diagram of revision and exam performance and revision time on semi-partial correlation - (2)

A
  • purple/red + purple + white+ orange
  • When we use semi-partial correlation to look at this relationship, we partial out the variance accounted for by exam anxiety (the orange bit) and look for the variance explained by revision time (the purple bit).
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
361
Q

Summary of partial correlation and semi-partial correlation - (2)

A

A partial correlation quantifies the relationship between two variables while accounting for the effects of a third variable on both variables in the original correlation.

A semi-partial correlation quantifies the relationship between two variables while accounting for the effects of a third variable on only one of the variables in the original correlation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
362
Q

Pearson’s product-moment correlation coefficient (described earlier) and Spearman’s rho (see section 6.5.3) are examples of

A

of bivariate correlation coefficients.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
363
Q

Non-parametric tests of correlations are… (2)

A
  • Spearman’s roh
  • Kendall’s tau test
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
364
Q

In spearman’s rho the variables are not normally distributed and measures are on a

A

ordinal scale (e.g., grades)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
365
Q

Spearman’s rho works on by

A

first ranking the data n(numbers converted into ranks), and then running Pearson’s r on the ranked data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
366
Q

Spearman’s correlation coefficient, rs, is a non-parametric statistic and so can be used when the data have

A

data have violated parametric assumptions such as nonnormally distributed data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
367
Q

In spearman correlation coefficient is sometimes called

A

Spearman’s rho

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
368
Q

For spearman’s r we can get R squared but it is interpreted slightly different as

A

proportion of
variance in the ranks that two variables share.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
369
Q

Kendall’s tau used rather than Spearman’s coefficient when - (2)

A

when you have a small data set with a large number of
tied ranks.

This means that if you rank all of the scores and many scores have the same rank, then Kendall’s tau should be used

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
370
Q

Kendall’s tau test - (2)

A

For small datasets, many tied ranks

Better estimate of correlation in population than Spearman’s ρ

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
371
Q

Kendall’s tau is not numerically similar to r or rs (spearman) and so tau squared does not tell us about

A

proportion of
variance shared by two variables (or the ranks of those two variables).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
372
Q

The Kendall’s tau is 66-75% smaller than both Spearman’s r and Pearson’s r so

A

tau is not comparable to r and r s

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
373
Q

There is a benefit using Kendall’s statistic than Spearman as it shows - (2)

A

Kendall’s statistic is actually a better estimate of the correlation in the population

we can draw more accurate generalizations from Kendall’s statistic than from Spearman’s.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
374
Q

Whats the decision tree for Spearman’s correlation? - (4)

A
  • What type of measurement = continous
  • How many predictor variables = one
  • What type of continous variable = continous
  • Meets assumption of parametric tests - No
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
375
Q

The output of Kendall and Spearman can be interpreted the same way as

A

Pearson’s correlation coefficient r output box

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
376
Q

The biserial and point-biserial correlation coefficients used when

A

one of the two variables is dichotomous (e.g., example of dichotomous variable is women being pregnant or not)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
377
Q

What is the difference between biserial and point-biserial correlations?

A

depends on whether the dichotomous variable is discrete or continuous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
378
Q

The point–biserial correlation coefficient (rpb) is used when

A

one variable is a
discrete dichotomy (e.g. pregnancy),

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
379
Q

biserial correlation coefficient (rb) is used
when - (2)

A

one variable is a continuous dichotomy (e.g. passing or failing an exam).

e.g. An example is passing or failing a statistics test: some people will only just fail while others will fail by
a large margin; likewise some people will scrape a pass while others will clearly excel.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
380
Q

Example of when point=biserial correlation used - (3)

A
  • Imagine interested in relationship between gender of a cat and how much time it spent away from home
  • Time spent away is measured in interval level –> mets assumptions of parametric data
  • Gender is discrete dichotomous variable coded with 0 for male and 1 for female
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
381
Q

Can convert point-biserial correlation coefficient into

A

biseral correlation coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
382
Q

Point biserial and biserial correlation differ in size as

A

biserial correlation bigger than point biserial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
383
Q

Example of queston conducting Pearson’s r (4) -

A

The researchers was interested in whether the amount someone gets paid and amount of holidays they take from work, whether these two variables would be related to their productivity at work

  • Pay: Annual salary
  • Holiday: Number of holiday days taken
  • Productivity: Productivity rating out of 10
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
384
Q

Example of Pearson’s r scatterplot :

relationship between pay and productivity

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
385
Q

If we have r = 0.313 what effect size is it?

A

medium effect size

±.1 = small effect
±.3 = medium effect
±.5 = large effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
386
Q

What does this scatterplot show?

A

o This indicates very little correlation between the 2 variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
387
Q

What will a matrix scatterplot show?

A

the relationship between all possible combinations of your variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
388
Q

What does this scatterplot matrix show? - (2)

A
  • For Pay and Holiday, we can see the line is very flat and indicates the correlation between the two variables is quite low
  • For pay and productivity, the line is steeper suggesting the correlation is fairly substantial between these 2 variables and same for holidays and pay and productivity and holidays here
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
389
Q

What is degrees of freedom for correlational analysis?

A

N-2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
390
Q

What does this Pearson’s correlation r output show? - (4)

A
    • The relationship between pay and holidays is very low correlation is -0.04
    • Between pay and productivity, there is a medium size correlation of r = 0.313
  • Between holidays and productivity there is medium going on large effect size of 0.435
  • Relationship between pay and productivity and also holidays and productivity is sig but correlation with pay and holidays was not sig
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
391
Q

Another examp;e of Pearson’s correlation r question - (3)

A

A student was interested in the relationship between the time spent preparing an essay, the interestingness of the essay topic and the essay mark received.

He got 45 of his friends and asked them to rate, using a scale from 1 to 7, how interesting they thought the essay topic was (1 - I’ll kill myself of boredom, 4 - it’s not too bad!, 7 - it’s the most interesting thing in the world!) (interesting).

He then timed how long they spent writing the essay (hours), and got their percentage score on the essay (essay).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
392
Q

Example of interval/ratio continous data needed for Pearson’s r for IV and DV - (2)

A
  • Interval scale: difference between 10 degrees C and 20 degrees is same as 80 F and 90 F, 0 degrees does not mean absence of temp
  • Ratio: Height as 0 cm means no weight and weight, time
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
393
Q

Pearson’s correlation r , spearman and kendall equires

A

one IV and one DV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
394
Q

What does this SPSS output show?

A. There was a non-significant positive correlation between interestingness of topic and the amount of time spent writing. There was a non-significant positive correlation between time spent writing an essay and essay mark
There was a significant positive correlation between interestingness of topic and essay mark, with a medium effect size

B. There was a significant positive correlation between interestingness of topic and the amount of time spent writing, with a small effect size.There was a significant positive correlation between time spent writing an essay and essay mark, with a large effect size. .There was a non-significant positive correlation between interestingness of topic and essay mark

C. There was a significant negative correlation between interestingness of topic and the amount of time spent writing, with a medium effect size.. There was a non-significant positive correlation between time spent writing an essay and essay mark. There was a non-significant positive correlation between interestingness of topic and essay mark

D. There was a significant positive correlation between interestingness of topic and the amount of time spent writing, with a large effect size. There was a non-significant positive correlation between time spent writing an essay and essay mark There was a non-significant positive correlation between interestingness of topic and essay mark

A

D. There was a significant positive correlation between interestingness of topic and the amount of time spent writing, with a large effect size. There was a non-significant positive correlation between time spent writing an essay and essay mark There was a non-significant positive correlation between interestingness of topic and essay mark

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
395
Q

r = 0.21 effect size is..

A

in between small and medium effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
396
Q

Effect size is only meaningful if you evaluatte it witth regards to

A

your own research area

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
397
Q

Biserial correlaion is when

A

one variable is dichotomous, but there is an underlying continuum (e.g. pass/fail on an exam)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
398
Q

Pointt biserial correlation is when

A

When one variable is dichotomous, and it is a true dichotomy (e.g. pregnancy)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
399
Q

Example of dichotomous relationship - (3)

A
  • example of a true dichotomous relationship.
  • We can compare the differences in height between males and females.
  • Use dichotomous predictor of gender
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
400
Q

What is the decision tree for multiple regression? - (4)

A
  • Continous
  • Two or more predictors that are continous
  • Multiple regression
  • Meets assumptions of parametric tests
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
401
Q

Multiple regression is the same as simple linear regression expect for - (2)

A

every extra predictor you include, you have to add a coefficient;

so, each predictor variable has its own coefficient, and the outcome variable is predicted from a combination of all the variables multiplied by their respective coefficients plus a residual term

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
402
Q

Multiple regression equation

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
403
Q

In multiple regression equation, list all the terms - (5)

A
  • Y is the outcome variable,
  • b1 is the coefficient of the first predictor (X1),
  • b2 is the coefficient of the second predictor (X2),
  • bn is the coefficient of the nth predictor (Xn),
  • εi is the difference between the predicted and the observed value of Y for the ith participant.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
404
Q

Multiple regression uses the same principle as linear regression in a way that

A

we seek to find the linear combination of predictors that correlate maximally with the outcome variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
405
Q

Regression is a way of predicting things that you have not measured by predicting

A

an outcome variable from one or more predictor variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
406
Q

Can’t plot a 3D plot of MR as shown here

A

for more than 2 predictor (X) variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
407
Q

If you got two prediictors thart overlap and correlate a lot then it is a .. model

A

bad model can’t uniquely explain the outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
408
Q

In Hierarchical regression, we are seeing whether

A

one model explains significantly more variance than the other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
409
Q

In hierarchical regression predictors are selected based on

A

past work and the experimenter
decides in which order to enter the predictors into the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
410
Q

As a general rule for hierarchical regression, - (3)

A

known predictors (from other research) should be entered into the model first in order of their importance in predicting the outcome.

After known predictors have been entered, the
experimenter can add any new predictors into the model.

New predictors can be entered either all in one go, in a stepwise manner, or hierarchically (such that the new predictor
suspected to be the most important is entered first).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
411
Q

Example of hierarchical regression in terms of album sales - (2)

A

The first model allows all the shared variance between Ad budget and Album sales to be accounted for.

The second model then only has the option to explain more variance by the unique contribution from the added predictor Plays on the radio.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
412
Q

What is forced entry MR?

A

method in which all predictors are forced
into the model simultaneously.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
413
Q

Like HR, forced entry MR relies on

A

good theoretical reasons for including the chosen predictors,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
414
Q

Different from HR, forced entry MR

A

makes no decision about the order in which variables are entered.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
415
Q

Some researchers believe that about forced entry MR that

A

this method is the only appropriate method for theory testing because stepwise techniques are influenced by random variation in the data and so rarely give replicable results if the model is retested.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
416
Q

Why select colinearity diagnostics in statistics box for multiple regression? - (2)

A

This option is for obtaining collinearity statistics such as the
VIF, tolerance,

Checking assumption of no multicolinearity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
417
Q

Multicollinearity poses a problem only for multiple regression because

A

simple regression requires only one predictor.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
418
Q

Perfect collinearity exists in multiple regression when at least

A

e.g., two predictors are perfectly correlated , have a correlation coefficient of 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
419
Q

If there is perfect collinearity in multiple regression between predictors it
becomes impossible

A

to obtain unique estimates of the regression coefficients because there are an infinite number of combinations of coefficients that would work equally well.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
420
Q

Good news is perfect colinearity in multiple regression is rare in

A

real-life data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
421
Q

If two predictors are perfectly correlated in multiple regression then the values of b for each variable are

A

interchangable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
422
Q

As colinearity increases in multiple regression, there are 3 problems that arise - (3)

A
  • Untrustory bs
  • Limit size of R
  • Importance of predictors
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
423
Q

One way of identifying multicollinearity in multiple regression is to scan a

A

a correlation matrix of all of the predictor
variables and see if any correlate very highly (by very highly I mean correlations of above .80
or .90)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
424
Q

The VIF indicates in multiple regression whether a

A

predictor has a strong linear relationship with the other predictor(s).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
425
Q

If VIF statistic above 10 or approaching 10 in multiple regression then what you would want to do is have a - (2)

A

look at your variables to see if you need to include all variables whether all need to go in model

if high correlation between 2 predictors (measuring same thing) then decide whether its important to include both vars or take one out and simplify regression model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
426
Q

Related to the VIF in multiple regression is the tolerance
statistic, which is its

A

reciporal (1/VIF) = inverse of VIF

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
427
Q

In Plots in SPSS, you put in multiple regression - (2)

A

ZRESID on Y and ZPRED on X

Plot of residuals against predicted to asses homoscedasticity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
428
Q

What is ZPRED in MR? - (2)

A

(the standardized predicted values of the dependent variable based on the model).

These values are standardized forms of the values predicted by the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
429
Q

What is ZRESID in MR? - (2)

A

(the standardized residuals, or errors).

These values are the standardized differences between the observed data and the values that the model predicts).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
430
Q

SPSS in multiple linear regression gives descriptive outcoems which is - (2)

A
  • basics means and also a table of correlations between variables.
  • This is a first opportunity to determine whether there is high correlation between predictors, otherwise known as multi-collinearity
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
431
Q

In model summary of SPSS, it captures how the model or models explain in MR

A

variance in terms of R squared, and more importantly how R squared changes between models and whether those changes are significant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
432
Q

Diagram of model summary

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
433
Q

What is the measure of R^2 in multiple regression

A

measure of how much of the variability in the outcome is accounted for
by the predictors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
434
Q

The adjusted R^2 gives us an estimate of in multiple regression

A

fit in the general population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
435
Q

The Durbin-Watson statistic if specificed in multiple regresion tells us whether the - (2)

A

assumption of independent errors is tenable (value less than 1 or greater than 3 raise alarm bells)

value closer to 2 the better = assumption met

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
436
Q

SPSS output for MR = ANOVA table which performs

A

F-tests for each model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
437
Q

SPSS output for MR contains ANOVA that tests whether the model is

A

significantly beter at predicting the outcome than using the mean as a ‘best guess’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
438
Q

The F-ratio represents the ratio of

A

improvement in prediction that results from fitting the model, relative to the inaccuracy that still exists in the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
439
Q

We are told the sum of squares for model (SSM) - MR regression line in output which represents

A

improvement in prediction resulting from fitting a regression line to the data rather than using the mean as an estimate of the outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
440
Q

We are told residual sum of squares (Residual line) in this MR output which represents

A

total difference between
the model and the observed data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
441
Q

DF for Sum of squares Model for MR regression line is equal to

A

number of predictors (e.g., 1 for first model, 3 for second)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
442
Q

DF for Sum of Squares Residual for MR is - (2)

A

Number of observations (N) minus number of coefficients in regression model

(e.g., M1 has 2 coefficents - one for predictor and one for constant, M2 has 4 - one for each 3 predictor and one for constant)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
443
Q

The average sum of squares in ANOVA table is calculated by

A

calculated for each term (SSM, SSR) by dividing the SS by the df. T

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
444
Q

How is the F ratio calculated in this ANOVA table?

A

F-ratio is calculated by dividing the average improvement in prediction by the model (MSM) by the average
difference between the model and the observed data (MSR)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
445
Q

If the improvement due to fitting the regression model is much greater than the inaccuracy within the model then value of F will be

A

greater than 1 and SPSS calculates exact prob (p-value) of obtaining value of F by change

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
446
Q

What happens if b values are positive in multiple regression?

A

there is a positive relationship between the predictor and the outcome,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
447
Q

What happens if the b value is negative in multiple regression?

A

represents a negative relationship between predictor and outcome variable?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
448
Q

What do the b values in this table tell us what relationships between predictor and outcome variable in multiple regression? (3)

A

Indicating positive relationships so as advertising budget increases, record sales increases (outcome)

plays on ratio increase as do record sales

attractiveness of band increases record sales

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
449
Q

The b-values also tell us, in addition to direction of relationship (pos/neg) , to what degree each in multiple regression

A

predictor affects the outcome if the effects of all other predictors are held constant:

450
Q

B-values tell us to what degree each predictor affects the outcome if the effects of all other predictors held constant in multiple regression

e.g., advertising budget - (3)

A

(b = 0.085):

This value indicates that as advertising budget (x)
increases by one unit, record sales (outcome, y) increase by 0.085 units.

This interpretation is true only if the
effects of attractiveness of the band and airplay are held constant.

451
Q

Standardised versions of b-values are much more easier to interpret as in muliple regression

A

not dependent on the units of measurements of variables

452
Q

The standardised beta values tell us that in multiple regression

A

the number of standard deviations that the outcome will change as a result of one standard deviation change
in the predictor.

453
Q

The standardized beta values are all measured in standard deviation
units and so are directly comparable: therefore, they provide a in MR

A

a better insight into the
‘importance’ of a predictor in the mode

454
Q

If two predictor variables (e.g., advertising budget and airplay) have virtually identical standardised beta values (0.512, and 0.511) it shows that in MR

A

both variables have a comparable degree of importance in the model

455
Q

If we collected 100 samples and in MR calculated CI for b, we are saying that 95% of these CIs of samples would contain the

A

true (pop) value of b

456
Q

A good regression model will have a narrow and small CI interval indicating in MR

A

value of b in this sample is close to the true value of b in the populatio

457
Q

A bad regression model have CI that cross zero indicating that in MR

A

in some samples the predictor has a negative
relationship to the outcome whereas in others it has a positive relationship

458
Q

In image below, which are the two best predictors based on CIs and one that isn’t as (2) in MR

A

two best predictors (advertising and airplay) have very tight confidence intervals indicating that the estimates for the current model are likely to be representative of the true population values

interval for attractiveness is wider (but still does not cross zero) indicating that the parameter for this variable is less representative, but nevertheless significant.

459
Q

If you do part and partial correlations in descriptive box, there will be another coefficients table which looks this in MR like:

A
460
Q

The zero-order correlations are the simple in MR

A

Pearson’s correlation coefficients

461
Q

The partial correlations represent the in MR

A

represent the relationships between each predictor and the outcome variable, controlling for the effects of the other two predictors.

462
Q

The part correlations in MR - (2)

A

represent the relationship between each predictor and the outcome, controlling for the effect that the other two variables have on the outcome.

representing the unique relationship each predictor has with otucome

463
Q

Partial correlations in example is calculated by in MR- (2)

A

unique variance in outcome (ignore all other predictors) explained by predictor divided by variance in outcome not explained by all other predictors

A/A+E

464
Q

Part correlations are calculated by - (2) in MR

A

unique variance in outcome explained by predictor divided by total variance in outcome

A/A+B+C+E

465
Q

If the average VIF is substantially greater than 10 then the MR regression

A

may be biased

466
Q

MR Tolerance below 0.1 indicates a

A

serious problem.

467
Q

Tolerance below 0.2 indicates a in MR

A

a potential problem

468
Q

How to interpret this image in terms of colinearity - VIF and tolerance in MR

A

For our current model the VIF values are all well below 10 and the tolerance statistics all well above 0.2;

therefore, we can safely conclude that there is no collinearity within our data.

469
Q

We can produce casewise diagnostics to see a in MR to see (2)

A

summary of residuals statistics to be examined of extreme cases

To see whether individual scores (cases) influence the modelling of data too much

470
Q

SPSS casewise diagnostics shows cases that have a standardised residuals that are in MR (2)

A

less than -2 or greater than 2

(We expect about 5% of our cases to do tha and 95% to have standardised residuals within about +/- 2.)

471
Q

If we have a sample of 200 then expect about .. to have standardised residuals outside limits in MR

A

10 cases (5% of 200)

472
Q

What does this casewise diagnostic show? - (2) MR

A
  • 99% of cases should lie within ±2.5 so expect 1% of cases lie outside limits
  • From cases listed, clear two cases (1%) lie outside of limits (case, 164 [investigate further has residual 3] and 179) - 1% which isconform to accurate model
473
Q

If there are many more cases we likely have (more than 5% of sample size) in case wise then in MR

A

broken the assumptions of the regression

474
Q

If cases are a large number of standard deviations from the mean, we may want to in casewise diagnostics in MR

A

investigate and potentially remove them because they are ‘outliers’

475
Q

Assumptions we need to check for MR - (8)

A
  • Continous outcome variable and continous or dichotomous predictor variables
  • Independence = all values of outcome variable should come from different participant
  • Non-zero variance as predictors should have some variation in value e.g., variance ≠ 0
  • No outliers
  • No perfect or high collinearity
  • Histogram to check for normality of errors
  • Scatterplot of ZRES against ZPRED to check for linearity and homoscedasticity = looking for random scatter
  • Independent errors (Durbin-Watson)
476
Q

Diagram of assumption of homoscedasticity and linearity of ZRESID againsr ZPRED in MR

A
477
Q

Obvious outliers on a partial plot represent cases that might have in MR

A

undue influence on a predictor’s b coefficient

478
Q

What does this partial plot show? - (2) in MR

A

the partial plot shows the strong positive relationship to album sales.

There are no obvious outliers and the cloud of dots is evenly spaced out around the line, indicating homoscedasticity.

479
Q

What does this plot show in MR(2)

A

the plot again shows a positive relationship to album sales, but the dots show funnelling,

There are no obvious outliers on this plot, but the funnel-shaped cloud indicates a violation of the assumption of homoscedasticity.

480
Q

P plot and histogram of normally distributed in MR

A
481
Q

P plot for skewed distirbution histogram for MR

A
482
Q

What if assumptions for regression is volated? in MR

A

you cannot generalize your findings beyond your sample

483
Q

If residuals show problems
with heteroscedasticity or non-normality then try to in MR

A

transforming the raw data – but
this won’t necessarily affect the residuals!

484
Q

If you have a violation of the linearity assumption then you could see whether you can in MR do l

A

logistic regression instead

485
Q

If R^2 is 0.374 (outcome var in productivity and 3 predictors) then it shows that in MR

A

37.4% of the variance in productivity scores was accounted for by 3 predictor variables

486
Q
  • In ANOVA table, tells whether model is sig improved from baseline model which is in MR
A

if we assumed no relation between predictor variables and outcome variable – flat regression line no association between these variables)

487
Q

This table tells us in terms of standardised beta values that (outcome is productivity in MR)

A

holidays had standardized beta coefficient of 0.031 whereas cake had a much higher standardized beta coefficient of 0.499 which tells us that amount of cake given out much better predictor of productivity than the amount of holidays taken

For pay we have a beta coefficient of 0.323 which tells us that pay was also a pretty good predictor in the model of productivity but slightly less than cake

488
Q

What does this table tells us in terms of signifiance? - (3) in MR

A
  • P value for holidays is 0.891 which is not significant
  • P value for cake is 0.032 is significant
  • P value for pay is 0.012 is significant
489
Q

In ANOVA it is comparing M2 with all its predictor variables with in MR

A

baseline not M1

490
Q

To see if M2 is an improvement of M1 in HR we need to look at … in model summary in MR

A

change statistics

491
Q

What does this change statistic show in terms of M2 and M1 in MR

A

M2 explains an extra 7.5% which is sig

492
Q

In MR, the smaller value of sig, the larger value of t the greater

A

contribution of that predictor.

493
Q

For this output interpret whether predicotrs are sig predictors of record scales and magnitude t statistic on impact of record sales in MR - (2)

A

For this model, the advertising budget (t(196) = 12.26, p < .001), the amount of radio play prior to release (t(196) = 12.12, p < .001) and attractiveness of the band (t(196) =4.55, p < .001) are all significant predictors of record sales.

From the magnitude of the t-statistics we can see that the advertising budget and radio play had a similar impact,
whereas the attractiveness of the band had less impact.

494
Q

What is example of contintous variable?

A

we are talking about a variable with a infinante number of real numbers within a given interval so something like height or age

495
Q

What is an example of dichotomous variable?

A

variable that can only hold two distinct values like male and female

496
Q

If outliers are present in data then impact the

A

line of best fit in MR

497
Q

You would expect that 1% of cases to lie outside the line of best fit so in large sample if you have in MR

A

one or two outliers then could be okay

498
Q

Rule of thumb to check for outliers is to check if there are any data points that in MR

A

are over 3 SD from the mean

499
Q

All residuals should lie within ….. SDs for no outliers /normal amount of outliers in MR

A

-3 and 3 SD

500
Q

Which variables (if any) are highly correlated in MR?

A

Weight, Activity, and the interaction between them are statistically significant

501
Q

What does homoscedasticity and hetrodasticity mean in MR? - (2)

A

Homoscedasticity: similar variance of residuals (errors) across the variable continuum, e.g. equally accurate.

Heteroscedasticity: variance of residuals (errors) differs across the variable continuum, e.g. not equally accurate

502
Q

P plot plots a normal distribution against

A

your distribution

503
Q

Diagram of normal, skewed to left (pos) and skewed to right (neg) of p-plots in MR

A
504
Q

Durbin-Watson test values of 0,2,4 show that… in MR- (3)

A
  • 0 = errors between pairs of obsers are pos correl
  • 2 = independent error
  • 4 = errors between pairs of observs are neg correl
505
Q

A Durbin-Watson statistic between … and … is considered to indicate that the data is not cause for concern = independent errors in MR

A

1.5 and 2.5

506
Q

If R2 and adjusted R2 are similar, it means that your regression model

A

‘generalizes’ to the entire population.

507
Q

If R2 and adjusted R2 are similar, it means that your regression model ‘generalizes’ to the entire population.
Particularly for MR

A

for small N and where results are to be generalized use the adjusted R2

508
Q

3 types of multiple regression - (3)

A
  1. Standard: To assess impact of all predictor variables simultaneously
  2. Hierarchical: To test predictor variables in a specific order based on hypotheses derived from theory
  3. Stepwise: If the goal is accurate statistical prediction from a large number of predictor variables – computer driven
509
Q

Diagram of excluded variables table in SPSS - (3) in MR

A
  • Tells that OCD interpretiotn of intrustrions would have not have a significant impact on model’s ability to predict social anxiety

Beta value of Interpretation of Intrusions is very small, indicating small influence on outcome variable

Beta is the degree of change in the outcome variable for every 1 unit of change in the predictor variable.

510
Q

What is multicollinearity in MR

A

When predictor variables correlate very highly with each other

511
Q

When checking assumption fo regression, what does this graph tell you in MR

A

Normality of residuals

512
Q

Which of the following statements about the t-statistic in regression is not true?

The t-statistic is equal to the regression coefficient divided by its standard deviation

The t-statistic tests whether the regression coefficient, b, is significantly different from 0

The t-statistic provides some idea of how well a predictor predicts the outcome variable

The t-statistic can be used to see whether a predictor variables makes a statistically significant contribution to the regression model

A

The t-statistic is equal to the regression coefficient divided by its standard deviation

513
Q

A consumer researcher was interested in what factors influence people’s fear responses to horror films. She measured gender and how much a person is prone to believe in things that are not real (fantasy proneness). Fear responses were measured too. In this table, what does the value 847.685 represent in MR

A

The residual error in the prediction of fear scores when both gender and fantasy proneness are included as predictors in the model.

514
Q

A psychologist was interested in whether the amount of news people watch predicts how depressed they are. In this table, what does the value 3.030 represent in MR

A

The improvement in the prediction of depression by fitting the model

515
Q

A consumer researcher was interested in what factors influence people’s fear responses to horror films. She measured gender (0 = female, 1 = male) and how much a person is prone to believe in things that are not real (fantasy proneness) on a scale from 0 to 4 (0 = not at all fantasy prone, 4 = very fantasy prone). Fear responses were measured on a scale from 0 (not at all scared) to 15 (the most scared I have ever felt).

Based on the information from model 2 in the table, what is the likely population value of the parameter describing the relationship between gender and fear in MR

A

Somewhere between −3.369 and −0.517

516
Q

A consumer researcher was interested in what factors influence people’s fear responses to horror films. She measured gender (0 = female, 1 = male) and how much a person is prone to believe in things that are not real (fantasy proneness) on a scale from 0 to 4 (0 = not at all fantasy prone, 4 = very fantasy prone). Fear responses were measured on a scale from 0 (not at all scared) to 15 (the most scared I have ever felt).

How much variance (as a percentage) in fear is shared by gender and fantasy proneness in the population in MR

A

13.5%

517
Q

Recent research has shown that lecturers are among the most stressed workers. A researcher wanted to know exactly what it was about being a lecturer that created this stress and subsequent burnout. She recruited 75 lecturers and administered several questionnaires that measured: Burnout (high score = burnt out), Perceived Control (high score = low perceived control), Coping Ability (high score = low ability to cope with stress), Stress from Teaching (high score = teaching creates a lot of stress for the person), Stress from Research (high score = research creates a lot of stress for the person), and Stress from Providing Pastoral Care (high score = providing pastoral care creates a lot of stress for the person). The outcome of interest was burnout, and Cooper’s (1988) model of stress indicates that perceived control and coping style are important predictors of this variable. The remaining predictors were measured to see the unique contribution of different aspects of a lecturer’s work to their burnout.

Which of the predictor variables does not predict burnout in MR

A

Stress from research

518
Q

Using the information from model 3, how would you interpret the beta value for ‘stress from teaching’ in MR

A

As stress from teaching increases by one unit, burnout decreases by 0.36 of a unit.

519
Q

How much variance in burnout does the final model explain for the sample in MR

A

80.3%

520
Q

A psychologist was interested in predicting how depressed people are from the amount of news they watch. Based on the output, do you think the psychologist will end up with a model that can be generalized beyond the sample?

A

No, because the errors show heteroscedasticity.

521
Q

Diagram of no outliers for one assumption of MR

A

Note that you expect 1% of cases to lie outside this area so in a large sample, if you have one or two, that could be ok

522
Q

Example of multiple regression - (3)

A

A record company boss was interested in predicting album sales from advertising.
Data
200 different album releases

Outcome variable:
Sales (CDs and Downloads) in the week after release

Predictor variables
The amount (in £s) spent promoting the album before release
Number of plays on the radio

523
Q

R is the correlation between

A

observed values of the outcome, and the values predicted by the model.

524
Q

Output diagram what does output show in MR? - (2)

A

Difference between no predictors and model 1 (a).
Difference between model 1 (a) and model 2 (b).

Our model 2 is significantly better at predicting the value of the outcome variable than the null model and model 1 (F (2, 197) = 167.2, p<.001) and explains 66% of the variance in our data (R2=.66

525
Q

What does this output show in terms of regression model in MR? - (3)

A

y = 0.09x1 + 3.59x2 + 41.12

For every £1,000 increase in advertising budget there is an increase of 87 record sales (B = 0.09, t = 11.99, p<.001).

For every number of plays on Radio 1 per week there is an increase of 3,589 record sales (B = 3.59, t = 12.51, p<.001).

526
Q

Report R^2, F statistic and p-value to 2DP for overall model - (3)

A

o R squared = 0.09
o F statistic = 22.54
o P value = p < 0.001

527
Q

Report beta and b values for video games, resitrctions and parental aggression to 2DP and p-value in MR

A
528
Q

Which of the following assumptions of homosecdasticity and linearity is correct?

AThere is non-linearity in the data

BThere is heteroscedasticity in the data

CThere is both heteroscedasticity and non-linearity in the data

DThere are no problems with either heteroscedasticity or non-linearity

A

D - data poiints show random pattern

529
Q

Determine the proportion of variance in salary that the number of years spent modelling uniquely explains once the models’ age was taken into account:

Hierarchical regression

A 2.0%

b17.8%

c39.7%

d42.2%

A

A –>The R square change in step 2 was .020,

530
Q

Test for multicollinearity (select tolerance and VIF statistics).

Based on this information, what can you conclude about the suitability of your regression model?

AThe VIF statistic is above 10 and the tolerance statistic is below 0.2, indicating that there is no multicolinearity.

BThe VIF statistic is above 10 and the tolerance statistic is below 0.2, indicating that there is a potential problem withmulticolinearity.

CThe VIF statistic is below 10 and the tolerance statistic is above 0.2, indicating that there is nomulticolinearity.

DThe VIF statistic is below 10 and the tolerance statistic is above 0.2, indicating that there is a potential problem withmulticolinearity.

A

B

531
Q

Example of question using hierarchical regression - (2)

A

A fashion student was interested in factors that predicted the salaries of catwalk models. He collected data from 231 models. For each model he asked how much they earned per day (salary), their age (age), and how many years they had worked as a model (years_modelling).

The student wanted to know if the number of years spent modelling predicted the models’ salary after the models’ age was taken into account.

532
Q

The following graph shows:

A. Regression assumptions met

B. Non-linearity = could indicate curve

C. Hetrodasticity + Non-linearity

D. hetrodasticity

A

A

533
Q

A consumer researcher was interested in what factors influence people’s fear responses to horror films. She measured gender (0 = female, 1 = male) and how much a person is prone to believe in things that are not real (fantasy proneness) on a scale from 0 to 4 (0 = not at all fantasy prone, 4 = very fantasy prone). Fear responses were measured on a scale from 0 (not at all scared) to 15 (the most scared I have ever felt). What is the likely population value of the parameter describing the relationship between gender and fear?

A

Somewhere between 3.369 and 0.517

534
Q

What are the 3 types of t-tests? - (3)

A
  1. One-samples t-test
  2. Paired t-test
  3. Independent t-test
535
Q

Whats a one-sample t-test?

A

Compares the mean of the sample data to a known value

536
Q

What is the assumptions of one-sample t-test? - (4)

A
  • DV = Continous (interval or ratio)
  • Independent scores (no relation between scores on test variable)
  • Normal distribution via frequency histogram (normal shape) and Q-plot (straight line) and non significant Shaprio Wilk
  • Homogenity of variances
537
Q

Example of one-sample t-test RQ - (2)

A

Is the average IQ of Psychology students higher than that of the general population (100)?

A particular factory’s machines are supposed to fill bottles with 150 millilitres of product. A plant manager wants to test a random sample of bottles to ensure that the machines are not under- or over-filling the bottles.

538
Q

What is the assumptions of independent samples t-tests (listing all of them) - (7)

A
  1. Independence. – no relationship between the groups
  2. Normal distribution via frequency histogram (normal shape) and Q-plot (straight line) and non significant Shaprio Wilk
  3. Equal variances
  4. Homogeneity of variances (i.e., variances approximately equal across groups) via non significant Levene’s test
  5. DV = Interval or continuous
  6. IV = Categorical
  7. No significant outliers
539
Q

What is an RQ example of independent samples t-tesT?

A

Do dog owners in the country spend more time walking their
dogs than dog owners in the city?

540
Q

What is the assumptions of paired t-test? (listing all) - 3

A

DV is continuous

Related samples: The subjects in each sample, or group, are the same. This means that the subjects in the first group are also in the second group

Normal distribution via frequency histogram (normal shape) and Q-plot (straight line) and non significant Shaprio Wilk

541
Q

What is an example of RQ of paired t-test?

A

Do cats learn more tricks when given food or praise as positive feedback?

542
Q

What is the decision framework for choosing a paired-sample (dependent) t-test? - (5)

A
  1. What sort of measurement = continous
  2. How many predictor variables = one
  3. What type of predictor variables = categorical
  4. How many levels of categorical predictor = two
  5. Same or different participants for each predictor level = same
543
Q

What is the decision framework for choosing independent-t-test? (5)

A
  1. What sort of measurement = continous
  2. How many predictor variables = one
  3. What type of predictor variables = categorical
  4. How many levels of categorical predictor = two
  5. Same or different participants for each predictor level = different
544
Q

If we are comparing differences between means of two groups in independent/paired t-test then all we are doing is

A

predicting an outcome based on membership of two groups

545
Q

Indepdnent and paired t-tests can fit into an ideal of a

A

linear model

546
Q

The t-distributed is defined by its

A

degrees of freedom - related to the sample size.

547
Q

The t distribution has heavier tails for - (2)

A

lower degrees of freedom (small N studies)

increased uncertainty and a higher likelihood of observing extreme values than large N studies with less heavy tails as t distribution goes to normal

548
Q

Independent and Paired T-tests have one predictor (X) variable with 2 levels and only …. outcome variable (Y)

A

one

549
Q

When is an independent-means t-test used?

A

When 2 experimental conditions and different participants are assigned to each conditiont

550
Q

What is independent-means t-test sometimes called as well?

A

independent-samples t-test

551
Q

When is a dependent-means t-test used?

A

Used when there are 2 experimental conditions and same participants took part in both conditions of the experiment

552
Q

What is dependent-means t-test sometimes referred to?

A

Matched pairs or paired samples t-test

553
Q

For independent and paired t-tests we compare between the sample means that we collected to the difference between sample means that we would expect if

A

there was no effect (i.e., null hypothesis was true)

554
Q

Formula of calculating t- test statistic (form depend on whether same or different participants used in each experimental condition) in independent/paired

A
555
Q

Formula of calculating t-statistic shows obtaining t-test statistic by diving the model/effect by the in independent /apried

A

error in the model

556
Q

Expected difference in calculating t-test statistic in most cases is

A

0 - expect differences between sample group means we colelcted to be different to 0

557
Q

If observed difference between sample means get larger in t-tests then more confident we become that

A

null hypothesis is rejected and two sample means differ because of experimental manipulation

558
Q

Both independent t-test and paired t-test are … tests based on normal distribution

A

parametric tests

559
Q

Since independent and paired t-tests are parametric tests they assume that the - (2)

A
  • Sampling distribution is normally distributed - in paired it means sampling distribution of differences of scores is normal not the socres itself!
  • Data measured at least interval level
560
Q

Since independent-tests used to test different groups of people it also assumes - (2)

A
  • Variances in populations are roughly equal (homegenity of variance) = Leven’s test
  • Scores are independent since they come from different people
561
Q

Diagram of equation of calculating t-statistic from paired t-test and explain - (2)

A
  • Compares mean differences betwen our samples (–D) to the differences we would expect to find between population means (uD) which is divided by standard error of differences (sD / square root N)
  • If H0 is ture, then expect no difference between population means hence uD = 0
562
Q

A small standard error of differences tells us that in paired-t-test

A

pairs of samples from a population have similar means to population

563
Q

A large standard error of differences tells us that in paired t-test - (2)

A

that sample means can deviate quite a lot from the populatio mean and

sampling distribution of differences is more spread out

564
Q

The average difference between person’s socre in condition 1 and condition 2 -(¯D) in paired t-test is an indicator of

A

systematic variation in the data (represents experimental effect)

565
Q

If average differences (–D) between our samples is large and standard error of differences is small in paired-t test then we can be confident that

A

the difference we observed in our sample is not a chance result and caused by experimental manipulation

566
Q

How do we normally calculate the standard error?

A

SD divided by square root of sample size

567
Q

How to calcuate the standard error of differences in paired-test?(σ –D)

A

Standard deviation of differences divided by square root of sample size

568
Q

the t-statistic in paired t-test is

A

ratio of systematic variation in experiment (average difference D) and unsystematic variation (standard erro of differences)

569
Q

When would we expect t statistic greater than 1 in paired-t-test equation?

A

If the experimental manipulation creates any kind of effect,

570
Q

When would we expect t statistic less than 1 in paired t-test equation?

A

If the experimental manipulation is unsuccessful then we might expect the variation caused by individual differences to be much greater than that caused by the
experiment

571
Q

In pairered and generally independent t-tests we can compare the obtainee value of t against thmaximum value we would expect to get by chance alone in t distribution with same DFs and if value we obtain exceeds the

A

critical value then conflict if reflects an effect in our IV

572
Q

What does this paired samples correlation show?

A

people doing well in first exam likely doing well in second exam regardless of condition they are in and significantly correlated (r= 0.664)

573
Q

What does this SPSS output show? = paired t- test

A

t(19) = 2.72, p = 0.012

574
Q

What does negative t-value mean?

paired t-test.

A

First condition had smaller mean than second condition

575
Q

What does 95% confiderence interval of difference mean in SPSS output of paired t-test?- (3)

A
  • 95% of the samples (e.g., if we had 100 samples then 95 of those samples..) the constructucted CIs contain true value (population) of the mean difference
  • CIs tell us boundaries within which true mean difference is likely to lie
  • The true value of mean difference is unlikely to be 0 if Cis does not contain 0
576
Q

How to calculate effect size for independent and paired t-tests?

A

Using cohen’s D

577
Q

Diagram of calculating Cohen’s D Statistic for sleep vs no sleep for paired

A

Minus big mean from small mean divided by smallest SD (control group)

578
Q

What does Cohen’s d of 0.20 represent

A

difference between groups is a 1/5 of SD

579
Q

Diagram of writing up paired t-test result

A
580
Q

To calculate effect size for independent and paired t-tests, beside Cohen’s D, we can also

A

calculate effect size r (above 0.50 is large effect) by converting t-value to r-value

581
Q

With independent t-test there are two different equations that can be used depending on whether the samples

A

contain an equal number of people

582
Q

With independent t-test since different participants participate in different condition, the pairs of scores will differ not just of experimental manipulation but also because of

A

other sources of variance (such as individual differences between participants’ motivation, IQ etc..)

583
Q

With dependent t-test we look at differences between pairs of scores because

A

scores came from same participant and so individual differences were eliminated

584
Q

Equation of independent t-test of equal N sizes for each condition

A
585
Q

Equation of independent t-test of equal N sizes becomes like the final form since - (3)

A
  • We are looking at differences between the overall means of 2 samples and compare with differences we would expect to get between means of 2 populations from which sampels come from
  • If H0 is true, samples drawn from same population
  • Therefore under H0, u1 = u2 therefore u1 - u2 = 0
586
Q

Equation of independent t-test in numbers for equal N sizes

A
587
Q

We use variance of sum law to obtain the estimate of standard error for each … in independent t-test equation for equal N sizes

A

sample group

588
Q

What does variance sum of law state?

A

variance of the sampling distribution is equal to the sum of the variances of the two populations from which the samples
were taken

589
Q

This independent t-test standard error formula combines the

A

standard error for two samples

590
Q

In independent t-test when we want to compare two groups that contain different number of participants then equation … is not appropriate

A
591
Q

For comparing two groups with unequal number of participants in independent t-test then we use the

A

pooled variance estimate t-test

592
Q

The pooled variance estimate t-test is used which takes into account of the

A

differnece in sample size by weighting the variance of each sample

593
Q

Formula of pooled variance estimate t-test - (2)

A

Each variance of sample is multipled by its DF and added together and divided by the sum of weights (sum of two DFs)

Larger samples better than small ones as close to population

594
Q

In formula of pooled variance estimate t-test it weights the variance of each sample by the

A

number of degrees of freedom (N-1)

595
Q

As with dependent t-test we compare obtained value of t in independent sample against the

A

maximum value we would expect to get by chance alone in t distribution with same DFs

596
Q

What does this output show? - in independent t-test - (2)

A

Sleep condition scored an average exam score of 66.200 and no sleep condition earned an average of 58.73

Effect size (Cohen’s D) = Mean of sleep minus mean of no sleep divided by standard deviation of sleep (control grp) = 66.20-58/73/7.12

597
Q

In independent samples t-test we check for Levene’s test for quality of variances which determine whether

A

we got equal variance across the groups or whether the variances are unequal

598
Q

In independent t-test, Levene’s test we are looking for a non-significant p-value which shows that

A

no statistically significant difference in variances between the two groups - report results from equal variances assumed

599
Q

In independent t-test if Levene’s test was significant then it means that

A

variances between the 2 groups are different and they are statistically significantly different - report data from equal variances not assumed

600
Q

What does this output show in independent t-test? - (2)

A
  • Levene’s test is not significant (p = 0.970) so no stats sig differences in variance between two groups
  • t(28) = 2.87, p = 0.008
601
Q

Diagram of reporting independent t-test

A
602
Q

Paired vs independent t-tests - who has better power?

A

Paired t-t ests

603
Q

Since paired-t-tests use same participants across conditions the … is reduced dramatically compared to independent t-test

A

unsystematic variance

604
Q

The non-parametric counterpart of dependent t-test is called

A

Wilcoxon signed rank test

605
Q

The non-parametric tests of the independent t-test is

A

Wilcoxon rank sum test and Mank Whitney test

606
Q

What does this SPSS output of independent-test of Levene’s show

A

homogeneity of variance as assessed by Levene’s Test for Equality of Variances (F = 1.58, p = .219)

607
Q

Cohen’s d for diet was 4.25

Is this a:
Small effect
Medium effect
Large effect

A

Large effect

608
Q

The probability of a value of t occurring yields the p value for the difference between the means occurring by

A

chance

609
Q

Although there are 2 ways to compute effect, use

A

Cohen D

610
Q

Another example of two samples independent t-test scenario

RQ

Sample

DV

Hypothesis

Test

Sig

  • (6)
A

Research question: Which of the two diet formulas is better for puppies?

Sample: 15 were randomly assigned to each of the two diets (A and B).

Dependent variable: Average daily weight gain (ADG, g/day) between 12 to 28 weeks of age.

Hypotheses:
Ho: µA = µB
Ha: µA ≠ µB.

Statistical Test: Two samples

independent t-test
Significance level: .05

611
Q

We can check if there is no outliers in independent t-test by looking at

A

boxplots - no outlier here

612
Q

To check normality of distribution for both independent groups for two-samples independent t-test, we can use..

A

histogram, q-qplot and tests of normality

613
Q

Checking normality for independent, is it - (3)

Research question: Which of the two diet formulas is better for puppies?

Dependent variable: Average daily weight gain (ADG, g/day)

A

We don’t have sig values for either group in the test of normality, histogram and plots look normal

So we have normality of distribution for both independent groups

Inspection of Q-Q Plots and the non-significant Shapiro-Wilk tests (p > .05) indicate that the ADG is normally distributed for both groups

614
Q

For checking homogeneity of variances in independent/paired we use

A

levene’s test

615
Q

Checking homogenity of variance in this two-sample independent t-test, what does it show?

Research question: Which of the two diet formulas is better for puppies?

Dependent variable: Average daily weight gain (ADG, g/day)

A

was homogeneity of variance as assessed by Levene’s Test for Equality of Variances (F = 1.58, p = .219)

616
Q

What does results of two-sample independent results show?

Research question: Which of the two diet formulas is better for puppies?

Dependent variable: Average daily weight gain (ADG, g/day)

A

This study found that puppies in diet B had statistically significantly higher average daily weight gain (89.29 ± 9.93 g/day) between 12 and 28 weeks of age compared to puppies in diet A (60.20 ± 6.85 g/day), t(27)= -9.24, p < .001.

617
Q

In Cohen’s D theoretically 3 SDs can be used - (3) which make very little difference

A
  1. Pooled SD (over conditions)
    2.Averaged SD
  2. Control group SD
618
Q

To calculate Cohen D for independent/paired t-test we need to use

A

control group SD

619
Q

How to calculate Cohen’s D for independent two samples t test for this group?

Research question: Which of the two diet formulas is better for puppies?

Dependent variable: Average daily weight gain (ADG, g/day) - (2)

A

d = (89.29 - 60.20) / 6.85

d = 4.25

620
Q

Cohen’s D guidelines for small, medium large - (3)

A

d = 0.2 be considered a ‘small’ effect size,

d = 0.5 represents a ‘medium’ effect size

d = 0.8 a ‘large’ effect size

621
Q

What does ANOVA stand for?

A

Analysis of Variance

622
Q

What

What is the decision tree for choosing a one-way ANOVA? - (5)

A

Q: What sort of measurement? A: Continuous
Q:How many predictor variables? A: One
Q: What type of predictor variable? A: Categorical
Q: How many levels of the categorical predictor? A: More than two
Q: Same or Different participants for each predictor level? A: Different

623
Q

When does ANOVA be used?

A

if you are comparing more than 2 groups in IV

624
Q

Example of ANOVA RQ

A

Which is the fastest animal in a maze experiment - cats, dogs or rats?

625
Q

We can’t do three separate t-tests for example what is the fastest animal in a maze experiment - cats, dogs or rats as - (2)

A

Doing separate t-tests inflates the type I error (false positive - e.g., pregnant man)

The repetition of the multiple tests adds multiple chances of error, which may result in a larger α error level than the pre-set α level - Family wise error

626
Q

What is familywise or experimentwise error rate?

A

This error rate across statistical tests conducted on the same experimental data

627
Q

Family wise error is related to

A

type 1 error

628
Q

What is the alpha level probability

A

probability of making a wrong decision in accepting the alternate hypothesis = type 1 error

629
Q

If we conduct 3 separate t-tests to test the comparison of which is the fastest animal in experiment - cats, dogs or rats with alpha level of 0.05 - (4)

A
  • 5% of type 1 error of falsely rejecting H0
  • Probability of no. of Type 1 errors is 95% for a single test
  • However, for multiple tests the probability of type 1 error decreases as 3 tests together => 0.950.950.95 = 0.857
  • This means probability of a type 1 error increases: 1- 0.857 = 0.143 (14.3% of not making a type 1 error)
630
Q

Much like model for t-tests we can write a general linear model for

A

ANOVA - 3 levels of categorical variable with dummy variables

631
Q

When we perform a t-test, we test the hypothesis that the two samples have the same

A

mean

632
Q

ANOVA tells us whether three or more means are the same so tests H0 that

A

all group means are equal

633
Q

An ANOVA produces an

A

F statistic or F ratio

634
Q

The F ratio produced in ANOVA is similar to t-statistic in a way that it compares the

A

amount of systematic variance in data to the amount of unsystematic variance i.e., ratio of model to its error

635
Q

ANOVA is an omnibus test which means it tests for and tells us - (2)

A

overall experimental effect

tells whether experimental manipulation was successful

636
Q

An ANOVA is omnibus test and its F ratio does not provide specific informaiton about which

A

groups were affected due to experimental manipulation

637
Q

Just like t-test can be represented by linear regression equation, ANOVA can be represented by a

A

multiple regression equation for three means and models acocunt for 3 levels of categorical variable with dummy variables

638
Q

As compared to independent samples t-test that compares means of two groups, one-way ANOVA compares means of

A

3 or more independent groups

639
Q

In one-way ANOVA we use … … to test assumption of equal variances across groups

A

Levene’s test

640
Q

What does this one-way ANOVA output show?

A

Leven’s test is non-significant so equal variances are assumed

641
Q

What does this SPSS output show in one-way ANOVA?

A

F(2,42) = 5.94, p = 0.005, eta-squared = 0.22

642
Q

How is effect size (eta-squared) calculated in one-way ANOVA?

A

Between groups sum of squares divided by total sum of squares

643
Q

What is the eta-squared/effect size for this SPSS output and what does this value mean? - (2)

A

830.207/3763.632 = 0.22
22% of the variance in exam scores is accounted for by the model

644
Q

Interpreting eta-squared, what does 0.01, 0.06 and 0.14 eta-sqaured in one way ANOVA means? - (3)

A
  1. 0.01 = small effect
  2. 0.06 = medium effect
  3. 0.14 = large effect
645
Q

What happens if the Levene’s test is significant in the one-way ANOVA?

A

then use statistics in Welch or Brown-Forsythe test

646
Q

The Welch or Brown-Forsythe test make adjustements to DF which affects

in one way ANOVA if Levene’s test is sig

A

statistics you get and affect if p value is sig or not

647
Q

What does this post-hoc table of Bonferroni tests show in one-way ANOVA ? - (3)

A
  • Full sleep vs partial sleep, p = 1.00, not sig
    • Full sleep vs no sleep , p = 0.007 so sig
    • Partial sleep vs no sleep = p = 0.032 so sig
648
Q

Diagram of example of grand mean

A

Mean of all scores regardless pp’s condition

649
Q

What are the total sum of squares (SST) in one-way ANOVA?

A

difference of the participant’s score from the grand mean squared and summed over all participants

650
Q

What is model sum of squares (SSM) in one-way ANOVA?

A

difference of the model score from the grand mean squared and summed over all participants

651
Q

What is residual sum of squares (SSR) in one-way ANOVA?

A

difference of the participant’s score from the model score squared and summed over all participants

652
Q

The residuals sum of squares (SSR) tells us how much of the variation cannot be

A

explained by the model and amount of variation caused by extraneous factors

653
Q

We divide each sum of squares by its

A

DF to calculate them

654
Q

For SST its DF we divide by is in one-way ANOVA

A

N-1

655
Q

For SSM its DF we divide by is one-way ANOVA so

A

number of group (parameters), k,

656
Q

For SSM if we have this design then… then its DF will be in one way ANOVA

A

3-1 = 2

657
Q

For SSR we divivde by its DF to calculate which will be the in one way ANOVA

A

total sample size, N, minus the number of groups, k

658
Q

Formulas of dividing each sum of squares by its DF to calculate it in one way ANOVA- (3)

A
  • MST = SST (N-1)
  • MSR = SSR (N-k)
  • MSM = SSM/k
659
Q

SSM tells us the total variation that the

A

exp manipulation explains

660
Q

What does MSM represent?

A

average amount of variation explained by the model (e.g. the systematic variation),

661
Q

What does MSR represent?

A

average amount of variation explained by extraneous variables (the unsystematic variation).

662
Q

The F ratio in one-way ANOVA can be calculated by

A
663
Q

If F ratio in one-way ANOVA is less than 1 then it represents a

A

non-significant effect

664
Q

Why F less than 1 in one-way ANOVA represents a non-significant effect?

A

F ratio is less than 1 means that MSR is greater than MSM = more unsystematic than systematic

665
Q

If F is greater than 1 in one-way ANOVA then shows likelhood … but doesn’t tell us - (2)

A

indicates that experimental manipulation had some effect above and beyond effect of individual differences in performance

Does not tell us whether F-ratio is large enough to not be a chance result

666
Q

When F statistic is large in one-way ANOVA then it tells us that the

A

MSM is greater than MSR

667
Q

To discover if F statistic is large enough not to be a chance result in one-way ANOVA then

A

compare the obtained value of F against the maximum value we would expect to get by chance if the group means were equal in an F-distribution with the same degrees
of freedom

668
Q

High values of F are rare by in one way ANOVA are rare - (3)

A

by chance

. Low degrees of freedom result in long tails of the distribution, so much like other statistics

large values of F are more common to crop up by chance in studies with low numbers of participants.

669
Q

The F-ratio tells us in one-way ANOVA whether model fitted to data accounts for more variation thane extraneous and does not tell us where

A

differences between groups lie

670
Q

If F-ratio in one-way ANOVA is large enough to be statistically significant then we know

A

that one or more of the differences between means is statistically significant (e.g. either b2 or b1 i statistically significant)

671
Q

It is necessary after conducting an one-way ANOVA to carry out further analysis to find out

A

which groups differ

672
Q

The power of F statistic is relatively unaffected by

A

non-normality

673
Q

when group sizes are not equal the accuracy of F is

A

affected by skew, and non-normality also affects the power of F in quite unpredictable ways

674
Q

When group sizes are equal, the F statistic can be quite robust to

A

violations of normality

675
Q

What tests do you do after performing a one-way ANOVA and finding significant F test? - (2)

A
  • Planned contrasts
  • Post-hoc tests
676
Q

What do post-hoc tests do? - (2)

A
  • compare all pairwise differences in mean
  • Used if no specific hypotheses concerning differences has been made
677
Q

What is the issue with post-hoc tests?

A
  • because every pairwise combination is considered the type 1 error rate increases, so normally the type 1 error rate is reduced by modifying the critical value of p
678
Q

Post-hoc tests are like two or one tailed hypothesis?

A

two-tailed

679
Q

Planned contrasts are like one or two-tailed hypothesos?

A

One-tailed hypothesis

680
Q

What is the most common modification of the critical value for p in post-hoc in one-way ANOVA?

A

Bonferroni correction, which divides the standard critical value of p=0.05 by the number of pairwise comparisons performed

681
Q

Planned contrasts are used to investigate a specific

A

hypothesis

682
Q

Planned contrasts do not test for every

A

pairwise difference so are not penalized as heavily as post hoc tests that do test for every difference

683
Q

With planned contrasts test you dervivie the hypotheses before the

A

data is collected

684
Q

In planned contrasts when one condition is used it is

A

never used again

685
Q

In planned contrasts the number of independent contrasts you can make can be defined with one way ANOVA

A

k (number of groups) minus 1

686
Q

How does planned contrasts work in SPSS?

A

Coefficients add to 0 for each contrast (-2 + 1 +1) and once group used alone in contrast then enxt contrasts set coefficient to 0 (e.g., -2 to 0)|

687
Q

Polynominal contrasts can also look at more complex trends other than linear such as in one way ANOVA?

A

quadratic, cubic and quartic

688
Q

The Bonferroni post-hoc ensures that the type 1 error is below in one-way ANOVA?

A

0.05

689
Q

With Bonferroni correction it reduces type 1 (being conserative in type 1 error for each comparison) it also in one way ANOVA?

A

lacks statistical power (probability of type II error will be high [ false negative]) so increasing chance of missing a genuine difference in data

690
Q

What post hoc-tests to use if you have equal sample sizes and confident that your group variances are similar? in one way ANOVA

A

Use REGWQ or Tukey as good power and tight control over Type 1 error rate

691
Q

What post hoc tests to use if your sample sizes are slightly different in one way ANOVA?

A

Gabriel’s procedure because it has greater power,

692
Q

What post-hoc tests to use if your sample sizes are very different? ine one way ANOVA?

A

if sample sizes are very different use
Hochberg’s GT2

693
Q

What post-hoc test to run if Levene’s test of homeogenity of variance is significant in one way ANOVA?

A

Games-Howell

694
Q

**

What post=hoc test to use if you want gurantee control over type 1 errror rate in one wau ANOVA?

A

Bonferroni

695
Q

What does this ANOVA error line graph show? - (2)

A
  • Linear trend as dose of Viagra increases so does mean level of libido
  • Error bars overlap indicating no between group differences
696
Q

What does the within groups gives deails of in ANOVA table?

A

SSR (unsystematci variation)

697
Q

The between groups label in ANOVA table tells us

A

SSM (systematic variation)

698
Q

What does this ANOVA table demonstrate? - (2)

A
  • Linear trend is significant (p = 0.008)
  • Quadratic trend is not significant (p = 0.612)
699
Q

When we do planned contrasts we arrange the weights in such that we compare any group with a positive weight

A

with a negative weight

700
Q

What does this output show if we conduct two planned comparisons of:

one to test whether the control group was different to the two groups which received Viagra, and one to see
whether the two doses of Viagra made a difference to libido

  • (2)
A

the table of weights shows that contrast 1 compares the placebo group against the two experimental groups,

contrast 2 compares the low-dose group to the high-dose group

701
Q

What does this table show if levene’s test is non significant =equal variances assumed

To test hypothesis that experimental groups would increase libido above the levels seen in the placebo group (one-tailed)

To test another hypothesis that a high dose of Viagra would increase libido significantly more than a low dose

one-way ANOVA

  • (3)
A

Signifiance value given in table is two-tailed and since hypothesis one-tail we divide by 2

for contrast 1, we can say that taking Viagra significantly increased libido compared to the control group (p = .0029/2 = 0.0145)

. The significance of contrast 2 tells us that a high dose of Viagra increased libido significantly more than a low dose (p(one-tailed) = .065/2 = .0325)

702
Q

If making a few pairwise comparisons and equal umber of pps in each condition then … if making a lot then use. in one way ANOVA - (2).

A

Bonferroni
Tukey

703
Q

Assumptions of ANOVA - (5)

A
  • Independence of data
  • DV is continuous; IV categorical (3 groups)
  • No significant outliers;
  • DV approximately normally distributed for each category of the IV
  • Homogenity of variance = Levene’s test not significant
704
Q

ANOVA compares many means without increasing the chance of

A

type 1 error

705
Q

In one-way ANOVA, we partiton the total variance into

A

IV and DV

706
Q

An independent t-test is used to test for:

A Differences between means of groups containing different participants when the sampling distribution is normal, the groups have equal variances and data are at least interval.

B Differences between means of groups containing different participants when the data are not normally distributed or have unequal variances.

C Differences between means of groups containing the same participants when the data are normally distributed, have equal variances and data are at least interval.

D Differences between means of groups containing the same participants when the sampling distribution is not normally distributed and the data do not have unequal variances.

A

A differences between means of groups containing different participants when sampling distribution is normal and the groups have equal variances and data are at least interva

707
Q

If you use a piared samples t-test

A The same participants take part in both experimental conditions.

BThere ought to be less unsystematic variance compared to the independent t-test.

C Other things being equal, you do not need as many participants as you would for an independent samples design.

D All of these are correct.

A

D All of these are correct

708
Q

Which of the following statements about the t distribution is correct?

A It is skewed

BIn small samples it is narrower than the normal distribution

CAs the degrees of freedom increase, the distribution becomes closer to normal

DIt follows an exponential curve

A

C As the DF increase, the distribution becomes closer to normal

709
Q

Which of the following sentences is an accurate description of the standard error?

AIt is the same as the standard deviation

BIt is the observed difference between sample means minus the expected difference between population means (if the null hypothesis is true)

CIt is the standard deviation of the sampling distribution of a statistic

D It is the standard deviation squared

A

CIt is the standard deviation of the sampling distribution of a statistic

710
Q

A psychologist was interested in whether there was a gender difference in the use of email. She hypothesized that because women are generally better communicators than men, they would spend longer using email than their male counterparts. To test this hypothesis, the researcher sat by the computers in her research methods laboratory and when someone started using email, she noted whether they were male or female and then timed how long they spent using email (in minutes). Based on the output, what should she report?

(NOTE: Check for the assumption of equality of variances).

A Females spent significantly longer using email than males, t(14) = –1.90, p = .079

BFemales and males did not significantly differ in the time spent using email,t(7.18) = –1.90,p= .099

CFemales and males did not significantly differ in the time spent using email, t(7.18) = –1.90, p < .003

DFemales and males did not significantly differ in the time spent using email, t(14) = –1.90, p = .079

A

BFemales and males did not significantly differ in the time spent using email,t(7.18) = –1.90,p= .099

711
Q

Other things being equal, compared to the paired-samples (or dependent)t-test, the independentt-test:

A Has more power to find an effect.

BHas the same amount of power, the data are just collected differently.

CHas less power to find an effect.

D Is less robust.

A

CHas less power to find an effect.

712
Q

Differences between group means can be characterized as a regression (linear) model if:

AThe outcome variable is categorical.

BThe groups have equal sample size.

CThe experimental groups are represented by a binary variable (i.e. code 1 and 0).

DThe difference between group means cannot be characterized as a llinear model, they must be analyzed as an independent t-test.

A

The experimental groups are represented by a binary variable (i.e. code 1 and 0)

713
Q

An experiment was done to look at whether different relaxation techniques could predict sleep quality better than nothing. A sample of 400 participants were randomly allocated to one of four groups: massage, hot bath, reading or nothing. For one month each participant received one of these relaxation techniques for 30 minutes before going to bed each night. A special device was attached to the participant’s wrist that recorded their quality of sleep, providing them with a score out of 100. The outcome was the average quality of sleep score over the course of the month.

Which test could we use to analyse these data?

A Regression only

B ANOVA only

C Regression or ANOVA

D Chi-square

A

C (multiple) Regression or ANOVA (independent) as regression and ANOVA is the same

Did not mention the hypothesis of prediction or it would be regression

Chi-square only used when you have one categorical predictor and outcome is categorical

714
Q

A researcher testing the effects of two treatments for anxiety computed a 95% confidence interval for the difference between the mean of treatment 1 and the mean of treatment 2. If this confidence interval includes the value of zero, then she cannot conclude that there is a significant difference in the treatment means: true or false.

TRUE OR FALSE

A

TRUE

715
Q

The student welfare office was interested in trying to enhance students’ exam performance by investigating the effects of various interventions. They took five groups of students before their statistics exams and gave them one of five interventions: (1) a control group just sat in a room contemplating the task ahead; (2) the second group had a yoga class to relax them; (3) the third group were told they would get monetary rewards contingent upon the grade they received in the exam; (4) the fourth group were given beta-blockers to calm their nerves; and (5) the fifth group were encouraged to sit around winding each other up about how much revision they had/hadn’t done (a bit like what usually happens). The final percentage obtained in the exam was the dependent variable. Using the critical values for F, how would you report the result in the table below?

AType of intervention did not have a significant effect on levels of exam performance, F(4, 29) = 12.43, p > .05.

BType of intervention had a significant effect on levels of exam performance, F(4, 29) = 12.43, p < .01.

CType of intervention did not have a significant effect on levels of exam performance, F(4, 33) = 12.43, p > .01.

DType of intervention had a significant effect on levels of exam performance, F(4, 33) = 12.43, p < .01.

A

Type of intervention had a significant effect on levels of exam performance, F(4, 29) = 12.43, p < .01.

716
Q

Imagine you compare the effectiveness of four different types of stimulant to keep you awake while revising statistics using a one-way ANOVA. The null hypothesis would be that all four treatments have the same effect on the mean time kept awake. How would you interpret the alternative hypothesis?

A. All four stimulants have different effects on the mean time spent awake

B, All stimulants will increase mean time spent awake compared to taking nothing

C. At least two of the stimulants will have different effects on the mean time spent awake

D, None of the above

A

C. At least two of the stimulants will have different effects on the mean time spent awake

717
Q

When the between-groups variance is a lot larger than the within-groups variance, the F-value is ____ and the likelihood of such a result occurring because of sampling error is _____

A small; high

B small; low

C. large; high

D. large; low

A

D. large; low

718
Q

Subsequent to obtaining a significant result from an exploratory one-way independent ANOVA, a researcher decided to conduct three post hoc t-tests to investigate where the differences between groups lie.

Which of the following statements is correct?

A. The researcher should accept as statistically significant tests with a probability value of less than 0.016 to avoid making a Type I error

B. The researcher should have conducted orthogonal contrasts instead of t-tests to avoid making a Type I error

C. This is the wrong method to use. The researcher did not make any predictions about which groups will differ before running the experiment, therefore contrasts and post hoc tests cannot be used

D. None of these options are correct

A

The researcher should accept as statistically significant tests with a probability value of less than 0.016 to avoid making a Type I error

719
Q

A psychologist was looking at the effects of an intervention on depression levels. Three groups were used: waiting list control, treatment and post-treatment (a group who had had the treatment 6 months before). The SPSS output is below. Based on this output, what should the researcher report?

A. The treatment groups had a significant effect on depression levels,F(2, 45) = 5.11.

B. The treatment groups did not have a significant effect on the change in depression levels,F(2, 35.10) = 5.11.

C. The treatment groups did not have a significant effect on depression levels,F(2, 26.44) = 4.35.

D. The treatment groups had a significant effect on the depression levels,F(2, 26.44) = 4.35.

A

D. The treatment groups had a significant effect on the depression levels,F(2, 26.44) = 4.35.

720
Q

Imagine we conduct a one-way independent ANOVA with four levels on our independent variable and obtain a significant result. Given that we had equal sample sizes, we did not make any predictions about which groups would differ before the experiment and we want guaranteed control over the Type I error rate, which would be the best test to investigate which groups differ?

A. Orthogonal contrasts

B. Helmert

C. Bonferroni

D. Hochberg’s GT2

A

C. Bonferroni

721
Q

The student welfare office was interested in trying to enhance students’ exam performance by investigating the effects of various interventions.

They took five groups of students before their statistics exams and gave them one of five interventions: (1) a control group just sat in a room contemplating the task ahead (Control); (2) the second group had a yoga class to relax them (Yoga); (3) the third group were told they would get monetary rewards contingent upon the grade they received in the exam (Bribes); (4) the fourth group were given beta-blockers to calm their nerves (Beta-Blockers); and (5) the fifth group were encouraged to sit around winding each other up about how much revision they had/hadn’t done (You’re all going to fail).

The student welfare office made four predictions: (1) all interventions should be different from the control; (2) yoga, bribery and beta-blockers should lead to higher exam scores than panic; (3) yoga and bribery should have different effects than the beta-blocker drugs; and (4) yoga and bribery should also differ.

Which of the following planned contrasts (with the appropriate group codings) are correct to test these hypotheses?

ANSWER 1
ANSWER 2
ANSWER 3
ANSWER 4

A

ANSWER 1 - sum of all weights should be 0

722
Q

Deciding what post hoc tests to run

A
723
Q

Example of RQ for one way ANOVA - (3)

A

Is there a statistically significant difference in Frisbee throwing distance with respect to education status

IV = Education with 3 levels = high school, graduate, postgrad

DV = Frisbee throwing distance

724
Q

What does this one-way ANOVA output show?

Research question: Is there a statistically significant difference in Frisbee throwing distance with respect to education status?

Variables:
IV - Education, which has three levels:
High School, Graduate and PostGrad;
DV - Frisbee Throwing Distance

A

There was homogeneity of variance as assessed by Levene’s Test for Equality of Variances (F (2,47) = 1.94, p = .155)

725
Q

What does the results of one-way ANOVA show?

Research question: Is there a statistically significant difference in Frisbee throwing distance with respect to education status?

Variables:
IV - Education, which has three levels:
High School, Graduate and PostGrad;
DV - Frisbee Throwing Distance

A

There was a statistically significant difference between groups as demonstrated by one-way ANOVA (F(2, 47) = 3.50, p = .038).

726
Q

What does the results of one-way ANOVA show? –> post hoc

Research question: Is there a statistically significant difference in Frisbee throwing distance with respect to education status?

Variables:
IV - Education, which has three levels:
High School, Graduate and PostGrad;
DV - Frisbee Throwing Distance

A

A Tukey post hoc test shows that the PostGrad group was able to throw the frisbee statistically significantly further than the High School group (p = .034). There was no statistically significant difference between the Graduate and High School groups (p = . 691) nor between the Graduate and PostGrad groups (p = .099).

727
Q

What is IV and DV of one -way ANOVA?

A

IV = 1 predicto Categorical with more than 2 levels

DV = 1 Continous

728
Q

one-way ANOVA is also called

A

between subject

729
Q

regression equation for ANOVA
can be extended to include one or more continuous variables that predict the outcome (or dependent variable).

these continous variables are not

A

part of the main experimental manipulation but have an influence on the dependent variable, are known as covariates and they can be included
in an ANOVA analysis.

730
Q

What does ANCOVA involve?

A

When we measure covariates and include them in an
analysis of variance

731
Q

Continuous variables, that are not part of the main experimental manipulation (don’t want to study them) but have an influence on the dependent variable, are known as

A

covariates

732
Q

From what we know from hierarchical regression model, if we enter covariate into regression model first then dummy variables representing exp manipulation after… - (2)

A

then we can see what effect an IV has after the effect of covariate

We partial out the effect of covariate

733
Q

What are the two reasons for including covariates in ANOVA? - (2)

A
  • To reduce within-group error variance = if we can explain unexplained variance , SSR, in terms of other variables (covariates)then reduce SSR to accurately assess effects of SSM
  • Elimination of confoundd = remove bias of unmeasured variables that confound results and influence DV
734
Q

ANCOVOA has same assumptions of ANOVA, e.g., normality and homogenity of variance (Levene’s test) expect has two more important assumptions which are… - (2)

A
  • Independence of the covariate and treatment effect
  • Homogeneity of regression slopes
735
Q

For ANCOVA to reduce within-group variance by allowing the covariate to explain some of the error variance the covariate must be

A

independent from the experimental/treatment effect - (IVs - categorical predictors) ( ANCOVA assumption)

736
Q

People should no use ANCOVA when the effect of covariate overlaps with the experimental effect as it means the

A

experimental effect is confounded with the effect of covariate = interpretation of ANCOVA is compromised

737
Q

In ANCOVA, the effect of the covariate should be independent of the

A

experimental effect

738
Q

When an ANCOVA is conducted we look at the overall relationship between DV and covariate meaning we fit a regression line to

A

entire dataset and ignore which groups pps fit in

739
Q

When is homogenity of regression slope is not satisifed in ANCOVA?

A

the relationship between the
outcome (dependent variable) and covariate differs across the groups then the overall regression model is inaccurate (it does not represent all of the groups).

740
Q

What is best way to test homeogenity of regression slopes assumption in ANCOVA?

A

imagine plotting a scatterplot for each experimental condition with the covariate on one axis and the outcome on the other and calculate its regression line

741
Q

Diagram of regression slopes satisfying homogenity of regression slopes in ANCOVA

A
  • exhibits the same slopes for control and 15 minute group
742
Q

Diagram of a regression slopes not satisfying homogenity of regression slopes in ANOCVA

A
  • 30 minutes of therapy exhibts a different slope compared to others
743
Q

What is design, variables and test would you use to test this researh scenario? - (5)

A
  • ANCOVA
  • Independent samples-design
  • One IV , two conditions, interval regime and steady state
  • One covariate (age in years)
  • One DV (Race time)
744
Q

What does this ANCOVA output show?

  • IV = Regime –> steady or interval
  • Covariate = Age
  • DV = Racetime- (2)
A
  • Age F(1,27) = 5.36, p = 0.028, partial eta-squared = 0.17 (large and sig main effect)
  • Regime F(1,27) = 4.28, p = 0.048, partial eta-squared = 0.14 (large and sig main effect)
745
Q

What DF do you report from this ANCOVA table for age for example…

A

DF for age and DF for error

746
Q

Guidelines for interpreting partial eta-squared - (3)

A

η2 = 0.01 indicates a small effect.
η2 = 0.06 indicates a medium effect.
η2 = 0.14 indicates a large effect

747
Q

What does this SPSS output for ANCOVA show? - (3)

A
  • Interval has a marginal mean of race times of 56.57
  • Steady state has a marginal mean of race times 62.97
  • Estimated marginal means partialled out the effects of age and view mean scores of race times in interval and steady state if mean age scores (30.07) across two groups was held constant
748
Q

What does this output show in terms of homogenity of regression slopes?

age is covariate and regime is IV and DV is race times - (2)

A
  • Interaction effect of regime * age has a p-value of 0.980
  • Since p-value is not significant the assumption of homogeneity of regression slopes has been met
749
Q

What happens if the interaction effect of IV and covariate is significant in testing homogenity of regression slopes

A

relationship between covariate and DV differ significantly between two groups or many groups you got and assumption is not satisfied

750
Q

For testing assumption of independence of covariate and experimental effect (IV) in SPSS, we need to add

A

IV (e.g., regime) and covariate (e.g., age) in DV instead of covariate box

751
Q

What does this SPSS output show in terms of independence of covariate and exp effect (IV)?

age is covariate (treated as DV) , regime is IV - (2)

A
  • P-value is not signifcant (p=0.528) so effect of variable age is not sig difference of age across training regime
  • and so independent variable are assumed to be independent.
752
Q

What does positive and negative b-value for covariates in ANCOVA parameter estimate box indicate? - (2)

A

f the b-value for the covariate is positive then it means that the covariate and the outcome variable have a positive relationship

If the b-value is negative it means the opposite: that the covariate and the outcome variable have a negative relationship

753
Q

What does this table of parameter estimates show for ANCOVA where..

DV = PP’sLibido, IV = Dose of Viagara, Covariate is Partner’sLibido - (3)

A
  • b for covariate is 0.416
  • Besides other things being equal, if a a partner’s libido increases by one unit, then the person’s libido should
    increase by just 0.416 units
  • Since b is positive then partner’s libido ahs pos relation with pps’s libido
754
Q

How is DF calculated for these t-tests in ANCOVA table? - (2)

A

N - p -1
N is total sample size, p is number of predictors (2 dummy variables and covariate )

755
Q

What post-hoc tests can you do with ANCOVA? - (3)

A
  • Tukey LSD with no adjustments (not reccomended)
  • Bonferroni correction (reccomended)
  • Sidak correction
756
Q

The sidak correction is similar to what correction?

A

Bonferroni correction

757
Q

Sidak correction is less conserative than

A

Bonferroni correction

758
Q

The Sidak correction should be selected if you are concerned about

A

loss of power associated with Bonferroni corrected values.

759
Q

What does these planned contrast results show in ANCOVA?

DV = Pp’s Libido, IV = Dose of Viagara, Covariate is Partner’s Libido -

IV Dose: Level 3 = high dose, level 2 = low dose, level 1 = placebo

(3)

A
  • Contrast 1 of comparing level 2 (low dose) against level 1 (placebo) is significant (p = 0.045)
  • Contrast 2 of comparing level 3 (high dose) with level 1 (placebo) is significant (p - 0.010)
760
Q

What does this Sidak correction post-hoc comparison in ANCOVA output show?

DV = Libido, IV = Dose of Viagara, Covariate is Libido -

IV Dose: Level 3 = high dose, level 2 = low dose, level 1 = placebo

    • (3)
A
  • The significant difference between the high-dose and placebo groups remains (p = .030)
  • high-dose and low-dose groups do not significantly differ (p = .93)
  • Low dose and placebo groups do not significantly differ (p value = 0.130)
761
Q

What do these scatterplot of regression lines show in terms of homogenity of regression slopes?

DV = Libido, IV = Dose of Viagara, Covariate is Libido -

IV Dose: Level 3 = high dose, level 2 = low dose, level 1 = placebo

(3)

A

For placebo and low dose there appears to be a positive relationship between pp’s libido and that of their partner

However, in the high-dose condition there appears to be no relationship at all between participant’s libido and that of their partner - shows negative relationship

Doubts whether homogenity of regression slopes is satisfied as not all the slopes are the same (go same direction)

762
Q

What effect sizes can we use for ANCOVA/ANOVA? - (4)

A
  • eta-squared
  • partial-eta squared (ANCOVA)
  • omega squared = used when equal numbe of pps in each grp
  • r
763
Q

How is eta-squared calcuated?

A

Dividing the effect of interest SSM by total variance in the data SST

764
Q

How is partial eta-squared calculated for ANCOVA??

A

SS Effect/ SS Effect + SS Residual

765
Q

What is the difference between partial and eta-squared?

A

This differs from eta squared in that it looks
not at the proportion of total variance that a variable explains, but at the proportion of variance that a variable explains that is not explained by other variables in the analysis

766
Q

What test is used to investigate this question and how is it conducted? - (2)

We want to know whether or not studying technique (3 levels) has an impact on exam scores,but we want to account for the grade that the student already has in the class.

A
  • ANCOVA
  • ANCOVA is conducted to determine i f there is a statistically significant difference between different studying techniques (IV) on exam score (DV) after controlling for current grade (covariate)
767
Q

In ANCOVA, we partion the total variance into

A

IV, DV and covariate

768
Q

In ANCOVA, examine influence of categorical IVs on DV while removing the effect of

A

covariate factor(s)

769
Q

In ANCOVA, the covariate correlates with the … but not the ..

A

correlates with outcome DV but not with IV

770
Q

What is an example of covariate?

A

baseline pre-test scores can be used as a covariate to control for inital grp differences on test performance

771
Q

In ANCOVA, the IVS, Covariates and DVs are.. - (2)

A
  • IVs are categorical
  • Covariates are metric (quantiatively) independent of IV
  • DV is metric
772
Q

In ANCOVA, you have - (2)

A

1 DV: Continous

2 predictor variables with 2 levels or more that are categorical and continous

773
Q

What is example of continous? - (3)

A

infinite number of possible values variables can take on

e.g., interval = equal intervals on variable represent equal difference measured like diff between 600ms and 800ms is = difference between 1300ms and 1500ms

e.g., ratio = same as interval but clear definition of 0 like height or weight

774
Q

What is example of categorical variable? - (3)

A

A variable that cannot take on all values within the limits of the variable - entities are divided into distinct categories

e.g., nominal = 2 or more caegories e.g., whether someone is vegan or vegetarian

e.g., ordinal categories have order like people who got fail, pass, merit or distinction

775
Q

What does independence of covariate mean in ANCOVA?

A

Independence of the covariate and treatment effect means that the categorical predictors and the covariate should not be dependent on each other

776
Q

What does homogenity of regression slopes mean in ANCOVA?

A

Homogeneity of regression slopes means that the covariate has a similar relationship with the outcome measure, irrespective of the level of the categorical variable - in this case the group

777
Q

For homogeneity of regression slopes in ANCOVA, there are

A

There are alternative, a bit more advanced, methods to account for such differences as they are not, in general, uninteresting, but for the ANCOVA analysis they do present an issue

778
Q

In ANCOVA, between subject effects we quote DF such as for dose as…

A

Quote df for the effect and error, e.g. 2,26

779
Q

In ANCOVA, adjusted means table in SPSS shows.. - (2)

outcome/DV = happiness measure ranging from 0 to 10 (as happy as I can image) = continous = interval

The fixed factor (IV) which is dose of therapy which is people have 15 minutes of puppy therapy or 30 minutes

Covariate is control group is how much they love puppies = continous = interval

A

The group means can be recalculated once the effect of the covariate is ‘discounted’ = impact of covariate is taken into account and adjusted into each level of predictor variable in mean column

These values can differ markedly from the original group means and help with interpretation.

780
Q

ANCOVA is extension of ANOVA as - (2)

A
  1. Control for Covariances (continuous variables you may not necessarily want to measure)
  2. Study combinations of categorical and continuous variables – covariate becomes the variable of interest rather than the one you control
781
Q

What ANCOVA was conducted?

We want to know whether or not studying technique has an impact on exam scores,but we want to account for the grade that the student already has in the class.

A

A three-way ANCOVA was conducted to determine a statistically significant difference between different study techniques on students exam scores after controlling for their current grades.

782
Q

Assumptions of ANCOVA - (8)

A

Independent variablesshould becategorical variables.

Thedependent variableand covariate should be continuous variables(measured on aninterval scaleorratio scale.)

Make sureobservations are independent - don’t put people into more than one group.

Normality: the dependent variable should be roughlynormalfor each of category ofindependent variables.

Data (and regression slopes) should showhomogeneity of variance.

The covariate and dependent variable (at eachlevelof independent variable) should belinearly related.

Your data should behomoscedastic

The covariate and theindependent variableshouldn’t interact.In other words, there should be homogeneity of regression slopes.

783
Q

In one-way ANOVA we partition the total variance into

A

IV and DV

784
Q

A psychologist was interested in the effects of different fear information on children’s beliefs about an animal. Three groups of children were shown a picture of an animal that they had never seen before (a quoll). Then one group was told a negative story (in which the quoll is described as a vicious, disease-ridden bundle of nastiness that eats children’s brains), one group a positive story (in which the quoll is described as a harmless, docile creature who likes nothing more than to be stroked), and a final group weren’t told a story at all. After the story children rated how scared they would be if they met a quoll, on a scale ranging from 1 (not at all scared) to 5 (very scared indeed). To account for the natural anxiousness of each child, a questionnaire measure of trait anxiety was given to the children and used in the analysis

what analysis has been used -

Independent analysis of variance

Repeated-measures analysis of variance

Mixed analysis of variance

Analysis of covariance

A

Analysis of covariance (ANCOVA)

785
Q

A psychologist was interested in the effects of different fear information on children’s beliefs about an animal. Three groups of children were shown a picture of an animal that they had never seen before (a quoll). Then one group was told a negative story (in which the quoll is described as a vicious, disease-ridden bundle of nastiness that eats children’s brains), one group a positive story (in which the quoll is described as a harmless, docile creature who likes nothing more than to be stroked), and a final group weren’t told a story at all. After the story children rated how scared they would be if they met a quoll, on a scale ranging from 1 (not at all scared) to 5 (very scared indeed). To account for the natural anxiousness of each child, a questionnaire measure of trait anxiety was given to the children and used in the analysis

what is covariate?

A

Natural Fear Level

786
Q

Which of the designs below would be best suited for ANCOVA?

A. Participants were randomly allocated to one of two stress management therapy groups, or a waiting list control group. Their levels of stress were measured and compared after 3 months of weekly therapy sessions.

B. Participants were allocated to one of two stress management therapy groups, or a waiting list control group based on their baseline levels of stress. The researcher was interested in investigating whether stress after the therapy was successful partialling out their baseline anxiety.

C. Participants were randomly allocated to one of two stress management therapy groups, or a waiting list control group. The researcher was interested in the relationship between the therapist’s ratings of improvement and stress levels over a 3-month treatment period.

D.Participants were randomly allocated to one of two stress management therapy groups, or a waiting list control group. Their baseline levels of stress were measured before treatment, and again after 3 months of weekly therapy sessions.

(2)

A

D since baseline levels of stress used as covariate and use this as a control when looking at impact treatment has had over 3 month assessment

Not B since grps allocated based on baseline levels of stress (covariate and IV correlated - problematic) and A and C is one-way independent ANOVA

787
Q

A psychologist was interested in finding a cure for hangovers. She took 50 people out on the town one night and got them drunk. The next morning, she allocated them to either a control condition (drink water only) or an experimental hangover cure condition (a beetroot, raw egg and chilli smoothie). This is the variable ‘Group’. Two hours later she then measured how well they felt on a scale from 0 (‘I feel fine’) to 10 (‘I am about to die’)(Variable = Hangover).

She also realized she ought to ask them how drunk they were the night before and control for this in the analysis, so she measured this on another scale of 0 (‘sober’) to 10 (‘very drunk’) (Variable = Drunk). The psychologist hypothesised that the smoothie drink would lead to participants feeling better, after having accounted for the previous night’s drunkenness.

What test?

A

ANCOVA

788
Q

A psychologist was interested in finding a cure for hangovers. She took 50 people out on the town one night and got them drunk. The next morning, she allocated them to either a control condition (drink water only) or an experimental hangover cure condition (a beetroot, raw egg and chilli smoothie). This is the variable ‘Group’. Two hours later she then measured how well they felt on a scale from 0 (‘I feel fine’) to 10 (‘I am about to die’)(Variable = Hangover).

She also realized she ought to ask them how drunk they were the night before and control for this in the analysis, so she measured this on another scale of 0 (‘sober’) to 10 (‘very drunk’) (Variable = Drunk). The psychologist hypothesised that the smoothie drink would lead to participants feeling better, after having accounted for the previous night’s drunkenness.

Identify IV (fixed), DV and covariate - (3)

A
  • IV: Group
  • DV: Hangover
  • Covariate: Drunk
789
Q

What is the decision tree of choosing a two-way independent ANOVA? - (5)

A

Q: What sort of measurement? A: Continuous
Q:How many predictor variables? A: Two or more
Q: What type of predictor variable? A: Categorical
Q: How many levels of the categorical predictor? A: Not relevant
Q: Same or Different participants for each predictor level? A: Different

790
Q

Partial eta-squared should be reported for

A

ANOVA and ANCOVA

791
Q

What is the two drawbacks of eta-squared?

A

as you add more variables to the model, the proportion explained by any one variable will automatically decrease.

792
Q

How is eta-squared calculated?

A

Sum of squares between (squares of effect M) divided by sum of squared total (squares of everything - effects, errors and interactions)

793
Q

In one-way ANOVA eta-squared and partial-eta squared will be eequal but not true in models with

A

more than one IV

794
Q

Two-way Independent ANOVA is also called an

A

Independent Factorial ANOVA

795
Q

What is a factorial design?

A

When experiment has two or more IVs

796
Q

What are the 3 types of factorial design? - (3)

A
  1. Independent factorial design
  2. Repeated-measures (related) factorial design
  3. Mixed design
797
Q

What is independent factorial design?

A
  • There is many IVs or predictors that each have been measured using different pps (between grps)
798
Q

What is repeated-measures (related) factorial design?

A
  • Many IVs or predictors have been measured but same pps used in all conditions
799
Q

What is mixed design?

A
  • Many IVs or predictors have been measured; some measured with diff pps whereas others used same pps
800
Q

Which design does independent factorial ANOVA use?

A

Independent factorial design

801
Q

What is factorial ANOVA?

A

When we use ANOVA to analyse a situation in which there is two or more IVs

802
Q

What is difference between one way and two way ANOVA?

A

A one-way ANOVA has one independent variable, while a two-way ANOVA has two.

803
Q

Example of two-way independent factorial ANOVA

The study tested the prediction that subjective perceptions of physical attractiveness become inaccurate after drinking alcohol which is IV, DVs

  • What are the IVs, DVs- (3)
A

IV = Alcohol - 3 levels = Placebo, Low dose, High dose
Iv = face type 2 levels = unattractive, attractive
DV = Physical attractiveness score

804
Q

Two way independent ANOVA can be fit into the idea of

A

linear model

805
Q

The study tested the prediction that subjective perceptions of physical attractiveness become inaccurate after drinking alcohol

IV = Alcohol - 3 levels = Placebo, Low dose, High dose
Iv = face type 2 levels = unattractive, attractive
DV = Physical attractiveness score

Create a linear model for this two-way ANVOA scenario which adds interaction term and explain why is it important - (3)

A
  • The first equation models the two predictors in a way that allows them to account for variance in the outcome separately, much like a multiple regression model
  • The second equation adds a term that models how the two predictor variables interact with each other to account for variance in the outcome that neither predictor can account for alone.
  • The interaction is important to us because it tests our hypothesis that alcohol will have a stronger effect on the ratings of unattractive than attractive faces
806
Q

How do we know coefficients in model are significant in two-way ANCOVA?

A

We follow the same routine , similar to one-way ANOVA, to compute sums of squares for each factor of the model (and their interaction) and compare them to the residual sum of squares, which measures what the model cannot explain

807
Q

How is two-way independent ANOVA similar to one-way ANOVA?

A

, we still find the total sum of squared errors (SST) and break this variance down into variance that can be explained by the experiment (SSM) and variance that cannot be explained (SSR).

808
Q

How is two-way INDEPENDENT ANOVA different to one-way INDEPENDENT ANOVA? - (3)

A

in two-way ANOVA, the variance explained by the experiment is made up of not one experimental manipulation but two.

Therefore, we break the model sum of squares down
into variance explained by the first independent variable (SSA), variance explained by the second independent variable (SSB) and variance explained by the interaction of these two
variables (SSA × B)

809
Q

How to calculate total sum of squares SST in two-way independent ANOVA?

A
810
Q

What is SST DF in two-way independent ANOVA?

A

N- 1

811
Q

How to compute model sum of squares SSM in two-way independent ANOVA? - (2)

A

sum of all grps (pairing each level of IV with another)

n = number of scores in each grp which is multipled by the mean value of each group subtracted by grand mean of all pps regardless of grp squared

812
Q

How to compute degrees of freedom of SSM in two-way independent ANOVA?

A

(g-1)

813
Q

How many groups are there in this research two-way independent ANOVA?

IV = Alcohol - 3 levels = Placebo, Low dose, High dose
Iv = face type 2 levels = unattractive, attractive
DV = Physical attractiveness score

A

placebo + attractiveness
placebo + untractiveness
low dose +attractiveness
low dose + unattractiveness
high dose +attractiveness
high dose +unattractiveness - 6 grps

814
Q

How is SSA (face type) computed in two-way independent ANOVA?

IV = Alcohol - 3 levels = Placebo, Low dose, High dose
Iv = face type 2 levels = unattractive, attractive
DV = Physical attractiveness score - (2)

A

considering only two groups at a time and add together - for first IV variable (SSA) (e.g., grps of pps rated attractive and grp of pps that rated unattractive)

number of pps in that grp multiplied by mean of grp subtracted by grand mean overall of all pps squared

815
Q

What is the degrees of freedom in SSA in TWO-WAY INDEPENDENT ANOVA?

A

DF = (g-1) so if male and female then 2 -1 = 1

816
Q

How to compute SSB in two-way independent ANOVA for alcohol type

IV = Alcohol - 3 levels = Placebo, Low dose, High dose
Iv = face type 2 levels = unattractive, attractive
DV = Physical attractiveness score - (2) - (3)

A

same formula as SSA but for the second IV

added for all grps of pps in second IV

number of pps in one grp of secondIV(mean score of that grp subtract by grand mean of all pps regardless of grp) squared

817
Q

What is DF for SSB in two-way independent ANOVA?

A

number of grps in second IV minus 1

818
Q

SS A X B in two-way independent ANOVA is calculating how much variance is explaiend

A

by the interaction of 2 variables

819
Q

How is SS A X B (interaction term) calculated in two-way ANOVA?

A

SS A X B = SSM - SSA - SSB

820
Q

How is SS A X B’S DF calculated in two-way independent ANOVA?

A

df A X B = df M - df A - df B

821
Q

The SSR in two-way independent ANOVA, is similar to one-way ANOVA as it represents the

A

individual differences in performance or the variance that can’t be explained by factors that were systematically manipulated.

822
Q

How to calculate SSR in two-way independent ANOVA?

A
  • use individual variances of each grp (e.g., attractiveness face type + placebo) and multiply by one less than number of people within the group (n - in this case 6) and do it for each group and add it together
823
Q

How to calculate SSR in two-way independent ANOVA?

A

number of grps you have in study(number of scores you have per group minus 1)

824
Q

Diagram of calculating mean sums of squares in two-way ANOVA independent

A
825
Q

What effect sizes can we calculate with two-way independent ANOVA? - (2)

A
  • Partial eta-squared
  • Omega-squared if advised
826
Q

What to do whe assumptions are violated in factorial independent ANOVA? - (3)

A
  • There is not a simple non-parametric counterpart of factorial ANOVA
  • If assumption of normality is violated then use robust methods described in Wilcox’s and files in R
  • If assumptions of homogenity of variance then implement corrections based on Welch procedure
827
Q

Example of a research scenario of two-way independent ANCOVA

Pick out IVs and DVs - (4)

A
  • Independent samples design
  • Two Ivs, both 2 conditions: drug type (A, B) and onset (early, late)
  • One DV is cognitive performance
  • Two way ANOVA
828
Q

What does this two-way ANOVA independent design SPSS output show?

A
  • The levene’s test is not significant so assume equal variances
829
Q

What happens if Levene’s test is significant in two-way independent ANOVA?

A

steps taken to equalise variances through data transformation

830
Q

What does this two-way independent ANOVA table show - (4)

A
  • Drug : F(1,24) = 5.58, p = 0.027, partial eta-squared = 0.19 (large effect + sig effect)
  • Onset: F(1,24) = 14.43, p = 0.001, partial eta-squared = 0.38 (large effect + sig effect)
  • Interaction Drug * Onset: F(1,24) = 9.40, p = 0.005, partial eta-squared = 0.28 (large effect + sig effect)
  • We got two sig main effects and sig interaction effect which are all quite large effect sizes
831
Q

What does this SPSS output show for two-way independent ANOVA? - (3)

A

drug B has higher score on cognitive test than A and is sig main effect (CI does not contain 0 and also main effect analysis)

early onset scoring higher on average than late onset (CI does not contain 0 and also main effect analysis)

Important of these main effect as main effects ignoring the effec tof other IV so results for drug at top is regardless of whether late/onset for example , does not tell anything for interaction

832
Q

What does this interaction plot show TWO WAY ANOVA? - (6)

A
  • Blue line is early onset
  • Green line is late onset
  • For late onset, drug B lead to higher mean scores on test than drug A
  • For early onset, drug A led to slightly higher mean scores than drug B
  • Drug A more effective then drug b for early onset but different marginal
  • Drug B was substantially more effective than Drug A for late
833
Q

Non-parallel lines in interaction plot indicate an

A

sig interaction effect

834
Q

We can follow interactions in two-way ANOVA with simple effects analysis which - (2)

A
  • looks at the effect of one IV at individuals levels of other IV
  • Seeing whether differences margina/substantial is sig
835
Q

The SSM in two-way independent ANOVA is broken down into three components:

A

variance explained by the first independent
variable (SSA), variance explained by the second independent variable (SSB ) and variance explained by the interaction of these two variables (SSA × B
).

836
Q

Example of difference of one-way ANOVA vs two-way ANOVA (independent) - (2)

A
  • One-way ANOVA have one IV categorical variable (level of educaiton - college degree, grad degree, high school)
  • Two-way ANOVA , you have 2 categorical IV variables - level of education (college degree, grad degree, high school) and zodaic sign (libra, pisces)
837
Q

In two-way independent ANOVA, you need how many DV and IV?

A

1 DV and 2 or more categorical predictors

838
Q

What test is used for this scenario?

A psychologist wanted to test a new type of drug treatment for ADHD called RitaloutTM. The makers of this drug claimed that it improved concentration without the side effects of the current leading brand of ADHD medication.

To test this, the psychologist allocated children with ADHD to two experimental groups, one group took RitaloutTM(New drug), the other took the current leading brand of medication (Old drug) (Variable = Drug).

To test the drugs’ effectiveness, concentration was measured using the Parker-Stone Concentration Scale, which ranges from 0 (low concentration) to 12 (high concentration) (Variable = Concentration).

In addition, the psychologist was interested in whether the effectiveness of the drug would be affected by whether children had ‘inattentive type’ ADHD or ‘hyperactive type’ ADHD (Variable = ADHD subtype).

A

Two-way independent ANOVA

839
Q

A researcher was interested in measuring the effect of 3 different anxiety medications on patients diagnosed with anxiety disorder. They measured anxiety levels before and after treatment of 3 different treatment groups plus a control group. The researchers also collected data on depression levels.

Identify the IV, DV, and covariates! - and design (3)

A

IV = 3 different types anxiety medications and control grp
DV: Anxiety levels after treatment of grps
Covariate = anxiety before treatment, depression levels

ANCOVA

840
Q

Researchers wanted to see how much people of different education levels are interested in politics. They also believed that there might be an effect of gender. They measured political interest with a questionnaire in males and females that had either school, college or university education.

Identify the IVs and DV and design - (3)

A
  • IV: Level of education - school, college or uni edu and gender (m, f)
  • DV: Political interest in questionnaire
  • Two-way independent ANOVA
841
Q

An experiment was done to look at whether there is an effect of both gender and the number of hours spent practising a musical instrument on the level of musical ability.
A sample of 30 participants (15 men and 15 women) who had never learnt to play a musical instrument before were recruited. Participants were randomly allocated to one of three groups that varied in the number of hours they would spend practising every day for 1 year (0 hours, 1 hours, 2 hours). Men and women were divided equally across groups.
All participants had a one-hour lesson each week over the course of the year, after which their level of musical skill was measured on a 10-point scale ranging from 0 (you can’t play for toffee) to 10 (‘Are you Mozart reincarnated?’).

Identify IVs and DV and design - (3)

A
  • IV: Gender (m,f) , number of hrs spent practicisng
  • DV: Level of muscial skill after a year
  • Two-way independent ANOVA, not t-tests since more than one IV
842
Q

In these outputs is there a effect of gender, education or interaction level TWO WAY ANOVA INDEPENDENT

A
  • Is there an effect of gender overall?
    No, F(1,54) = 1.63, p = .207
    Is there an effect of education level?
    Yes, F(2,54) = 147.52, p < .001
    Is there an interaction effect?
    Yes, F(2,54) = 4.64, p = .014
843
Q

How to interpret these findings?

A
  • Main effect of Aspirin: Aspirin reduces heart attackes compard to placebo (1)
  • Main effect of carotene: Beta carotene reduces heart attack (2)
  • Interaction effect: Yes, bigger effect when aspirin and beta carotene taken together (3) - also lines drawn more its an interaction
844
Q

WHICH STATEMENT BEST DESCRIBES A COVARIATE?

A variable that is not able to be measured directly.

A variable that shares some of the variance of another variable in which the researcher is interested.

A pair of variables that share exactly the same amount of variance of another variable in which the researcher is interested.

A variable that correlates highly with the dependent variable.

A

A variable that shares some of the variance of another variable in which the researcher is interested.

845
Q

TWO-WAY ANOVA IS BASICALLY THE SAME AS ONE-WAY ANOVA, EXCEPT THAT:

The model sum of squares is partitioned into two parts

The residual sum of squares represents individual differences in performance

The model sum of squares is partitioned into three parts

We calculate the model sum of squares by looking at the difference between each group mean and the overall mean

A

C. The model sum of squares is partitioned into three parts

The model sum of squares is partitioned into the effect of each of the independent variables and the effect of how these variables interact (see Section 13.2.7)

D is also true, but we also do this for both one-way and two-way ANOVA (see Section 13.2.7).

846
Q

IF WE WERE TO RUN A FOUR-WAY BETWEEN-GROUPS ANOVA, HOW MANY SOURCES OF VARIANCE WOULD THERE BE?

4
16
12
15

A

16 because 4*4 = 16 (if it was 3x2 then would be 6)

847
Q

Which of the following sentences best describes a covariate?

A. A variable that shares some of the variance of another variable in which the researcher is interested.

B. A variable that correlates highly with the dependent variable

C. A variable that is not able to be measured directly

D. A pair of variables that share exactly the same amount of variance of another variable in which the researcher is interested

A

A

848
Q

An experiment was done to look at whether there is an effect of both gender and the number of hours spent practising a musical instrument on the level of musical ability.
A sample of 30 participants (15 men and 15 women) who had never learnt to play a musical instrument before were recruited. Participants were randomly allocated to one of three groups that varied in the number of hours they would spend practising every day for 1 year (0 hours, 1 hours, 2 hours). Men and women were divided equally across groups.
All participants had a one-hour lesson each week over the course of the year, after which their level of musical skill was measured on a 10-point scale ranging from 0 (you can’t play for toffee) to 10 (‘Are you Mozart reincarnated?’).

A. Two-way independent ANOVA

B. Two-way repeated ANOVA

C. Three way ANOVA = only 2 IVs so no

D. T-test

A

A

849
Q

Which of the designs below would be best suited to ANCOVA?

A. Participants were randomly allocated to one of twostress management therapy groups, or a waiting listcontrol group. Their baseline levels of stress weremeasured before treatment, and again after 3months of weekly therapy sessions

B. Participants were randomly allocated to one of twostress management therapy groups, or a waiting listcontrol group. Their levels of stress were measuredand compared after 3 months of weekly therapysessions.

C. Participants were randomly allocated to one of twostress management therapy groups, or a waiting listcontrol group. The researcher was interested in therelationship between the therapist’s ratings ofimprovement and stress levels over a 3-monthtreatment period.

D. Participants were allocated to one of two stressmanagement therapy groups, or a waiting listcontrol group based on their baseline levels ofstress. The researcher was interested ininvestigating whether stress after the therapy wassuccessful partialling out their baseline anxiety

A

A - baseline levels of stress used as covariate

. We can use the baseline, pre-treatment measures as a control when looking at the impact the treatment has on the 3-month assessment.

850
Q

A music teacher had noticed that some students went to pieces during exams. He wanted to testwhether this performance anxiety was different for people playing different instruments. He tookgroups of guitarists, drummers and pianists (variable = ‘Instru’) and measured their anxiety(variable = ‘Anxiety’) during the exam. He also noted the type of exam they were performing (inthe UK, musical instrument exams are known as ‘grades’ and range from 1 to 8). He wanted tosee whether the type of instrument played affected performance anxiety when accounting for thegrade of the exam. Which of the following statements best reflects what the effect of ‘Instru’ in theoutput table below tells us?

(Hint: ANCOVA looks at the relationship between an independent and dependent variable, takinginto account the effect of a covariate.

A. The type of instrument played in the exam had asignificant effect on the level of anxietyexperienced, even after the effect of the grade ofthe exam had been accounted for

B. The type of instrument played in the exam had asignificant effect on the level of anxietyexperienced

C. The type of instrument played in the exam did nothave a significant effect on the level of anxietyexperienced

A

A

851
Q

Question 5A psychologist was interested in the effects of different fear information on children’s beliefs aboutan animal. Three groups of children were shown a picture of an animal that they had never seenbefore (a quoll). Then one group was told a negative story (in which the quoll is described as avicious, disease-ridden bundle of nastiness that eats children’s brains), one group a positive story(in which the quoll is described as a harmless, docile creature who likes nothing more than to bestroked), and a final group weren’t told a story at all. After the story children rated how scared theywould be if they met a quoll, on a scale ranging from 1 (not at all scared) to 5 (very scaredindeed). To account for the natural anxiousness of each child, a questionnaire measure of traitanxiety was given to the children and used in the analysis. The SPSS output is below.

Whatanalysis has been used?

(Hint: The analysis is looking at the effects of fear information on children’s beliefs about an animal, taking into account children’s natural fear levels.)

A. ANCOVA

B. Independent analysis of variance

C. Repeated measures analysis of variance

A

A

852
Q

Imagine we wanted to investigate the effects of three different conflict styles (avoiding, compromising and competing) on relationship satisfaction, but we discover that relationship satisfaction is known to covary with self-esteem. Which of the following questions would be appropriate for this analysis?

A. What would the mean relationship satisfaction be for the three conflict style groups, if their levels of self-esteem were held constant?

B. What would the mean relationship satisfaction be if levels of self-esteem were held constant?

C. What would the mean self-esteem score be for the three groups if their levels of relationship satisfaction were held constant?

D. Does relationship satisfaction have a significant effect on the relationship between conflict style and self-esteem?

A

A

853
Q

A study was conducted to look at whether caffeine improves productivity at work in different conditions. There were two independent variables. The first independent variable was email, which had two levels: ‘email access’ and ‘no email access’. The second independent variable was caffeine, which also had two levels: ‘caffeinated drink’ and ‘decaffeinated drink’. Different participants took part in each condition. Productivity was recorded at the end of the day on a scale of 0 (I may as well have stayed in bed) to 20 (wow! I got enough work done today to last all year). Looking at the group means in the table below, which of the following statements best describes the data?

A. A significant interaction effect is likely to be present between caffeine consumption and email access.

B. There is likely to be a significant main effect of caffeine.

C. The effect of email is relatively unaffected by whether the drink was caffeinated.

D. The effect of caffeine is about the same regardless of whether the person had email access.

A

A = for decaffeinated drinks there is little difference between email and no email, but for caffeinated drinks there is

854
Q

What are the two main reasons for including covariates in ANOVA?

A. 1. To reduce within-group error variance
2. Elimination of confounds

B. 1. To increase within-group error variance
2. To reduce between-group error variance

C. 1. To increase within-group error variance
2. To correct the means for the covariate

D. 1. To increase between-group variance
2. To reduce within-group error variance

A

A

855
Q

A psychologist was interested in the effects of different fear information on children’s beliefs about an animal. Three groups of children were shown a picture of an animal that they had never seen before (a quoll). Then one group was told a negative story (in which the quoll is described as a vicious, disease-ridden bundle of nastiness that eats children’s brains), one group a positive story (in which the quoll is described as a harmless, docile creature who likes nothing more than to be stroked), and a final group weren’t told a story at all. After the story children rated how scared they would be if they met a quoll, on a scale ranging from 1 (not at all scared) to 5 (very scared indeed). To account for the natural anxiousness of each child, a questionnaire measure of trait anxiety was given to the children and used in the analysis. Which of the following statements best reflects what the ‘pairwise comparisons’ tell us?

A. Fear beliefs were significantly higher after negative information compared to positive information and no information, and fear beliefs were not significantly different after positive information compared to no information.

B. Fear beliefs were significantly lower after positive information compared to negative information and no information; fear beliefs were not significantly different after negative information compared to no information.

C. Fear beliefs were significantly higher after negative information compared to positive information; fear beliefs were significantly lower after positive information compared to no information.

D. Fear beliefs were all about the same after different types of information.

A

A

856
Q

a musical instrument and gender on the level of musical ability. A sample of 30 (15 men and 15 women) participants who had never learnt to play a musical instrument before were recruited. Participants were randomly allocated to one of three groups that varied in the number of hours they would spend practising every day for 1 year (0 hours, 1 hours, 2 hours). Men and women were divided equally across groups. All participants had a one-hour lesson each week over the course of the year, after which their level of musical skill was measured on a 10-point scale ranging from 0 (you can’t play for toffee) to 10 (‘Are you Mozart reincarnated?’). An ANOVA was conducted on the data from the experiment. Which of the following sentences best describes the pattern of results shown in the graph?

A. The graph shows that the relationship between musical skill and time spent practising was different for men and women.

B. The graph shows that the relationship between musical skill and time spent practising was the same for men and women.

C. The graph indicates that men and women were most musically skilled when they practised for 2 hours per day.

D. Women were more musically skilled than men.

A

A

857
Q

What is the decision tree for choosing one-way repeated measures ANOVA? - (5)

A

Q: What sort of measurement? A: Continuous
Q:How many predictor variables? ONE IV
Q: What type of predictor variable? A: Categorical
Q: How many levels of the categorical predictor? More than two
Q: Same or Different participants for each predictor level? A: Same

858
Q

The assumption of sphericity in within-subject design ANOVA can be likened to

A

the assumption of homogeneity of variance in
between-group ANOVA

859
Q

Sphericity is sometimes denoted as IN REPEATED ANOVA

A

ε or circularity

860
Q

What does spherecity refer to repeated anova?

A

equality of variances of the differences between treatment levels.

861
Q

you need at least … conditions for spherecity to be an issue in repeated ANOVA

A

three

862
Q

How is sphereicty assumed in this datasetA (USED IN REPEATED ANOVA

A
863
Q

How is spherecity calculated? - (2) REPEATED ANOVA

A
  • Calculating differences between between pairs of scores for all treatment levels e.g., A-B, A-C , B-C
  • Calculating variances of these differences e.g., variances of A-B, A-C, B-C
864
Q

What does the data from table show in terms of assumption of spherecity (calculated by hand) REPEATED ANOVA? - (3)

A

there is some deviation from sphericity because the variance of the differences between conditions A and B (15.7) is greater than the variance of the differences
between A and C (10.3) and between B and C (10.3).

However, these data have local circularity (or local sphericity) because two of the variances of differences are identical.

The deviation from spherecity in the data does not seem too severe (all variances roughly equal) but here assess deviation is serve to warrant an action

865
Q

How to assess the assumption of sphereicity in SPSS REPEATED ANOVA?

A

via Mauchly’s test

866
Q

If Mauchly’s test statistic is significant (p < 0.05) then REPEATED ANOVA

A

variance of differences between conditions are significnatly different - must be vary of F-ratios produced by computer

867
Q

If Mauchly’s test statistisc is non significant (p > 0.05) then it is reasonable to conclude that the REPEATED ANOVA

A

varainces of the differences between conditions are equal and does not significantly differ

868
Q

Signifiance of Mauchly’s test REPEATED ANOVA is dependent on

A

sample size

869
Q

Example of signifiance of Maulchy’s test dependent on sample size REPEATED ANOVA - (2)

A

in big samples small deviations from sphericity can be
significant,

small samples large violations can be non-significant

870
Q

What happens if the data violates the sphereicity assumption REPEAED ANOVA? - (2)

A

several corrections that can be applied to
produce a valid F-ratio

or

use multivariate test statistics (MANOVA)

871
Q

What corrections to apply to produce valid F-ratio when data violates sphereicity REPEATED ANOVA? - (2)

A
  • Greenhouse-Geisser correction ε
  • Huynh-Feldt correction
872
Q

The Greenhouse-Geisser correction varies between REPEATED ANOVA

A

1/k-1 (k is number of repeated measures conditions) and 1

873
Q

The closer that Greenhouse Geisser correction is to 1, the REPEATED ANOVA

A

more homogeneous the variances of differences, and hence the closer the data are to being spherical.

874
Q

How to calculate lower-bound estimate fo spherecity for Greenhouse-Geisser correction when there is 5 conditions REPEATED ANOVA? - (2)

A

Limit of f ε^ is 1/k (number of repeated-measures conditions)

so… 1/(5-1) = 1/4 = 0.25

875
Q

Huynh and Feldt (1976) reported that when the

A

Greenhouse-Geisser correction is too conservative

876
Q

Huynh-Feldt correction is less conservative than

A

Greenhouse-Geisser correction

877
Q

Why is MANOVA used when data that violates spherecity IN REEATED ANOVA?

A

MANOVA is not dependent upon the assumption of sphericity

878
Q

In repeated measures ANOVA, the effect of our experiment is shown up in within participant variance than

A

between group variance

879
Q

In independent ANOVA, the within-group variance is our…. and it is not contaimed by… - (2)

A

residual variance (SSR) = variance produced by individual differences in performance

SSR is not contaimined by experimental effect as study carried out by different people

880
Q

In repeated-measures ANOVA, the within-participant variability is made up of

A

the effect of experimental manipulation SSM and individual differences in performance (random factors outside of our control) - this is error SSR

881
Q

Similar to independent ANOVA, repeated-measures ANOVA uses F-ratio to - (2)

A

compares the size of the variation due to our experimental
manipulations to the size of the variation due to random factors

has same type of variances in independent - total sum of squares (SST), model sum of squares (SSM) and a residual sum of squares (SSR)

882
Q

What is the differences between independent ANOVA and repeated-measures design ANOVA?

A

repeated-measures ANOVA the model and residual sums of squares are both part of the within-participant variance.

883
Q

In repeated-measures ANOVA

If the variance due to our manipulations is big relative to
the variation due to random factors, we get a .. and conclude - (2)

A

big value of F ratio

we can conclude that the observed results are unlikely to have occurred if there was no effect in the population.

884
Q

To compute F-ratios we first compute the sum of squares which is the following REPEATED ANOVA… - (5)

A
  • SST
  • SSB
  • SSW
  • SSM
  • SSR
885
Q

How is SST calculated in one-way repeated measures ANOVA? REPEATED ANOVA

A

SST = grand variance (N-1)

886
Q

What is the DF’s of SST? REPEATD ANOVA

A

N-1

887
Q

The SSW (within-participant) sum of squares is calculated in one-way repeated ANOVA by…

A

square of the standard deviation of each participant’s scores multiplied by the number of conditions minus 1, summed over all participants.

888
Q

What is the DF of SSW of one-way repeated ANOVA? - (2)

A

DF = N(n-1)

number of participants multiplied by the number of conditions minus 1;

889
Q

How is SSM calculated in one-way repeated ANOVA? - (2)

A

square of the differences between the mean of the participant scores for each condition and the grand mean multiplied by the number of participants tested, summed over all conditions.

do this for each condition grp

890
Q

What is the DF of SSM in one-way repeated ANOVA? - (2)

A

DF = n-1
n is number of conditions

891
Q

How is SSR calculated in one-way repeated ANOVA?

A

the difference between the within-participant sum of squares and the sum of squares for the model.

892
Q

What is the DF for SSR in one-way repeated ANOVA?

A

DF of SSW minus DF of SSM

893
Q

How do we calculated mean squares (MS) and mean residuals (MR) to acalculate F-ratio in one-way repeated ANOVA?

A
894
Q

We don’t need to use SSB (between-subject variation )to calculate F-ratio in

A

one-way repeated ANOVA

895
Q

What does SSB represent in one-way ANOVA?

A

individual differences between cases

896
Q

Not only does sphereicity produces problems for F in repeated measures ANOVA but causes complications for

A

post-hoc tests

897
Q

When spereicity is violated in one-way repeated ANOVA , what post-hoc test to use and why - (2)

A

Bonferroni method seems to be generally the
most robust of the univariate techniques,

especially in terms of power and control of the Type I error rate.

898
Q

When sphereicity is not violated in one-way repeated ANOVA, then what post-hoc tests to use?

A

Tukey can be used

899
Q

In either case where sphereicity is violated or not in one-way repeated ANOVA, a post-hoc test called - (2)

A

Games–Howell procedure, which uses a pooled error term,

it is more preferable to Tukey’s test.

900
Q

Due to complications of sphereicity in one-way repeated ANOVA,

A

standard post hoc tests used for independent designs not avaliable for repeated measure designs

901
Q

Why is repeated contrast useful in repeated-measures design especially one-way repeated measures?

A

levels of the independent variable have a meaningful order e.g., meausred DV at successive time points or adminstered increasing doses of a drug

902
Q

When should Sidack correction as post hoc be selected for one-way repeated ANOVA?

A

concerned about the loss
of power associated with Bonferroni corrected values.

903
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

what does these SPSS outputs show? - (2)

A
  • Left shows variables represent each level of IV which is animal
  • Right shows descriptive statistics - higher mean time to retch when celebrity eating stick insect (8.12)
904
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - ONE WAY repeated anova

What does this Mauchly’s Test of Spherecity show? - (2)

A
  • P-value is 0.047 which is less than 0.05
  • Thus, reject the assumption of spherecity that variances of the differences between levels are equal
905
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub)

What to do if this Mauchly’s Test of Spherecity shows assumption of sphereicity is violated..? - (3)

one way repeated ANOVA

A
  • Since there are 4 conditions, lower limit of ε^ is 1/(4-1) = 0.333 (lower-bound estimate in table)
  • SPSS Output 13.2 shows that the calculated value of ε
    ^ is 0.533.
  • 0.533 is closer to the lower limit of 0.33 than it is to the upper limit of 1 and it therefore represents a substantial deviation from sphericity
906
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

What does this main ANOVA table show in terms of spherecity assumed? - (2)

A
  • The value of F = 3.97 which is compared against a critical value for 3 and 21 DF and p-value is 0.026
  • conclude there is significant difference between 4 animals in their capacity to induce retching when eaten
907
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

What has changed and kept the same in the table? - (2)

A
  • The F-ratios are the same across the rows
  • the D.F is changed as well as critical value the F-statistic is compared with
908
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

How is adjustments made to DF?

A
  • Adjustment made by multiplying the DF by the estimate of spherecity.
909
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

What does the results show in terms of Greenhouse-Geisser correction and Huynh-Fedt..? - (3)

A
  • Observed F statistic not significant using Greenhouse-Geisser ( p> 0.05)
  • Greenhouse-Geisser is quite conservative and miss true effects that exist
  • Thus, Huynh-Feldt showend F-statistic is still significant as p-value of 0.048
910
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

What happens if Greenhouse Geisser is not-significant (p>0.05) and Huynh-Feldt is significant in this example? - (2)

A
  • Taking average of two significant values e.g., 0.063+ 0.048/2 = 0.056
  • Thus, go with Greenhouse-Geisser correction and conclude F ratio is non-significant
911
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

What happens if two corrections - Greenhouse and Felt give same conclusion then you can choose which one to

A

report

912
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

Important to use valid critical value of F - choosing which p-value to report as it potentially makes a difference between making a

A

Type 1 error (False positive) or not

913
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

what does this summary table of repeated contrasts show? - (3)

Level 1 vs 2 is stick insect vs kangaroo testicle
Level 2 vs 3 is kangaroo testicle vs fish eyeball
Level 3 vs 4 is fish eyeball vs witchetty grub

A
  • celebrities took significantly longer to retch after
    eating the stick insect compared to the kangaroo testicle (Level 1 vs. Level 2) - p-value of 0.002
  • Time taken to retch was not significantly different in Level 2 vs 3 and Level 3 vs 4
914
Q

Researcher measures mean time taken for celebrities to retch for each animal (sticky insect, kangaroo testicle, fish eye, witchey grub) - one-way repeated ANOVA

If main effect is not significant in main ANOVA table for this data then significant contrasts in table below should be … but if MANOVA was significant then… - (2)

A

ignored

inclined to conclude main effects of animal was significant and proceed with further tetss like contrasts

915
Q

What IV, DV , design and test to use for this research scenario? - (4)

A
  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
916
Q

What does LSD correction (post-hoc option in SPSS)

A

does not actually make any adjustments to p-value in terms of critical value as what post-hoc test should do

917
Q

What does output show? - (3)

  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
A
  • sales are increasing across the weeks
  • Week 1 start at 427.93 and gradually rise by week 4 to 642,28 pounds
  • looks like incentives are having an effect and seem to generate higher sales
918
Q

What does this output show in terms of Maulchys Test of Spherecity? - (2)

  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
A
  • P-value is not significant ( p = 0.080)
  • Assumption of spherecity is satisfied so we got equal variances between differences across conditions
919
Q

If Maulchy’s test of spherecity is not significant in one-way repeated ANOVA, then which line do we use in main ANOVA table?

A
920
Q

If Maulchy’s test of spherecity is significant in one-way repeated ANOVA, then which line do we use in main ANOVA table?

A
921
Q

What does this main ANOVA table show? - (3)

  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
A
  • DF for week is 3 and 57 (spherecity assumed from week and error)
  • Week: F(3,57) = 26.30, p < 0.001 (p = 0.000), eta-squared is 0.58 - large effect
    • There is an overall effect going on and change across weeks
922
Q

What does this Sidmak correction table and table of means show you in this output? - (6)

  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
A
  • No sig difference betwen W1 and W2
  • Sig difference between W1 and W3 = ihigher sales in W3 (538.570) compared to W1 (427.933)
  • Sig difference between W1 and W4 = ihigher sales in W3 (642.284) compared to W1 (427.933)
    *Not sig diff with W2 and W3
  • Sig difference between W2 and W4 , higher sales in W4 (642.284) than W2 (481.388)
  • Sig difference between W3 and W4 , higher sales in W4 (642.284) than W3 (538.570)
923
Q

What does this output show in terms of repeated contrasts? - (3)

  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
A
    • Did sales increase from W1 to W2? = p = 0.010 significant
  • Did sales increase from W2 to W3? = p = 0.030
  • Did sales increase from W3 to W4? = p = 0.008
924
Q

What happens if post hoc and contrasts are telling a different story? - contrasts says weekly increase e.g. W1 to W2 increase, W2 to W3 increase , W3 to W4 increase but post-hoc W1 to W3 was increased sig, W1 to W4 was sig increase but W2 to W3 was not - (2)

  • Repeated measures design
  • One IV (Incentive) , four conditions (week 1, week 2, week 3, week 4)
  • One DV (Sales Generated)
  • One-way repeated ANOVA
A
  • Post hoc has lack of power due to many multiple comparisons
  • By limiting comparisons in contradt we get around problem
925
Q

Diagram of writing up one-way repeated ANOVA

A
926
Q

Two-way repeated ANOVA involves

A

more than one IV

927
Q

What does four-way ANOVA mean?

A

4 different IV

928
Q

What does 2x3 ANOVA means? - (2)

A
  • IV with 3 levels
  • IV with 2 levels
929
Q

What design, IV, DV and test would you to to investigate the follow scenario? - (4)

A
  • Repeated measures design
  • Two IVs: alcohol (3 conditions) and sleep (2 conditions)
  • DV: Reaction Times
  • Two-way repeated measures ANOVA
930
Q

What does this two-way repeated ANOVA SPSS output show? - (2)

  • Repeated measures design
  • Two IVs: alcohol (3 conditions) and sleep (2 conditions)
  • DV: Reaction Times
  • Two-way repeated measures ANOVA
A
  • large number for RT means slower RT
  • Alcohol seem to have an effect on RT but particularly for 2 pints + no sleep
931
Q

What does this two-way repeated ANOVA SPSS output show for Mauchly’s Test of Sphericity? - (2)

  • Repeated measures design
  • Two IVs: alcohol (3 conditions) and sleep (2 conditions)
  • DV: Reaction Times
  • Two-way repeated measures ANOVA
A
  • Two p-values: alcohol ( p = 0.00) and alcohol * sleep [ interaction effect] (p = 0.00) – > sig so assumption of spherecity is violated so report Grenhouse-Geisser values from main ANOVA table
  • No p-value for sleep as only 2 conditions and test of sphericity need more than 2
932
Q

What does this two-way repeated ANOVA main table show? - (3)

  • Repeated measures design
  • Two IVs: alcohol (3 conditions) and sleep (2 conditions)
  • DV: Reaction Times
  • Two-way repeated measures ANOVA
  • Error DF was 38.
  • Test of Spherecity was sig –> assumption violated
A
  • Main sig effect of alcohol: F(1.16,22.06) = 51.38, p < 0.001, partial eta-squared = 0.73
  • Main sig effect of sleep: F(1,19) = 88.61, p < 0.001, partial-eta-squared = 0.82
  • Interaction effect: F(1.15,21.91) = 23.36, p < 0.001, partial-eta squared = 0.55
933
Q

What does this two-way repeated ANOVA output show in post hocs? - Sidmak correction - (4)

  • Repeated measures design
  • Two IVs: alcohol (3 conditions) and sleep (2 conditions)
  • DV: Reaction Times
  • Two-way repeated measures ANOVA
A
  • condition 1 and condition 2 which was significant
  • Condition 1 vs 3 which was significant
  • Condition 2 with Condition 3 was significant
  • So all groups differing significantly from each other so interpret from that higher does of alcohol has more impact on RT
934
Q

What does this two-way repeated ANOVA interaction plot show? - (3)

  • Repeated measures design
  • Two IVs: alcohol (3 conditions) and sleep (2 conditions)
  • DV: Reaction Times
  • Two-way repeated measures ANOVA
A
  • Interaction effect is there = as line continue they cross
  • Most pronouned effect was in alcohol grp 3 (2 pints)
  • When alcohol grp 3 had full nights sleep (2), impairs their RT very slightly
  • When alcohol grp 3 had sleep deprivation (1) in combination with 2 pints, it impairs RT by a lot –> use simple effect analysis as well as two-way independent ANVOA to see if difference in grp 3 of blue and green line is sig
935
Q

What happens when assumptions vilated in repeated-measures ANOVA? - (2)

A

Can do non-parametric test called Friedman’s ANOVA if only one IV

There is no non-parametric counterpart for more than one IV in repeated design

936
Q

Assumption of repeated measures ANOVA - (3)

A
  1. Normal distribution
  2. Repeated measures design (same participants)
  3. Sphereicity - Mauchly’s test
937
Q

What does significant Mauchly’s test signify in repeated measures? - (2)

A

A significant effect means that corrections need to be made later on

Those corrections are listed in the main ANOVA output table

938
Q

What is decision tree for two-way repeated ANOVA

A

1 DV continous and 2 or more categorical predictors with 2 or more levels with same participants in each predictor level

939
Q

What is decision tree for one-way repeated ANOVA? - (3)

A

1 DV continous

1 Predictor categorical with more than 2 levels

Same participants in each predictor level

940
Q

Just like independent measure designs there can be more than one categorical predictor.

When all participants take part in all combinations of those predictors, we have a repeated measures factorial design and can use an ANOVA to test for

A

significant main effects and interactions

941
Q

Example of two-way repeated ANOVA - (3)

A

The variables are the type of drink (Beer - Wine - Water) and the type of imagery used in the advertisement (positive - negative - neutral)

The outcome is how much the participant likes the beverage on a scale from -100 (dislike very much) to 100 (like very much)

Participants went two conditions

942
Q

Equation of variance

A
943
Q

What is mixed design? - (2)

A

A mixture of between-subject and within-subject

Several independent variables or predictors have been measured; some have been measured with different entities, (pps) whereas others used the same entities (pps)

944
Q

You will need at least two IVs for

A

mixed design

945
Q

What is decision tree for mixed design ANOVA? - (7)

A

Q: What sort of measurement? A: Continuous
Q:How many predictor variables? A: Two or more
Q: What type of predictor variable? A: Categorical
Q: How many levels of the categorical predictor? A: Not relevant
Q: Same or Different participants for each predictor level? A: Both
This leads us to and Factorial mixed ANOVA

946
Q

Example of mixed design scenario for ANOVA - (2)

A

a mixed ANOVA is often used in studies where you have measured a dependent variable (e.g., “back pain” or “salary”) over two or more time points or when all subjects have undergone two or more conditions (i.e., where “time” or “conditions” are your “within-subjects” factor),

but also measure DV when your subjects have been assigned into two or more separate groups (e.g., based on some characteristic, such as subjects’ “gender” or “educational level”, or when they have undergone different interventions). These groups form your “between-subjects” factor.

947
Q

An organizational psychologist is hired as a consultant by a person planning to open a coffee house for college students. The coffee house owner wants to know if her customers will drink more coffee depending on the ambience of the coffee house. To test this, the psychologist sets up three similar rooms, each with its own theme (Tropical; Old Library; or New York Café ) then arranges to have thirty students spend an afternoon in each room while being allowed to drink all the coffee they like. (The order in which they sit in the rooms is counterbalanced.) The amount each participant drinks is recorded for each of the three themes.

  1. Independent variable(s)
  2. Is there more than 1 IV?
  3. The levels the independent variable(s)
  4. Dependent variable
  5. Between (BS) or within-subjects (WS)?
  6. What type of design is being used?
A

Theme

No

Tropical, Old Library,
New York Café

Amount of coffee consumed

Within-subjects

1-way Repeated measures

948
Q

A manager at a retail store in the mall wants to increase profit. The manager wants to see if the store’s layout (one main circular path vs. a grid system of paths) influences how much money is spent depending on whether there is a sale. The belief is that when there is a sale customers like a grid layout, while customers prefer a circular layout when there is no sale. Over two days the manager alternates the store layout, and has the same group of customers come each day. Based on random assignment, half of the customers told there is a sale (20 % will be taken off the final purchases), while the other half is told there is no sale. At the end of each day, the manager calculates the profit.

  1. Independent variable(s)
  2. Is there more than 1 IV?
  3. The levels the independent variable(s)
  4. Dependent variable
  5. Between (BS) or within-subjects (WS)?
  6. What type of design is being used?
A

Sale/ No Sale, Store’s layout

Yes

Sale-No Sale, Grid-Circular

Profit

BS (Sale) and WS (Layout)

2-way mixed Measures

949
Q

A researcher at a drug treatment center wanted to determine the best combination of treatments that would lead to more substance free days. This researcher believed there were two key factors in helping drug addiction: type of treatment and type of counseling. The researcher was interested in either residential or outpatient treatment programs; and either cognitive-behavioral, psychodynamic, or client-centered counseling approaches. As new clients enrolled at the center they were randomly assigned to one of six experimental groups. After 3 months of treatment, each client’s symptoms were measured.

  1. Independent variable(s)
  2. Is there more than 1 IV?
  3. The levels the independent variable(s)
  4. Dependent variable
  5. Between (BS) or within-subjects (WS)?
  6. What type of design is being used?
A

Type of treatment, Type of counseling.

Yes

Residential or outpatient/ cognitive-behavioural, psychodynamic or client-centered.

Substance-free days

Between subjects

2-way independent measures ANOVA.

950
Q

Assumptionsof mixed ANOVA - (3)

A

Normal Distribution

Independent and Repeated Factors

Homogeneity of Variance for the Independent factor
+
Sphericity for the Repeated factor

951
Q

Assumptions of repeated-measures ANOVA -(3)

A

Normal Distribution

Repeated Measure Design (same participants)

Sphericity (Mauchly’s Test)

952
Q

Assumptions of independent ANOVA - (3)

A

Normal Distribution

Independence of Scores

Homogeneity of Variance (Levene’s Test)

953
Q

Leven’s test tests if the variances in independent groups are similar, would levene’s test be significant in this case?

A

Levene’s test would likely be significant as the variance between the two groups are quite different.

954
Q

Sphereicity is an assumption of both

A

repeated and mixed models

955
Q

If p-value significant for checking for spherecity ten - (3)

A

If GG < 0.75 THEN USE GG

IF GG > 0.75 THEN USE HF

Since GG less than 0.75 report adjusted F, DF and sig which is F(1.24, 21.00) = 212.32 , p < 0.001

956
Q

Homogenity of variance is

A

distribution of groups are similar?

957
Q

Spherecity is asking are the disttibution of differences between groups are

A

similar

958
Q

The researcher hypothesized that there would be an interaction between dog breed (Collie or German Shepherd) and week of obedience school training (all dogs measured at 1 week and 5 weeks) as they relate to the number of times the dog growls per week. Specifically, it was hypothesized that Collies would show no difference in growls between 1 week and 5 weeks, but German Shepherds would growl less at 5 weeks than at 1 week.

  1. Independent variable(s)
  2. Is there more than 1 IV?
  3. The levels the independent variable(s)
  4. Dependent variable
  5. Between(BS) or within-subjects (WS)?
  6. What type of design is being used?
A
  1. Dog breed and measurement time
  2. Yes
  3. Collie-German Shepard/Week 1-Week 5
  4. Number of growls
  5. Dog breed Between and measurement time is within
  6. 2-WAY mixed ANOVA
959
Q

What does this 2-way mixed ANOVA show? - (3)

  1. Independent variable(s)
  2. Is there more than 1 IV?
  3. The levels the independent variable(s)
  4. Dependent variable
  5. Between(BS) or within-subjects (WS)?
  6. What type of design is being used?
A

1) is there an effect overall = Yes (green)

2) Is the effect bread = Yes (red)

3) Is there an interaction = Yes (blue)

960
Q

Partioning of variance of one way vs two way independent

A
961
Q

Rules of contrast coding - (5)

A

Rule 1: Groups coded with positive weights compared to groups coded with negative weights.

Rule 2: The sum of weights for a comparison should be zero.

Rule 3: For a given contrast, the weights assigned to the group(s) in one chunk of variation should be equal to the number of groups in the opposite chunk of variation.

Rule 4: If a group is not involved in a comparison, assign it a weight of zero

Rule 5: If a group is singled out in a comparison, then that group should not be used in any subsequent contrasts.

962
Q

Contrast coding example SPSS how to read

A
963
Q

When conducting a Repeated-Measures ANOVA, which of the following assumptions is NOT relevant?

A.Independent residuals
B.Homogeneity of variance
C.Sphericity
D.They are all relevant

A

B

964
Q

One advantage of repeated measures designs over independent designs is that we are able to calculate a degree of error for each effect, whereas in an independent design we are able to calculate only one degree of error: true or false?

True or False

A

True

965
Q

An experiment was conducted to see how people with eating disorders differ in their need to exert control in different domains. Participants were classified as not having an eating disorder (control), as having anorexia nervosa (anorexic), or as having bulimia nervosa (bulimic). Each participant underwent an experiment that indicated how much they felt the need to exert control in three domains: eating, friendships and the physical world (this final category was a control domain in which the need to have control over things like gravity or the weather was assessed). So all participants gave three responses in the form of a mean reaction time; a low reaction time meant that the person did feel the need to exert control in that domain. The variables have been labelled as group (control, anorexic, or bulimic) and domain (food, friends, or physical laws). Of the following options, which analysis should be conducted?

A. Analysis of covariance
B. Two-way repeated measures ANOVA

C. Two-way mixed ANOVA

D. Three-way independent ANOVA

A

C

Two IVs = Group (Control, Anroexic, Bullimic) and Domain (Food, Friends, Physical Laws)

Group is between

Each partiicpant underwent domains so within

DV = Participannts measured

966
Q

An experiment was done to compare the effect of having a conversation via a hands-free mobile phone, having a conversation with an in-car passenger, and no distraction (baseline) on driving accuracy. Twenty participants from two different age groups (18–25 years and 26–40 years) took part. All participants in both age groups took part in all three conditions of the experiment (in counterbalanced order), and their driving accuracy was measured by a layperson who remained unaware of the experimental hypothesis.

How do we interpret the main effect of distraction from the SPSS table (next slide)? - (2)

A

The assumption of sphericity has been met, indicated by Mauchly’s test (p > .05).

There was a significant main effect of distraction (F(2, 36) = 45.95, p < .001). This effect tells us that if we ignore the effect of age, driving accuracy was significantly different in at least two of the distraction groups.

967
Q

Two-way repeated-measures ANOVA compares:

A. Several means when there are two independent variables, and the same entities have been used in all conditions

B. Two means when there are more than two independent variables, and the same entities have been used in all conditions.

C. Several means when there are two independent variables, and the same entities have been used in some of the conditions.

D. Several means when there are more than two independent variables, and some have been manipulated using the same entities and others have used different entities.

A

A

968
Q

When conducting a repeated-measures ANOVA which of the following assumptions is not relevant?

A. Homogeneity of variance

B. Sphericity

C. Independent residuals

D. They are all relevant

A

A

969
Q

The table shows hypothetical data from 3 conditions

For these data, spherecity will hold when

(Hint: Sphericity refers to the equality of variances of the differences between treatment levels.)

A.The variances of the differences between treatment levels are roughly equal

B. The variance of each condition is roughly equal

C. The variance of each condition is not equal

D. The variances of the differences between treatment levels are not equal

A

A

970
Q

Imagine we were interested in the effect of supporters singing on the number of goals scored by soccer teams. We took 10 groups of supporters of 10 different soccer teams and asked them to attend three home games, one at which they were instructed to sing in support of their team (e.g., ‘Come on, you Reds!’), one at which they were instructed to sing negative songs towards the opposition (e.g., ‘You’re getting sacked in the morning!’) and one at which they were instructed to sit quietly. The order of chanting was counterbalanced across groups. Looking at the output below, which of the following sentences is correct?#

A.The results showed that the number of goals scored was significantly affected by the type of singing from the supporters, F(2, 18) = 11.24, p = .001.

B. The results showed that the number of goals scored was significantly affected by the type of singing from the supporters, F(1.58, 14.19) = 11.24, p = .002.

C. The results showed that the number of goals scored was significantly affected by the type of singing from the supporters, F(2, 12.4) = 11.24, p = .001.

D. The results showed that the number of goals scored was significantly higher when supporters sang positive songs towards their team than when they sat quietly, F(2, 18) = 11.24, p = .001.

A

A = Mauchly’s test was non-significant, so we can report the result in the row labelled ‘sphericity assumed’

971
Q

Imagine we were interested in the effect of supporters singing on the number of goals scored by soccer teams. We took 10 groups of supporters of 10 different soccer teams and asked them to attend three home games, one at which they were instructed to sing in support of their team (e.g., ‘Come on, you Reds!’), one at which they were instructed to sing negative songs towards the opposition (e.g., ‘You’re getting sacked in the morning!’) and one at which they were instructed to sit quietly. The order of chanting was counterbalanced across groups. An ANOVA with a simple contrasts using the last category as a reference was conducted. Looking at the output tables below, which of the following sentences regarding the contrasts is correct?

a.The first contrast revealed that soccer teams scored significantly more goals when their supporters sang positive songs compared to when they did not sing. The second contrast revealed that soccer teams scored significantly fewer goals when their supporters sang negative songs compared to when they did not sing.

b. The first contrast revealed that soccer teams scored significantly fewer goals when their supporters did not sing compared to when they sang negative songs. The second contrast revealed that soccer teams scored a similar amount of goals when their supporters sang positive songs compared to when they did not sing.

c. The first contrast revealed that soccer teams scored significantly more goals when their supporters sang positive songs compared to when they did not sing. The second contrast revealed that soccer teams scored significantly fewer goals when their supporters sang negative songs compared to when they sang positive songs.

d. The first contrast revealed that soccer teams scored significantly more goals when their supporters sang positive songs compared to when they did not sing. The second contrast revealed that soccer teams did not significantly differ in the number of goals scored when their supporters sang negative songs compared to when they did not sing.

A

a = see from the means in the Descriptive Statistics table that positive singing resulted in the highest number of goals scored and negative singing resulted in the least number of goals score

972
Q

An experiment was done to compare the effect of having a conversation via a hands-free mobile phone, having a conversation with an in-car passenger, and no distraction (baseline) on driving accuracy. Twenty participants from two different age groups (18–25 years and 26–40 years) took part. All participants in both age groups took part in all three conditions of the experiment (in counterbalanced order), and their driving accuracy was measured by a layperson who remained unaware of the experimental hypothesis.

Which of the following sentences is the correct interpretation of the main effect of distraction?

AThere was a significant main effect of distraction, F(2, 36) = 45.95, p < .001. This effect tells us that if we ignore the effect of age, driving accuracy was significantly different in at least two of the distraction groups.

B. There was no significant main effect of distraction, F(2, 36) = 45.95, p = .719. This effect tells us that if we ignore the effect of age, driving accuracy was the same for no distraction, hands-free conversation and in-car passenger conversation.

C. There was a significant main effect of distraction, F(2, 36) = 45.95, p < .001. This effect tells us that driving accuracy was different for no distraction, hands-free conversation and in-car passenger conversation in the two age groups.

D. There was no significant main effect of distraction, F(2, 36) = 45.95, p > .05. This effect tells us that none of the distraction groups significantly distracted participants across both age groups.

A

A = We can read the results in the row labelled ‘sphericity assumed’, as we can see from the output of Mauchly’s test that the assumption of sphericity has been met, p > .05. However, we would need to do some follow-up tests to investigate exactly where the differences between groups lie

973
Q

Field and Lawson (2003) reported the effects of giving children aged 7–9 years positive, negative or no information about novel animals (Australian marsupials). This variable was called ‘Infotype’. The gender of the child was also examined. The outcome was the time taken for the children to put their hand in a box in which they believed either the positive, negative, or no information animal was housed (positive values = longer than average approach times, negative values = shorter than average approach times). Based on the output below, what could you conclude?

A. Approach times were significantly different for the boxes containing the different animals, but the pattern of results was unaffected by gender.

B. Approach times were significantly different for the boxes containing the different animals, and the pattern of results was affected by gender.

C. Approach times were not significantly different for the boxes containing the different animals, but the pattern of results was affected by gender.

D.Approach times were not significantly different for the boxes containing the different animals, but the pattern of results was unaffected by gender.

A

A

974
Q

What leads to chi-squared test?

A

Q: What sort of measurement? A: Categorical (in this case counts or frequencies)
Q:How many predictor variables? A: One
Q: What type of predictor variable? A: Categorical
Q: How many levels of the categorical predictor? A: Not relevant
Q: Same or Different participants for each predictor level? A: Different
This leads us to and Chi-square test for independence of groups

975
Q

In chi-square test, participants is allocated to one and only one category such as - (3)

A

pass or fail,

pregnant or not pregnant,

win, draw or lose

976
Q

Since each participant is allocated to one category in chi-squared test each individual therefore

A

contributes to the frequency or count with which a category occurs

977
Q

Table scenario in which cats can be trained to dance more effectively with food or affection at reward - chi-squared test

A
978
Q

Table scenario in which cats can be trained to dance more effectively with food or affection at reward - chi-squared test

what are the four categories? - (4)

A
  • could they dance - yes
  • could they dance - no
  • food as reward
  • affection as reward
979
Q

Table scenario in which cats can be trained to dance more effectively with food or affection at reward - chi-squared test

highlight the frequencies for four categories

A
980
Q

Table scenario in which cats can be trained to dance more effectively with food or affection at reward - chi-squared test

what do the rows give?

A

Row totals give frequencies of dancing and non-dancing cats

981
Q

Table scenario in which cats can be trained to dance more effectively with food or affection at reward - chi-squared test

what do the columns give? - (2)

A

The column totals give frequencies of food and affection as reward

These are the numbers in each group

982
Q

IV and DV in chi-squared tests - (2)

A

One categorical DV (because of frequencies)

with one categorical IV with different participants at each predictor level

983
Q

In chi-squared categorical outcomes, the null hypothesis is set

A

up on the basis of expected frequencies, four all four variable combinations, based on the idea that the variable of interest has no effect on frequencies

984
Q

What does the chi-square tests?

A

whether there is a relationship between two categorical variables.

985
Q

In chi-square since we are using categorical variables we can not use

A

mean or any similar statistic hence cannot use any parametric tests

986
Q

What does chi-square compare?

A

observed frequencies from the data with frequencies which would be expected if there was no relationship between the two variables.

987
Q

In chi-square test when measuring categorical variables we are interested in

A

frequencies (number of items that fall into combination of categories)

988
Q

Example of scenario using chi-square

A

We have a list of movie genres; this is our first variable. Our second variable is whether or not the patrons of those genres bought snacks at the theatre. Our idea (or, in statistical terms, our null hypothesis) is that the type of movie and whether or not people bought snacks are unrelated. The owner of the movie theatre wants to estimate how many snacks to buy. If movie type and snack purchases are unrelated, estimating will be simpler than if the movie types impact snack sales.

989
Q

What is assumptions of chi-square test? - (3)

A

Data values that are a simple random sample from the population of interest.

Two categorical or nominal variables. Don’t use the independence test with continuous variables that define the category combinations. However, the counts for the combinations of the two categorical variables will be continuous.

For each combination of the levels of the two variables, we need at least five expected values. When we have fewer than five for any one combination, the test results are not reliable

990
Q

We have a list of movie genres; this is our first variable. Our second variable is whether or not the patrons of those genres bought snacks at the theatre. Our idea (or, in statistical terms, our null hypothesis) is that the type of movie and whether or not people bought snacks are unrelated. The owner of the movie theatre wants to estimate how many snacks to buy. If movie type and snack purchases are unrelated, estimating will be simpler than if the movie types impact snack sales.

Is the Chi-square test of independence an appropriate method to evaluate the relationship between movie type and snack purchases? - (3)

A

We have a simple random sample of 600 people who saw a movie at our theatre. We meet this requirement.

Our variables are the movie type and whether or not snacks were purchased. Both variables are categorical.

But last requirement is for more than five expected values for each combination of the two variables. To confirm this, we need to know the total counts for each type of movie and the total counts for whether snacks were bought or not. = check later

991
Q

We have a list of movie genres; this is our first variable. Our second variable is whether or not the patrons of those genres bought snacks at the theatre. Our idea (or, in statistical terms, our null hypothesis) is that the type of movie and whether or not people bought snacks are unrelated. The owner of the movie theatre wants to estimate how many snacks to buy. If movie type and snack purchases are unrelated, estimating will be simpler than if the movie types impact snack sales.

Diagram of contigency table in Chi-square and calculating row totals and colum and grand total - (7)

A

50 + 125 + 90 +45 = 310

75 + 175 + 30 + 10 = 290

50 + 75 = 125

125 + 175 = 300

90 + 30 = 120

45 + 10 = 55

310 + 290 = 600

992
Q

How to calculate chi-square test statistic? - (4)

A
  1. Calculate the difference from actual and expected for each Movie-Snacks combination.
  2. square that difference.
  3. Divide by the expected value for the combination.
  4. We add up these values for each Movie-Snacks combination. This gives us our test statistic.
993
Q

We have a list of movie genres; this is our first variable. Our second variable is whether or not the patrons of those genres bought snacks at the theatre. Our idea (or, in statistical terms, our null hypothesis) is that the type of movie and whether or not people bought snacks are unrelated. The owner of the movie theatre wants to estimate how many snacks to buy. If movie type and snack purchases are unrelated, estimating will be simpler than if the movie types impact snack sales.

Diagram of contigency table in Chi-square of calculating eexpected counts

A

e.g., for action and snacks it would be column total (310) * row total (125) divided by grand total of 600 = 65

994
Q

Example of calculating chi-square from table

A

For this it would be 65.03

995
Q

Does the area of psychology that a person prefers depend on whether they would select a cat or a dog as a pet? - chi-square test of independence

Chi-square example we need to check the assumptions below - (2)

A

Independence
Each item or entity contributes to only one cell of the contingency table.

The expected frequencies should be greater than 5.
In larger contingency tables up to 20% of expected frequencies can be below 5, but there is a loss of statistical power.
Even in larger contingency tables no expected frequencies should be below 1.

996
Q

How to understand your test statistic from chi-squared? - (5) if you have test statistic of 65.03

A
  1. Set your significance level = .05
  2. Calculate the test statistic -> 65.03
  3. Find your critical value from chi-squared distribution table based on df & significance level
  4. Degrees of freedom: df (r – 1) x (C-1)
    For the movie example this is; Df = (4-1) x (2-1) = 3 -> 7.815
  5. compare test statistic with critical level
    65.03 > 7.82 so reject the idea that movie type and snack purchases are independent
997
Q

Example of research question and hypothesis and sig level of chi-square test of independence- (4)

A

Research question:
Does the area of psychology that a person prefers depend on whether they would select a cat or a dog as a pet?

Hypotheses:
H0: The area of interest in psychology and type of pet preferred are independent of each other.
H1: The area of interest in psychology and type of pet preferred are not independent of each other. That is the primary area of interest in psychology depends on whether you prefer a cat or a dog.
Significance level: α = .05

998
Q

Does the area of psychology that a person prefers depend on whether they would select a cat or a dog as a pet? - chi-square test of independence

Chi-square example we need to check the assumptions of The expected frequencies should be greater than 5.

What does it show? - (4)

A

Here we see that all the expected counts in the cat group and one expected count in the dog group are below 5.

We also have one in the cat group that is below 1.

So, SPSS has flagged that we have 60% of the expected counts falling below 5.

So assmption of expected frequencies greater than 5 is not assumed

999
Q

If chi-square assumption that The expected frequencies should be greater than 5 is not satisfied then do - chi-square test of independence

A

We should use Fisher’s Exact Test which can correct for this.

1000
Q

Does the area of psychology that a person prefers depend on whether they would select a cat or a dog as a pet? - - chi-square test of independence

If assumptions were met (expected frequencies greater than 5) then.. report - (2)

A

A chi-square independence test was performed to examine whether there was a relationship between their area of studies in psychology and their preference for cats or dogs.

The relationship between these variables was not significant, χ²(4, N = 46) = 1.46, p = .834, so we fail to reject H0.

1001
Q

Are directional hypotheses possible with chi-square?

A.Yes, but only when you have a 2 × 2 design.
B.Yes, but only when there are 12 or more degrees of freedom.
C.Directional hypotheses are never possible with the chi-squared test.
D.Yes, but only when your sample is greater than 200.

A

A = only when you have 2 variables to compare and can’t do non-directional in chi-square have to use loglinear or goodness of fit tests

1002
Q

Example situations you can do chi-square directional and not possible - (5)

A

If we are just comparing pet preferences between males and females, we can make a directional hypothesis (2 x 2 – male/female, cats/dogs).

Males prefer cats or females prefer dogs.

However, when we start adding variables to the design it gets complicated.

If we wanted to compare drink preferences at different times of the day for students/lecturers, we couldn’t form a directional hypothesis.

This is because we have 3 main effects and several interactions to consider. We need to use loglinear analyses to do this.

1003
Q

Loglinear analysis is a …. of chi-square

A

extension

1004
Q

Chi-square only analyses two variables at a time, whilst log-linear models

A

can determine complex interactions in multidimensional contingency tables with more than two categorical variables.

1005
Q

Loglinear is appropriate when

A

there’s no clear distinction between response and explanatory variables

1006
Q

think of

A

Think of chi-square like t-tests (2 groups) and log-linear like ANOVA (more than 2 groups).

1007
Q

Example of RQ, hypothesis and sig level of loglinear - (3)

A

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Hypotheses:
H0: Treatment, type of animal and improvements are independent of each other.
H1: Treatment, type of animal and improvements are associated with each other.

Significance level: α = .05

1008
Q

Assumptions of log linear - (2)

A

Independence
Expected counts > 5

1009
Q

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Checking assumption of expected counts - (3):

A

Here we have 3 things we are comparing: animal (cat/dog), treatment (yes/no) and improvement (yes/no) all of which are categorical.

We look and see that all of the expected counts are above 5.

So met assumption of independence and expected counts

1010
Q

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Here we have 3 things we are comparing: animal (cat/dog), treatment (yes/no) and improvement (yes/no) all of which are categorical.

In loglinear model selection it begins with - (2)

A

all terms present (all main effects and all possible interactions

main effects: Animal, Treatment and Improvement

interactions: Animal * Treatment, Animal * Improvement, Treatment * Improvement and Treatment* Animal* Improvement

1011
Q

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Here we have 3 things we are comparing: animal (cat/dog), treatment (yes/no) and improvement (yes/no) all of which are categorical.

In loglinear model selection after including all main effects and interactions then it - (4)

A

Remove a term and compares the new model with the one in which the term was present.

Starts with the highest-order interaction (including max number of variables/categories)

Uses the likelihood ratio to ‘compare’ models below:

If the new model is no worse than the old, then the term is removed and the next highest-order interactions are examined, and so on.

1012
Q

Model selection of loglinear - what does it show?

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Here we have 3 things we are comparing: animal (cat/dog), treatment (yes/no) and improvement (yes/no) all of which are categorical. - (3)

A

We can see that the model selection worked in a way that it first tried to remove the 3-way interaction.

However, we can see here that it * affected the fit of the model, so it was left in.

Since removing the highest-order interaction made a * difference to the fit of the model, we get a final model that is the saturated model (it contains all main effects and interactions).

1013
Q

Loglinear SPSS K way and Higher order effects what does it show?

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Here we have 3 things we are comparing: animal (cat/dog), treatment (yes/no) and improvement (yes/no) all of which are categorical. - (2)

A

we are using the likelihood ratio here because that’s how we compare the models to find the best fit

. We see that all main effects and interactions are significantly contributing to explaining the variance in the data

1014
Q

loglinear what does K represent and what does K = 1,2 and 3 represent? - (4)

A

K represents the level of the terms.

For example, K=1 would be the main effects,

K=2 would be our 2-way interactions and

K=3 is our 3-way interaction.

1015
Q

Loglinear SPSS - what does parameter estimates show?

Research question: Is the new treatment associated with improvements in health in cats and dogs?

Here we have 3 things we are comparing: animal (cat/dog), treatment (yes/no) and improvement (yes/no) all of which are categorical. - (3)

A

There is a significant three-way interaction between animal, treatment and improvement, as well as two significant two-way interactions between animal and improvement and treatment and improvement (p < .001)

a * 3-way interaction between animal, treatment and improvement as well as two * 2-way interaction between animal/improvement and treatment/improvement.

Like our post-hoc tests, this is telling us where the * differences are.

1016
Q

Loglinear after seeing statistical tests we go to raw data showing that…

A

Based on the raw data, there seems to be indication that the cats responded better to treatment than dogs, this should be followed up by chi-square tests separately for cats and dogs to determine whether the association between treatment and improvement is present in both cats and dogs

1017
Q

When conducting a loglinear analysis, if our model is a good fit of the data then the goodness-of-fit statistic for the final model should be:

A. Significant (p should be smaller than .05)

B. Non-significant (p should be bigger than .05)

C. Less than 5 but greater than 1

D. Greater than 5

A

B

1018
Q

The goodness of fit tests in log linear tests

A

hypothesis that frequencies predicted by model (expected frequencies) are sig different from actual frequencies in data (obsevered)

1019
Q

A significant goodness of fit result mean

A

our model was significantly different from our data (i.e., the model is a bad fit to the data).

1020
Q

A recent story in the media has claimed that women who eat breakfast every day are more likely to have boy babies than girl babies. Imagine you conducted a study to investigate this in women from two different age groups (18–30 and 31–43 years).

Looking at the output tables below, which of the following sentences best describes the results? = chi-square

A. Women who ate breakfast were significantly more likely to give birth to baby boys than girls.
B. There was a significant two-way interaction between eating breakfast and age group of the mother.
C. Whether or not a woman eats breakfast significantly affects the gender of her baby at any age.
D. The model is a poor fit of the data.

A

C

1021
Q

Chi square and log linear are both

A

non-parametric methods

1022
Q

Non-parametric tests used when

A

When data violate the assumptions of parametric tests we can sometimes find a nonparametric equivalent
eg. normality of distribution

1023
Q

Non-parametric tests work on the principle of

A

randomization or ranking the data for each group

1024
Q

Ranking data gets rid of in non parametric

A

outliers and skew

1025
Q

How does ranking work in non-parametric? - (2)

A

Add up the ranks for the two groups and take the lowest of these sums to be our test statistic

The analysis is carried out on the ranks rather than the
actual data.

1026
Q

Non-parametric equivalent of independent/unrelated t-tests

A

Mann-Whitney or Wilcoxon
rank-sum test

1027
Q

Non-parametric equivalent of repeated t-test

A

Wilcoxon signed-rank test

1028
Q

Non-parametric equivalent of : One-way independent (between-subjects) ANOVA

A

Kruskall-Wallis or (for trends)
Jonckheere-Terpstra

1029
Q

Non-parametric equivalent of one-way repeated ANOVA

A

Friedmanʼs ANOVA

1030
Q

Non-parametric equivalent of Multi-way between or
within-subjects ANOVA

A

Loglinear analysis (categorical
outcome, with participants as a factor)

1031
Q

Non-parametric equivalent of correlation

A

Spearman’s Rho or Kendall’s Tau

1032
Q

Mann-Whitney/Wilcoxon rank-sum Test - Compares

A

two independent groups of scores

1033
Q

Wilcoxon signed rank Test - Compare

A

two dependent groups of scores

1034
Q

Kruskal-Wallis Test - Compares

A

> 2 independent groups of scores

1035
Q

Friedman’s Test - Compares

A

> 2 dependent groups of scores

1036
Q

Spearman’s Rho & Kendall’s Tau - Measures the extent to which

A

two continuous variables are related (pattern of responses across variables)

1037
Q

Logic behind Wilcoxon’s rank sum test, what does SPSS do? - (3)

A

Step 1: Get some not normally distributed data

Step 2: Rank it (regardless of group)

Step 3: Significance testing
Does one of the groups have more of the higher ranking scores than the other?

1038
Q

What is DF of chi-square?

A

(r-1)(c-1)

1039
Q

The likelhood ratio in loglinear model preferred

A

small sample sizes

1040
Q

DF of likelhood ratio in loglinear

A

df = (r-1)(c-1)

1041
Q

Decision tree of Mann Whitney - (4)

A

1 DV = Ordinal (e.g., high school, bachelors, order is meaningful) or continous
1 IV = Categorical and 2 levels
Different partiicpants
Does not meet assumption of parametric

1042
Q

Wilcoxon rank sum and Man Whitney U is

A

same procedure and used to compare two independent groups and assess whether samples come from same distribution

1043
Q

For Mann-Whitney U/Wilcoxon Rank Sum they comparing 2 independent conditions the two steps - (2)

A

Rank all the data on the the basis of the scores irrespective of the group

compute the sum of ranks of each group

1044
Q

For wilcoxon rank sum, the statistic Ws is

A

the lower of the two sums of ranks

1045
Q

For Mann-Whitney, the statistic U use the

A

sum of ranks for group 1, R1, as follows

1046
Q

Example of table where comparing 2 independent conditions of Wilcoxon rank sum or Mann Whitney U test

A

Here we have data for two groups; one taking alcohol, the other ecstasy. The scores for a measure of depression. Scores were obtained on two days; Sunday and Wednesday. The drugs were administered on Saturday.

1047
Q

Example of table where comparing 2 independent conditions of Wilcoxon rank sum or Mann Whitney U test

Here we have data for two groups; one taking alcohol, the other ecstasy. The scores for a measure of depression. Scores were obtained on two days; Sunday and Wednesday. The drugs were administered on Saturday.

Two steps for both statistics:
Rank all the data on the the basis of the scores irrespective of the group
compute the sum of ranks of each group - (5)

A

The graphic here shows how we can list the scores in order and as a result assign each score a rank.

When scores tie, we give them the average of the ranks.

If we ensure we keep track of the group the scores came from we can relatively easily add the ranks up for each group.

Note, that if there was little difference between the groups the sums of their ranks would be similar, as they are for the data shown her for Sunday.

However, the sum of ranks differ considerably for the data obtained on Wednesday.

1048
Q

For Wilcoxon sum of ranks = comparing 2 independent groups the W s statistic,

the group sizes of n1 and n2 the mean of W2 is given:

A
1049
Q

For Wilcoxon sum of ranks = comparing 2 independent groups the W s statistic,

the standard error of Ws is given

A
1050
Q

For Wilcoxon sum of ranks = comparing 2 independent groups the z score of Ws can be calculated

A
1051
Q

For Mann Whitney , the statistic U use the the sum of ranks for group 1, R1, as follows

A
1052
Q

For Mann Whitney , the statistic U use the the sum of ranks for group 1, R1, as follows

Specificy equation - (2)

A

The first terms involving n1 and n2 actually compute the maximum possible sum of ranks for group 1.

U is zero when all those in group one have scores that exceed the scores of those in group 2.

1053
Q

In Mann Whitney U there is a standardised test statistic which is z score that can allow you to compute

A

effect size so r = z / square root of N (number of pps

1054
Q

What is decision tree of Wilcoxon signed rank test? - (4)

A

1 IV categorical with 2 levels

Same participants in each predictor level

1 DV - Ordinal or continous

Does not meet assumption of parametric tests

1055
Q

Steps of Wilcoxon signed rank test - (4)

A
  1. Compute the difference between scores for the two conditions
  2. Note the sign of the difference (positive or negative)
  3. Rank the differences ignoring the sign and also exclude any zero differences from the ranking
  4. Sum the ranks for positive and negative ranks
1056
Q

Example of Wilcoxon signed rank test carrying out steps - (9)

A

The table shown here has the Depression Scores taken on Sunday and Wednesday for those taking ecstasy on Saturday.

Data for Sunday are in the first column and Wednesday in the second column.

The third column shows the difference between scores obtained on Sunday and Wednesday.

NOte some could be negative, some positive. In this example however the difference is always positive apart from two values when the difference is zero.

The fourth column notes the sign of the difference or notes it is going to be excluded because the difference was zero.

The fifth column ranks the differences in terms of their size, but not sign.

The sixth and seventh column list the ranks that were for positive and negative differences, respectively.

It is these two columns that are summed to get the relevant statistics, called T+ and T-.

Because T+ and T- are not independent, we take only the T+ value.

1057
Q

For Wilcoxon signed rank test the group size n the mean T is given

A
1058
Q

For Wilcoxon signed rank test thestandard error fo T is given:

A
1059
Q

For Wilcoxon signed rank test compute z score of T by

A
1060
Q

Kruskal Wallis decision tree like one-way independent ANOVA - (4)

A

1 DV of continous or ordinal

1 IV categorical predictor of more than 2 levels

Diff participants in each predictor level

Does not meet assumption of parametric

1061
Q

Kruskal Wallis steps - (2)

A

Rank all the data on the the basis of the scores irrespective of the group

Compute the sum of ranks of each group, Ri , where i is the group number

1062
Q

For Kruskal-Wallis, the statistic H is as follows

A
1063
Q

What is decision tree of Friedman test? - (4)

A

1 DV continous or ordinal

1 IV predictor categorical with more than 2 levels

Same participants in each predictor level

Doesnot meet assumption of parametric tests

1064
Q

What is steps of Friedman test? - (2)

A

Rank the scores or each individual - that means you will have ranks varying from 1 to the number of conditions the participants took part in

Compute the sum of ranks, Ri , for each condition

1065
Q

For Friedman, the statistic F is as follows

A

K = conditions
N = number of pps

1066
Q

Example when using chi-square test - (3)

A
  • In this example, they wanted to look at whether attendance at lectures had an impact on their exam performance on whether they passed or failed
  • Attendence was coded as 1 if participants generally attended lectures , barirng illness, and 2 if they did not attend
  • Exam was scored as 1 = Pass and 2 = Fail
1067
Q
  • In this example, they wanted to look at whether attendance at lectures had an impact on their exam performance on whether they passed or failed - chi square

What does it show? - (4)

A
  • Attendence, Attended Lectures , Count = this is people who attended lecture and number of people who passed was 84 and people who failed was 29
  • % Within attendence give same info so 74.3% passed and 25.7% failed when attended lectures
    -Going to didn’t attend lectures, 22 people passed and 35 failed and below is in percentages:
  • Easier using percentages writing up
1068
Q
  • In this example, they wanted to look at whether attendance at lectures had an impact on their exam performance on whether they passed or failed - chi square

What does it show? - (2)

A
  • At top row, pearson chi-squared which chi-square statistic which was 20.617, DF which is 1 and p-value was 0.000
  • 0 cells have a count less than 5  met assumption of chi-square test that expected counts greater than 5
1069
Q
  • DF is always … in two-by-two chi-square
A

1

1070
Q

If SPSS output shows below in chi-square that

A

0 cells have a count less than 5  met assumption of chi-square test

1071
Q
  • In this example, they wanted to look at whether attendance at lectures had an impact on their exam performance on whether they passed or failed - chi square

What does it this effect size show? - (2)

A
  • x^2 (1) = 20.62, p < 0.001
  • Cramer’s V = 0.35 , indicating a medium effect size
1072
Q

Effect size guideline of r correlation coefficient - (3)

A

Small effect = 0.1
Medium effect = 0.3
Large = 0.5 and above

1073
Q

Cramer’s V can be interpreted similar to

A

correlation coefficient:

1074
Q

In chi-square we can calculate odds

A

ratio

1075
Q

Example of calculating odds ratio for chi-square - (3)

A

odds of passing/failing for students who attended lecture = no. of students who attended and passed (84) / no. of students who attedned and failed (29) = 2.897

odds of passing/failing for students who did not attend = no. of students who did not attend lectures and passed (22) / no of students who did not attended and fail (35)

Odds ratio = odds of P/F of attended/ odds of P/F of not attended = 2.897/ 0.629 = 4.606 saying for an individual who attended lectures lead them to be more likely to pass exam

1076
Q

Example research scenario of Mann Whitney - (4)

A
  • Independent sample design
  • One IV, two conditions = existing vs new medication
  • One DV (symptoms) but this time on ordinal (scale from 1 to 5) and got combination of non-normally distributed data and small sample size (very problematic for t-tests)
  • Mann Whitney U Test
1077
Q

Example of using Mann Whitney U = skew

A
1078
Q

What does this Mann Whitney U show?

  • Independent sample design
  • One IV, two conditions = existing vs new medication
  • One DV (symptoms) but this time on ordinal (scale from 1 to 5) and got combination of non-normally distributed data and small sample size (very problematic for t-tests)
  • Mann Whitney U Test
A
  • This box summarises the p-value ( p = 0.026) and tells you whether to accept or reject the null hypothesis.
1079
Q

What does this output show of Mann Whitney U? - (3)

A
  • Mann Whitney U test statistic is 166.000 to report and also people report standardised test statistic 2.292 which is z score so handy to report as know if its above +/-1.96 then p-value we get out of test is significant
  • P-value of exact significant is p = 0.026
  • This is significant difference between the 2 groups
1080
Q
  • Next we would want to look at the median scores to see which group is scoring highest and lowest after sig Mann Whitney U test

What does this output show? - (3)

  • Independent sample design
  • One IV, two conditions = existing vs new medication
  • One DV (symptoms) but this time on ordinal (scale from 1 to 5) and got combination of non-normally distributed data and small sample size (very problematic for t-tests)
  • Mann Whitney U Test
A
  • For existing treatment, median score was 3.

And new treatment the median score was 4.

It suggests new treatment was more effective in reducing symptons than the existing treatment

1081
Q

Example research scenario of Friedman ANOVA - (5)

A
  • Again we got ordinal data for DV not sure distances between levels is going to be the same
  • Related design
  • One IV, 3 conditions
  • One DV (level reached to video game)
  • Friedman’s ANOVA = more than 2 groups in related design
1082
Q

What does this Friedman ANOVA output show?

A
  • We got total sample size which is 30 and test statistic which is 21.788, DF = 2 and p value is 0.000 so significant difference between the 3 groups
1083
Q

For Friedman’s ANOVA we do

A

post hoc tests for pairwise comparisons to look where the differences are

1084
Q

What do this Friedman ANOVA test post hoc tests show? - (7)

A
  • First one is Joy stick vs Vyper Max
  • Second one is Joystick vs Evo Pro etc…
  • Notice it gives two p-values of sig and adjusted sig
  • Adjusted sig control for multiple comparison and make correcitons to p-value (use this
  • Difference between joystick vs Vyper Max was sig at p = 0.005
  • Difference between Joystick vs Evo Pro was sig at p = 0.00
  • Difference between VyperMax vs Evopro is non-significant as p = 0.660
1085
Q
  • The problem with non-parametric tests is that they have less power
A

to detect sig effects compared to parametric effects so maybe issue of dealing with power so may have median scores higher in one then another but not sig

1086
Q

Non-parametric tests are used when

A. The assumptions of parametric tests have not been met.

B. You want to increase the power of your experiment.

C. You have more than the maximum number of tied scores in your data set.

D. All of these.

A

A = non parametric have fewer assumptions than parametric

1087
Q

With 2  2 contingency tables (i.e., two categorical variables both with two categories) no expected values should be below ____.

A. 5

B. 1

C. 0.8

D. 10

A

A

1088
Q

Which of the following statements about the chi-square test is false?

A. Which of the following statements about the chi-square test is false?

B. The chi-square test can be used to check how well a model fits the data.

C. The chi-square test is used to quantify the relationship between two categorical variables.

D. The chi-square test is based on the idea of comparing the frequencies you observe in certain categories to the frequencies you might expect to get in those categories by chance.

A

A = correct, because it is false. Chi-square can be used on categorical variables only

1089
Q

When conducting a loglinear analysis, if our model is a good fit of the data then the goodness-of-fit statistic for the final model should be:

Hint: The goodness-of-fit test tests the hypothesis that the frequencies predicted by the model (the expected frequencies) are significantly different from the actual frequencies in our data (the observed frequencies).)

A. Non-significant (p should be bigger than .05)

B. Significant (p should be smaller than .05)

C.Greater than 5

D. Less than 5 but greater than 1

A

A = If our model is a good fit of the data then the observed and expected frequencies should be very similar (i.e., not significantly different

1090
Q

What is the parametric equivalent of the Wilcoxon signed-rank test?

A. The paired samples t-test

B. The independent t-test

C. Independent ANOVA

D. Pearson’s r correlation

A

A

1091
Q

Are directional hypotheses possible with chi-square?

A. Yes, but only when you have a 2 × 2 design.

B. Yes, but only when there are 12 or more degrees of freedom.

C. Directional hypotheses are never possible with the chi-squared test.

D. Yes, but only when your sample is greater than 200.

A

A =

1092
Q

A psychologist was interested in whether there was a gender difference in the use of email. She hypothesized that because women are generally better communicators than men, they would spend longer using email than their male counterparts. To test this hypothesis, the researcher sat by the computers in her research methods laboratory and when someone started using email, she noted whether they were male or female and then timed how long they spent using email (in minutes). How should she analyse the differences in males and females (use the output below to help you decide)?

A. Mann–Whitney test

B. Paired t-test

C.Wilcoxon signed-rank test

D. Independent t-test

A
1093
Q

What is the Jonckheere–Terpstra test used for?

A. To test for an ordered pattern to the medians of the groups you’re comparing.

B. To test whether the variances in your data set are approximately equal.

C. To test for an ordered pattern to the means of the groups you’re comparing.

D.To control for the familywise error rate.

A

A

1094
Q

f the standard deviation of a distribution is 5, what is its variance?

A

25 = 5^2

1095
Q

If standard deviation of distribution is 5, what is its variance?

A

5^2 = 25

1096
Q

A distribution with positive kurtosis (leptokurtic) indicates that:
A Scores are tightly clustered around the centre of the distribution
B Scores are spread widely across the distribution
C Scores are clustered towards the left side of the distribution
D Scores are clustered towards the right side of the distribution

A

A

1097
Q

If the scores on a test have a mean of 28 and a standard deviation of 3, what is the z-score for a score of 34?

A 3
B 2
C -2
D -3.42

A

B = 34-28/3

1098
Q

Question 4
Which of the following is an assumption of a one-way repeated measures ANOVA but not a one-way independent ANOVA?

A Homogeneity of variance
B Homogeneity of regression slopes
C Sphericity
D Multicollinearity

A

C

1099
Q

A test statistic with an associated p value of p = .002 tells you that:

A The statistical power of your test is large
B The probability of getting this result by chance is 0.2%, assuming the null hypothesis is correct
C The effect size of this finding is large
D All of the above

A

B

1100
Q

Question 6
Of the following, which is the most appropriate reason to use a non-parametric test?

A When the DV is measured on an ordinal scale
B When you have unequal sample sizes between conditions of the IV
C When the sample size is small
D When you have a violation of the assumption of homogeneity of variance

A

A

1101
Q

Question 7
The following are all commonly stated assumptions/requirements for using ANOVA. Which of the 4 is the only one that the procedure always requires?
A Subjects are assigned to treatment conditions / groups using random allocation
B Data is from a normally distributed population
C DV is continuous (interval or ratio)
D Variance in each experimental condition is similar (assumption of homogeneity of variance)

A

C

1102
Q

Question 8
A researcher runs a single t test and obtains a p value of p = .04. The researcher rejects the null hypothesis and concludes that there is a significant effect of the experimental manipulation in the population. Which of the following are possible?

A The researcher may have made a type 1 error
B The researcher may have made a type 2 error
C The researcher may have made a familywise error
D All of the above are possible

A

A

1103
Q

Question 9

99% of z-scores lie between:

A  1.96
B  2.58
C  3.29
D  1

A

B

1104
Q

Question 10
If predictor X shows a correlation coefficient of -.45 with outcome Y, we can confidently say that:
A X is a significant predictor of Y
B Variance in X accounts for 20.25% (that’s -.45²) of the variance in Y
C X has a causal relationship with Y
D All of the above

A

B

1105
Q

Question 11
How much variance has been explained by a correlation of r = .50?

A 10%
B 25%
C 50%
D 70%

A

B = 0.50 squared

1106
Q

Question 12
The relationship between two variables partialling out the effect that a third variable has on both of those variables can be expressed using a:

A Bivariate correlation
B Semi-partial correlation
C Point-biserial correlation
D Partial correlation

A

D

1107
Q

Question 13
A regression model in which variables are entered into the model on the basis of a mathematical criterion is known as a:

A Forced entry regression
B Hierarchical regression
C Stepwise regression
D Logistic regression

A

C

1108
Q

Question 14

In the following regression equation, what does the parameter b_0 indicate?

A The predicted value of the outcome variable
B The regression slope
C The intercept
D Error variance

A

C

1109
Q

Question 15
In multiple regression, a high VIF statistic, a low tolerance statistic, and substantial correlations between predictor variables, ALL indicate:

A Multicollinearity
B Heteroscedasticity
C The presence of outliers
D Non-normality of the residuals

A

A

1110
Q

Question 16
In a multiple regression model, the t test statistic can be used to test:

A Differences between group means
B The significance of the overall model
C The significance of the regression coefficients for each predictor
D The t test statistic is not used in multiple regression

A

C

1110
Q

Question 17
A Mixed ANOVA design would be appropriate for which of the following situations?

A Different participants are tested in each condition
B All participants are tested in all conditions
C Participants are tested in all conditions for at least one IV, and different participants are tested in each condition for at least one IV
D None of the above

A

C

1111
Q

Question 18

In a one-way independent ANOVA with 40 participants and 5 conditions of the IV, what are the degrees of freedom for the between-groups Mean Squares (MSbetween)?

A 4
B 5
C 35
D 40

A

A = k (number of grps) - 1 = 5 - 1 = 4

1112
Q

Question 19
In a two-way ANOVA there are:

A Two IVs and two DVs
B Two IVs and one DV
C One IV and two DVs
D None of the above

A

B

1113
Q

Question 20
In a two-way factorial design, the SSR (residual sum of squares) consists of:

A Variance due to the independent variables and their interaction
B Variance due to the independent variables, dependent variable(s) and error variance
C Variance accounted for by the interaction only
D Variance which cannot be explained by the independent variables

A

D

1114
Q

Question 30

Statistics enthusiast and Dub Reggae legend ‘Mad Professor’ conducted a study into the effects of listening to music on a memory task. He ended up with three independent variables and one dependent variable, and he wished to analyse all possible main effects and interaction effects. How many model effects in total will he have?

A 1
B 3
C 6
D 7

A

C = 2^k - 1 = 2^3 - 1 = 7

1115
Q

Question 24
A nutritionist was interested in the effectiveness of two of the latest fad diets. The nutritionist took 30 people who wanted to lose weight and allocated them to either the SuperScienceMaxPro weight loss regime, or the SensiNutriPlus diet. He recorded their weight at 4 time points. (The start of the diet, and then every month after that for 3 months). In addition, the nutritionist was interested in whether males and females would differ in recorded weight loss over the 4 time points. What is the design of this study?
A Two factorial with one independent factor and one repeated measures factor
B Three-factorial with two independent factors and one repeated measures factor
C Three-factorial with one independent and two repeated measures factors
D Four-factorial with two independent factors and 4 repeated measures factors

A

B

1116
Q

Question 26
What is the non-parametric equivalent of a one-way repeated measures ANOVA?
A Wilcoxon sign test
B Mann-Whitney U test
C Kruskall-Wallis test
D Friedman test

A

D

1117
Q

Question 27
What is a limitation of the Chi-square test?
A It cannot be used when you have more than 2 categorical variables
B Directional hypotheses are not possible when you have more than two conditions of a variable
C A small sample size can result in an unreliable test statistic
D All of the above

A

D

1118
Q
A
1119
Q

Distribution of z

A