Research Methods II Flashcards

1
Q

Frequency distribution

A

Organized tabulation of individuals in each category on the scale of measurement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

f (frequency)

A

Frequency of a particular score.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Cumulative frequency

A

Accumulation of individual scores as one moves up the scale. Lowest score, sum of all frequencies, and all scores below it. Highest score should have cumulative frequency equal to total sample size.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Cumulative percentile rank

A

Accumulation of percent of all scores as one moves up the scale. Start with lowest score divide the cumulative frequency of a particular score by the total sample size.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Mean

A

Sum of scores divided by number of scores. Average.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Median

A

Score divides distribution in half. 50th percentile.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Mode

A

Score in distribution with greatest frequency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Degrees of Freedom (v or df)

A

Number of values used to estimate a parameter minus the number of parameters to be estimated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why use df in sample SD?

A

All scores in a set of scores are free to vary, EXCEPT for the last score. Last score is restricted if the mean (or the sum) and the number of scores are known. The correct way to get the UNBIASED estimators of the population is to divide the square deviations by N - 1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Transformation of the Scale for the SD

A
  • Adding or subtracting a constant to each score in a distribution will not change the standard deviation but will change the mean by the same constant.
  • Multiplying or dividing each score by a constant causes the standard deviation and the mean to be multiplied or divided by the same constant.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Symmetrical distribution

A

Distribution which the left side of the distribution “mirrors” the right side of the distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Skewed distribution

A

Distribution is skewed if one tail is longer than the other. Long tail in positive direction = positive skew. One tail in negative direction = negative.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Order of Mode, Median, & Mean for Positive skew

A

Left to right:
- Mode
- Median
- Mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Order of Mode, Median & Mean for negative skew

A

Left to right:
- Mean
- Median
- Mode

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Kurtosis

A

“Peakedness” or “flatness” of distribution. How fat or thin a distribution is. Degree of frequency distribution is flat (low kurtosis) or peaked (high kurtosis).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Mesokurtic

A

Distribution with zero kurtosis. Normal distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Leptokurtic

A

Distribution with positive kurtosis. Acute peak around the mean, fat tails.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Platykurtic

A

Distribution with negative kurtosis. Small peak around mean, thin tails.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Bimodal distribution

A

2 modes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Rectangular distribution

A

Mean, median & no mode. No mode because all scores have the same frequency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Sampling error

A

The amount of error between statistics calculated from sample to corresponding population parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

A sampling distribution

A

Distribution selected by all possible samples of a specific size from the population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Distribution of sample means

A

Collection of sample means for all possible random samples of a particular size (n) that can be obtained from the population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Central Limit Theorem

A

For any population with mean µ and standard deviation σ, the distribution of sample means for a sample size of n will approach normal distribution with a mean of µ and a standard deviation of σ/ square root n as n approaches infinity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

The Standard Error of Xbar

A

The standard deviation of the distribution of sample means.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

Law of Large Numbers

A

The larger the sample size, the more probable it is that the sample mean will be close to the population mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Confidence Intervals

A

Used to help estimate or feel confident that the actual µ is within a certain range of the mean of means from many samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Statistical Model

A

Statistical representation of the real world.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

A Simple Statistical Model

A

The mean is a hypothetical value (i.e., doesn’t have to be a value that actually exists in the data set). As such, the mean is a simple statistical model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

Measuring the “Fit” of the Model

A

The mean if a model of what happens in the real world: typical score. It is not a perfect representation of data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

Null Hypothesis

A

The predicted relationship does NOT exist. The symbol is Ho.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

Alternative Hypothesis

A

The predicted relationship does exist. The symbol is H1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Type I error

A

The rejection of the null hypothesis is true. Saying there is a relationship when it does not exist.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Type II error

A

The acceptance of the null hypothesis when the null hypothesis is false. Saying there is no relationship when there is a relationship.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

Alpha Level

A

Minimizing risk of a Type I error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

Power

A

The probability of correctly rejecting the null hypothesis when the null is false. It is the probability that Type II error is not committed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Factors Affecting Power

A
  • The alpha level. Increase the alpha level x increase power of statistical test.
  • One tailed vs. 2-tailed alpha test.
  • Sample size. As sample size increases, so does power.
  • Reduce error variance will increase power. (Test everyone in same quiet room rather than different room, different noises.)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

Increasing effect size of the independent variable…

A

Would increase power.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Greater subject variability will…

A

Decrease power.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

Parametric tests based on normal distribution requires 4 basic assumptions

A
  • Normally distributed sampling distribution.
  • Homogeneity of variance/ homoscedasticity.
  • Interval or ratio data.
  • Independence of scores.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

Normal dsitribution

A

A probability of distribution of a random variable with perfect symmetry and a skew of 0 and kurtosis of 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

Non-parametric tests

A

A family of statistical tests that do not rely on the restrictive assumptions of parametric test. Does not assume sampling distribution is normally distributed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

Homogeneity of variance (HOV)

A

Assumption that the variance 1 continuous variable is stable/consistent between treatment groups of a discrete variable for t-tests & ANOVAs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

Homoscedasticity

A

Assumption the variance of 1 continuous variable is stable/consistent across scores of another continuous variable. For regressions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

Independence of scores

A

One data point does not influence another data point.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

Big advantage of Parametric Tests

A

More powerful (statistically speaking in reflecting the null hypothesis when it is false) compared to non-parametric tests.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

Big advantage of Nonparametric tests

A

More freedom! Not restricted by assumptions to do data analysis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

Kolmogorov-Smirnov Test

A
  • Tests if data differ from a normal distribution.
  • Significant = non-normal data.
  • Non-significant = normal data.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

Histograms

A

Frequency distribution with bar drawn adjacent. Gives continuous figure that emphasizes the continuity of variables. Good for continuous variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

Q-Q Plots

A

Quantile-quantile plot. Plots quantiles of a particular distribution. If value falls on the diagonal plot, the value shares the same distribution as the normal distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

Quantile

A

The proportion of case we find below a certain value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

A perfect normal distribution would…

A

Have a skewness of 0 and a kurtosis of 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

What does a significant Levene’s test mean?

A
  • Tests if variances in different groups are the same.
  • Significant = variances not equal (bad)
  • Non-significant = variances are equal (good)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

Log transformation (log (xi)) to…

A

Reduce positive skew.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

Square root transformation (square root Xi) to reduce…

A

Positive skew and to stabilize variance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

Reciprocal transformation ( l / xi) can also reduce skewness.

A

Dividing 1 by each scores also reduces the impact of large scores. This transformation reverse the scores; you can avoid this by reversing the scores before the transformation, 1 (xhighest - xi)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

Potential problems with transforming data

A

Transforming the data helps as often as it hinders the accuracy of F.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

Use Robust methods (e.g.. Bootstrap) to…

A

account for violations of assumptions (e.g., normality).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

Correlations

A

Way of measuring the extent to which two continuous variables are related. Measures pattern of responses across variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

Scatterplot

A

A perfect linear relationship (r = +1.00 or - 1.00) is when all the data points lie on a straight line in a scatter plot.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

Covariance

A

Average cross-product deviations.
- Calculate the error between the mean and each subjects’ score for the first variable (x).
- Calculate the error between the mean and their score for the second variable (y).
- Multiply these error values.
- Add these values and you get the cross product deviations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

Problems with Covariance

A
  • Depends upon the units of measurement. e.g. the covariance of the two variables measured in miles might be 4.25, but if the same scores are converted to kilometers, the covariance is 11.
  • Standardize it. Divide by the standard deviations of both variables.
  • The standardized version of covariance is known as the correlation coefficient. H is relatively unaffected by units of measurement.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

The Correlation Coefficient

A

Measures the degree and direction of linear relationship between two variables in terms of standardized (z-scores with a mean of 0 and a SD of 1).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

Sum of Products

A

To calculate the Pearson Correlation, you need to first calculate the Sum of Products.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

Correlation simply describes… a relationship between 2 variables, not causality.

A

a relationship between 2 variables, not causality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

Third-variable Problem

A

Causality between two variables cannot be assumed because there may be other measured or unmeasured variables affecting the results.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

Direction of causality

A

Correlation coefficients say nothing about which variable causes the other to change.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

Measurement error affecting r

A

The more error there is in a measure, the smaller the correlation can be. This measures the relationship or correlation between two parallel measures of a variable of interest.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

Coefficient of Determination

A
  • r squared = coefficient of determination
  • It measures the proportion of variability in one variable that can be determined from the relationship with the other variable.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

Linear Transformations

A
  • Variables transformed into standard scores or added or multiplied by a constant, the correlation between the two variables will remain the same.
71
Q

Spearman’s rho

A
  • Pearson’s correlation on the ranked data
  • Good for non-normally distributed data
72
Q

Kendall’s tau

A
  • Also measures relationship between ordinal variables.
  • It gives more accurate p-values for small data sets, but less popular than Spearman’s rho.
73
Q

Regression Line

A

The straight line that best describes the linear relationship between two variables.
- Describes the relationship between two variables.
- Identifies the center or “central tendency” of the relation.
- Line used for prediction.

74
Q

Equation for Linear Relationship

A

Yi = bo + biXi + Ei

75
Q

Yi

A

criterion variable, dependent variable

76
Q

Xi

A

predictor variable, independent variable

77
Q

bi

A

Regression coefficient for the predictor (IV)
- Gradient (slope) of the regression line
- Direction/strength of the relationship

78
Q

bo

A

Y intercept (value of Y when X = 0). Also called the constant.
- Point at which the regression line crosses the Y-axis (ordinate)

79
Q

Ei

A

the error of the regression line

80
Q

Assumption of Regression:
Linearity

A

Since it is based on linear correlations, multiple regression assumes linear bivariate relationships between each x and y, and also between y and predicted y.

81
Q

Assumption of Regression:
Normality

A

Both univariate & multivariate distributions of residuals (actual scores minus predicted scores) are normally distributed. Thus, the Y scores are independent and normally distributed (Shapiro-Wilk).

82
Q

Assumptions of Regression:
Independence of scores

A

Independence of Y (outcome: DV) scores.

83
Q

Assumptions of Regression:
Independence of errors

A

The errors (residuals) from observations using the regression formula should not be correlated with each other (Durbin-Watson test).

84
Q

Assumption of Regression:
Minimal Multicollinearity

A

The predictors (IVs) should not be highly correlated with each other. Rule of thumb is no higher than r -.80 between predictors.

85
Q

Assumption of Regression:
Homoscedasticity

A

Variance of the residuals are uniform for all values of U.

86
Q

Assumption of Regression:
Heteroscedasticity

A

Spread of data points is smaller for low X values and wider for higher X values.

87
Q

Standard Error of Estimate

A
  • The measure of accuracy of regression.
  • The regression equation allows predictors, but does not provide info about accuracy.
  • Standard distance between the regression line and the actual data points.
  • The greater the correlation, the smaller the standard error of the estimate.
88
Q

The least squares criterion

A
  • Want the sums of the squared errors to be the smallest.
  • Yields values for the b-weights and the y-intercept that results in sum of the squared residuals being at the minimum.
  • The best fitting-line has the smallest total squared error.
89
Q

Regression to the Mean

A
  • Occurs when you have a nonrandom sample from a population and two measures that are imperfectly correlated.
  • The sample posttest mean is closer to the posttest population mean than their pretest mean to the pretest population mean.
90
Q

Characteristics of the Regression to the Mean

A
  • Can occur because the sample is not random.
  • Group phenomenon.
  • Happens between any two variables.
  • The more extreme the sample group; the greater the regression to the mean.
    r = 1, no regression to the mean
    r =.5, 50% regression to the mean
    r = .2, 80% regression to the mean
    r = 0, 100% regression to the mean
91
Q

Multiple regression

A

2 or more independent variables in the regression model
- include more than one predictor to enhance the prediction of Y

92
Q

Zero-Order Correlations

A

Relationship between two variables, ignoring influence of other variables in prediction.

93
Q

Highest order correlations:
First order

A

Relation between 2 variables after controlling for influence of 1 other variable.

94
Q

Highest order correlations:
Second order

A

Relation between 2 variables after controlling for influence of 2 other variables.

95
Q

Partial Correlation

A

Relationship between two variables after removing the overlap completely from both variables.

96
Q

Part (semi-partial) correlations

A

Relationship between two variables after removing a third variable from just the independent variable.

97
Q

Methods of Variable Entry into the Multiple Regression Model:
Simultaneous Entry

A

All variables are entered at the same time and the Beta weights are determined simultaneously.

98
Q

Methods of Variable Entry into the Multiple Regression:
Sequential (hierarchical) Entry

A

Used to build a subset of predictors.

99
Q

Methods of Variable Entry into the Multiple Regression Model:
Apriori

A

Variables entered are determined by some theory.

100
Q

Methods of Variable Entry into the Multiple Regression Model:
Statistical Criteria

A

Computer decides which variables are entered based on their unique predictive abilities.

101
Q

Problem of Shrinkage:
Low N:k Ratio

A

Enough participants for each predictors. If participants is low to number of predictors, sample estimated may not predict population.

102
Q

Problem of Shrinkage:
Multicollinearity

A

High correlations betwene predictors can cause instability of prediction.

103
Q

Problem of Shrinkage:
Measurement error

A

Measurement does not reflect true score, the application of Beta weights to ta new sample may not be accurate.

104
Q

Reading/Interpreting Regression Tables:
R squared

A

Variance explained by the regression model, aka coefficient of determination

105
Q

Reading/Interpreting Regression Tables:
B

A

B weight, raw score; the unique effect of the predictor

106
Q

Reading/Interpreting Regression Tables:
SE B

A

Standard error of B weight of the predictor

107
Q

Reading/Interpreting regression Tables:
beta

A

beta weight; the standardized unique effect of the predictor based on z-scores

108
Q

Bootstrapping

A

A method of resampling in which the model is re-run multiple (e.g., thousands of) times with different permutations of the same data.
- It is used to provide a more reliable estimate (s) of the statistic.
- It is especially useful for small samples and non-normal data.

109
Q

Moderator

A

When or how much the IV will affect the DV. (Levels, low, average, high, etc.)

110
Q

Mediator

A

The mechanism: how or why the effect of the IV occurs. (Third variable, affects the independent variable)

111
Q

Why center the continuous IVs to create the interaction terms?

A

To get the interaction term for 2 continuous independent variables in multiple regression, the score of the predictors must be centered around the mean.

112
Q

Disadvantage of Barron & Kenny

A

Low statistical power.

113
Q

Complete (full) mediation

A

Relationship between IV & DV completely disappears. Beta weight approaches 0.

114
Q

Partial mediation

A

Relationship with IV & DV, but it is reduced. Beta weight may be significant, but difference between new & old beta weights is significant.

115
Q

Mediation Package Method

A

More flexible and statistically powerful.

116
Q

Average Casual Mediation Effects (ACME)

A

Indirect effect of M (total effect - direct effect). Must be significant to show that mediation is significant.

117
Q

Average Direct Effects (ADE)

A

Must NOT be significant to prove complete mediation.

118
Q

Total effect

A

Combined indirect & direct effect.

119
Q

Bootstrapping for…

A

more accuracy & power.

120
Q

Rationale for the t-test or when do you use a t-test?

A

Only one independent variable is manipulated in only two ways and only one outcome is measured.

121
Q

Assumptions of t-test

A
  • Independent t-test & dependent t-test are PARAMETRIC TESTS based on the NORMAL DISTRIBUTION.
  • Sampling distribution is NORMALLY DISTRIBUTED.
  • Dependent variable data are CONTINUOUS (interval or ratio).
  • Independent t-test assumes:
    x Homogeneity of variances.
    x Scores in different conditions are independent.
122
Q

Independent t-test

A
  • Two means based on independent data.
  • Data from different groups of people.
123
Q

The t-test as a GLM (Regression equation)

A

Nominal IVs are usually done with t-tests or ANOVAs. Can be done in regression as long as predictors are coded correctly (Dummy coding).

124
Q

Dependent t-test

A

Compares two means based on related data.

125
Q

Cohen’s d for effect

A

0.2 = small
0.5 = medium
0.8 = large

126
Q

What to do when assumptions are broken:

A
  • Nonparametric version of independent t-test
    x Mann-Whitney test; aka Wilcoxon rank-sum test
  • Nonparametric version of dependent t-test
    x Wilcoxon signed-rank test
  • Robust test
    x Bootstrapping
127
Q

Mann-Whitney test; aka Wilcoxon rank-sum test

A
  • Non-parametric equivalent of the independent sample t-test.
  • Use to test differences between two conditions in which different participants have been used.
128
Q

Wilcoxon signed-rank test

A

To compare two sets of scores when these scores come from the same participants.

129
Q

Robust tests

A

These functions require the data to be in two different columns.

130
Q

Bootstrapping

A

Robust method to compare independent means.

131
Q

Trimmed means

A

“Where 20% of extreme scores are excluded from calculating the mean”
- Could be excluding a certain percentage from the mean.

132
Q

M estimator: median

A
  • Median rather than trimmed mean.
133
Q

Analysis of Variance (ANOVA)

A
  • Determines if mean differences exist for 2 or more treatments or populations. Involves comparison of variances that reflect different sources of variability.
  • Test the differences between samples are due to chance (sampling error) or whether they are systematic treatment effects that have caused the scores in one group to be different from the scores in another.
134
Q

F-ratio = t-squared under what conditions?

A

T-squared = F when there is 1 independent variable with only 2 treatment conditions.

135
Q

Components of the ANOVA

A
  • Between-Treatment Group Variability: Between treatment conditions, differences among sample means.
    x Treatment effect, individual differences & experimental error could explain differences.
  • Within-Treatment Group Variability: Variability within each sample.
    x Individual differences and experimental error.
136
Q

What is the F ratio when there is no treatment effect?

A

Fratio = Treatment effect + experimental error/individual differences + experimental error ~ 1

137
Q

What is the F ratio when there is a treatment effect?

A

F ratio = Treatment effect + individual differences + experimental error/ individual differences + experimental error > 1

138
Q

Assumptions of ANOVA

A
  • Independence
  • Normality
  • Homogeneity of variance
139
Q

Test homogeneity of variance assumption

A

Levene’s Test

140
Q

The ANOVA as a GLM (regression equation)

A
  • Requires data to be in wide format rather than long format.
  • Robust ANOVA based on trimmed means.
  • Compare medians rather than means.
  • Add bootstrap to the trimmed mean method.
141
Q

Planned Contrasts (Group Comparisons)

A
  • The variability explained by the model (experimental manipulation, SSm) is due to participants being assigned to different groups.
  • This variability can be broken down further to test specific hypotheses about which groups might differ.
  • We break down the variance according to hypotheses made a priori (before the experiment).
142
Q

Robust ANOVA

A
  • Compare medians rather than means
  • Add bootstrap to trimmed mean method
143
Q

Post Hoc tests

A

Determine specifically which treatment groups or samples are different from each other.

144
Q

Testwise alpha level

A

Alpha level you select for each individual hypotheses.

145
Q

Experimentwise alpha level

A

Total probability of a Type I error that is accumulated from all the separate tests in the experiment.

146
Q

Bonferroni

A
  • Calculates new pairwise alpha to keep the familywise alpha at .05.
  • Strictest correction procedure, lowest power.
  • Bonferroni a = a/number of tests
  • Most commonly used.
  • May overcome Type I.
147
Q

Scheffe Test

A
  • Allow researcher to conduct any & all comparisons while preventing experiment wise error from exceeding alpha level.
  • Strict, also low power.
148
Q

Tukey HSD Test

A
  • Help control for experimentwise Type I error when the set of comparisons consists of pairs of treatment means or pairwise comparison.
149
Q

Kruskal-Wallis test

A
  • Non-parametric counterpart of one way independent ANOVA.
  • Data violated assumptions. Based on ranked data.
150
Q

Friedman’s within-subjects ANOVA

A
  • Used for testing differences between condition when there are 2 or more conditions.
  • Same participants for all conditions. Repeated measures design.
  • Based on ranked data.
151
Q

Analysis of Covariance

A

Any ANOVA design can become an ANCOVA by the addition of a concomitant variable called a covariate (CV). ANCOVA is an extension of ANOVA where the effects of the IVs on the DV are assessed after the effects of one or more covariates are partialed out partitioning variance.

152
Q

Advantages of ANCOVA

A
  • Reduces error variance – by explaining some of the unexplained variance (SSr) the error variance in the model can be reduced.
  • Increases statistical power – by equating the treatment groups, we reduce error variance due to subject variability, and thus, increase power. Matching is a “procedural” way to equate the groups, whereas ANCOVA is a statistical way to equate them.
  • Greater experimental control – by controlling known extraneous variables, we gain greater insight into the effect of the predictor variables (s)
153
Q

Choosing a good covariate

A
  • A covariate is a source of variation that is not controlled for in the design of the experiment, but which does affect the dependent variable.
  • The covariate is correlated with the dependent variable. We want a correlation of at least r = .20.
  • The covariate should be independent of the independent variable(s), and it shouldn’t correlate highly with any other covariates.
  • Adding a covariate complicate the design. It also means you’ll probably need more subjects so you won’t get empty cells in the new design.
154
Q

Adjusted means

A
  • Use a regression equation to find the adjusted means for each treatment group.
  • Will be used when you want to do main group comparisons (contrasts).
155
Q

Homogeneity of regression slopes assumption

A
  • The regression slope has to be the same for each of the 3 groups.
  • Want parallel lines, not interactions.
156
Q

Assumption covariate (CV) is independent of the independent variable (IV)

A

No relationship between covariate and IV. Run regular ANOVA.

157
Q

Testing HOV assumption

A

Levene Test - same as ANOVA, want the value to be not significant

158
Q

Contrast (coding) to do Type III ANCOVA

A

R uses the Type I model to fit predictors into the model which uses the predictors based on how they were entered. Want to use Type III to get the unique effect of each predictor.
- Use contrast coding to write data in format for Type III model.
- Second line describes the relationship between covariate & DV.
- Third line indicates that the IV does has a significant effect on activity level, AFTER controlling for partner activity level.

159
Q

Robust ANCOVA being bootstrap ANCOVA

A

Do robust ANCOVA to free the analysis from the restriction of homogeneity of regression slopes.

160
Q

Benefit of Factorial Designs

A
  • We can look at how variables interact. Moderation model.
  • Interactions: Show how the effects of one IV on the DV might depend on the effects (level) of another IV.
161
Q

Main effects

A
  • The separate effect of each independent variable AVERAGED over the levels of the other independent variable (AVERAGED EFFECTS)
162
Q

Simple effects

A

The variability among the treatment means associated with one independent variable at a particular level of the other independent variable. AKA simple main effects.

163
Q

Interaction

A

An interaction is present when the effect of one of the independent variables on the dependent variable is not the same at all levels of the second independent variable.
The simple effects of one of the independent variables are not the same at all levels of the second independent variable.
When an interaction is present the main effect is NOT representative of the corresponding simple effects.
The main effects do not fully or accurately describe the data.

164
Q

Interpreting graphs:
Main effect A

A

If the average line representing each condition of A is flat or no slope then there is no main effect A. If there is a slope of that averaged line, then there is a significant main effect A.

165
Q

Interpreting graphs:
Main effect B

A

If the dots that represent the average of all the conditions of B are right on top of each other, then there is no main effect of B. If the dots are not directly on top of each other and they are spaced apart, then yes, there is a significant main effect of B.

166
Q

Interpreting graphs:
Interaction

A

If the lines are parallel, then there is no interaction. If the lines are NOT parallel, then there is a significant interaction.

167
Q

Interaction comparison

A

Within the 3-way ANOVA (A x B x C) are smaller interactions (e.g. A x B of level C1 and at level C2) are called simple interactions. The comparison of smaller interactions is called interaction comparisons.

168
Q

Interaction contrast

A
  • Any interaction analysis that analyses each factor only on 2 of its levels.
  • So for a 4 x 4 design, an interaction contrast would be 2 x 2 x 2 (where 1 level of A, 2 levels of B, and 1 level of C are ignored).
  • In a 3-factor design, 3 x 4 x 3, an interaction contrast would be 2 x 2 x 2 (where 1 level of A, 2 levels of B, and 1 level of C are ignored).
  • These types of interaction analyses are useful because like main & simple comparisons, they break the interaction terms into a single degree of freedom.
  • It would enable us to understand the interaction or the simple effects in more detail.
169
Q

Contrast coding to do Type III ANOVA

A
  • Can utilize planned (apriori) group comparisons in the model
  • Do contrast coding of the IVs first so that we can make group comparisons of no versus any alcohol & 2 pints versus 4 pints of alcohol for example.
170
Q

Goodness of fit test

A

Used to see if two or more categories of a nominal variable differ significantly from expectation.
- df + (c-1); number of categories minus 1

171
Q

Pearson’s Chi-Square Test/Chi-Square test of Association

A
  • Use to see whether there’s a relationship between two categorical variables.
  • t obs needs to be greater than t crit.
172
Q

Likelihood Ratio Statistics

A
  • An alternative to Pearson’s chi-square, based on maximum likelihood theory.
  • Create a model for which the probability of obtaining the observed set of data is maximized.
  • This model is compared to the probability of obtaining the observed set of data is maximized.
  • The resulting statistic compares observed frequencies with those predicted by the model.
  • i and j are the rows and columns of the contingency table and ln is the natural logarithm.
  • Preferred to Pearson’s chi-square when the total sample (N) is small.
173
Q

Assumptions of Chi-square

A
  • Independence
    x Each person, item, or entity contributes to only one cell of the contingency table.
  • The expected frequencies should be greater than n = 5.
    x In larger contingency tables, up to 20% of expected frequencies can be below 5 for a category or group, but there is a loss of statistical power.
    x Even in larger contingency tables, no expected frequencies should be below 1. If you find yourself in this situation, consider using Fisher’s exact test.
174
Q

Chi-square test

A

df = (rows - 1) x (columns - 1)