Biostats Flashcards

1
Q

Analysis of Variance

A

A method of hypothesis testing, normally used with more than two study groups, in which the likelihood of the outcome is estimated by a ratio of the variability between groups divided by the variability within groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Variance between Groups

A

The component of the total variability in ones data that measures the variability of group means around the grand mean. The between groups variance measures treatment effects if present. If there is no treatment effect the between variance component reflects only error variance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Variance within Groups

A

The component of the total variability in ones data that measures error (i.e., natural) variability. This is measured by the variability of group measurements around the group mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Statistical Error

A

A conclusion error in hypothesis testing. This happens when one incorrectly rejects Ho or incorrectly does not reject Ho.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Sum of Squarea

A

A component of variance that is formed by summing the squared differences of numbers around their mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Mean Square

A

Another term for variance. A mean square is formed by dividing a sum of squares term by its degrees of freedom.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Factors

A

The systematic sources of variability in an ANOVA.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Contrast Coefficients

A

Positive and negative integers that are used in a sum of squares equation to test a particular pattern of means in a one-way ANOVA.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Levels of a Factor

A

The divisions of a factor. Each factor must have at least two levels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

MSbg, MSwg

A

Another term for variance. A mean square is formed by dividing a sum of squares term by its degrees of freedom.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Type I Error

A

A conclusion error in hypothesis testing when one incorrectly rejects Ho. This is analogous to a false positive.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Planned Comparisons

A

A method of hypothesis testing within ANOVA in which a specific pattern of group means is tested. A planned comparison is also called an a priori comparison because a global F does not need to be computed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Post Hoc Comparisons

A

The hypothesis testing that occurs among group means following a statistically significant global ANOVA. Post-hoc comparisons are also called a posteriori tests because they are conducted only if a global F test reaches statistical significance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

F Ratio

A

Used in the analysis of variance, the F ratio is the quotient of variance between groups divided by the variance within groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Main Effect

A

In a factorial design a main effect is the influence of one factor on the dependent variable while disregarding the other factors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Interaction

A

In a factorial design an interaction is the combined influence of more than one factor on the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Test for Trend

A

A specific type of a priori comparison in which one tests for a particular type of relationship (e.g., linear, quadratic) with an independent and dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Orthogonal Components

A

Independent contributions to the total sum of squares in an ANOVA. In general, the term “orthogonal” means independent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Dunnett’s T-Test

A

This is one of the post-hoc tests in which all groups are compared with a reference group. The results of the Dunnett Test should be reported only if the global ANOVA reaches statistical significance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Regression Line

A

A type of analysis to reveal the nature and degree of association of two measurements.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Slope

A

The rate of change of a line. The slope describes how fast the line is rising or sinking. It is expressed as the change in y per unit change in x.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Y-Intercept

A

The point along the ordinate (the y axis) where a line cuts through when x = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Pearson Correlation Coefficient

A

A parametric measure of association, bounded by + or - 1 in which the observed covariance of two variables is divided by the maximum possible covariance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Homoscedasity

A

One of the main assumptions of regression analysis which states that the variability of the y variable is constant across values of x.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Heteroscedasity

A

In regression analysis when the variability of the y variable is not constant across values of x.

26
Q

Least Squares Criterion

A

The rule for identifying the best fitting line in a scatterplot. The criterion states that the best fitting line is the one which minimizes the squared deviations of points around the line.

27
Q

Prediction Error

A

look in notes

28
Q

Direct Relationship

A

A relationship in which two variables vary in the same manner. A direct relationship is observed when one variable increases and decreases as another variable does the same.

29
Q

Spearman Correlation Coefficient

A

A non-parametric measure of association, bounded by + or - 1 which is based upon the same concept of association in the Pearson Correlation, but is computed by way of ranks.

30
Q

Influence

A

look in notes

31
Q

Coefficient of Determination

A

The square of the Pearson r which describes the proportion of variance in one variable that is explained by the variance in the other.

32
Q

Coefficient of Alienation

A

1 - the square of the Pearson r which describes the proportion of variance in one variable that is unrelated to the variance in the other.

33
Q

Scatterplot

A

A graph in which the values of one variable are plotted as a function of another.

34
Q

Standard Error of Estimate

A

The average error (in y units) when predicting y from x.

35
Q

Residuals

A

The difference between a measurement and the value of the measurement that is predicted by some mathematical model.

36
Q

Covary

A

The general term to denote that Variables vary together (i.e., height and weight covary).

37
Q

Inverse Relationship

A

A relationship in which two variables vary in opposition. As one increases, the other decreases.

38
Q

Fisher Transformation

A

A non-linear re-expression that is used in assessing the confidence interval of a population Pearson Correlation coefficient.

39
Q

Linearity Assumption

A

One of the main assumptions of regression analysis which states that a straight line adequately represents the relationship between x and y.

40
Q

Simpson’s Paradox

A

The change in the magnitude or direction of a relationship that is due to a confounding variable.

41
Q

Observed Frequency

A

Frequencies that are determined from observation (i.e., ones data).

42
Q

Expected Frequency

A

Frequencies that are calculated from an expected pattern.

43
Q

Cell

A

The intersection of two levels in a contingency table.

44
Q

Column Percent

A

The proportion or percentage that is formed within a vertical column from the frequencies in a chi-square test for association.

45
Q

Row Percent

A

The proportion or percentage that is formed within a Horizontal row from the frequencies in a chi-square test for association.

46
Q

Odds Ratio

A

A quotient of two odds that is conditioned by a dichotomous variable.

47
Q

X2 Goodness of Fit Test

A

A form of hypothesis testing with frequencies in which the observed data are compared or “fitted” against expected values.

48
Q

X2 Test for Association

A

A chi-square goodness of fit test that is used to study the relationship between two categorical variables.

49
Q

McNemar Test for Symmetry

A

A specific type of chi-square test that analyzes the change in a dichotomous variable.

50
Q

Contingency Table

A

A table constructed with at least two factors that reveal the intersection of all levels. A contingency table is used in factorial ANOVA and with the chi-square test for association.

51
Q

Phi Coefficient

A

A measure of association bounded between + and - 1 that is calculated from a 2x2 contingency table. The Phi coefficient is a specific instance of the Cramer V.

52
Q

Cramer V

A

A measure of association bounded between + and - 1 that is calculated from a contingency table that is larger than 2x2.

53
Q

Relative Benefit (or risk)

A

The quotient of the probability of improving with treatment divided by the probability of improving without treatment.

54
Q

Effect Size

A

A measure of the degree of impact of a treatment.

55
Q

Type II Error

A

A conclusion error in hypothesis testing when one incorrectly does not reject Ho. This is analogous to a false negative.

56
Q

Alpha

A

The point in a sampling distribution in which one rejects Ho. Alpha is the probability of a Type I error.

57
Q

Statistical Decision Error

A

A conclusion error in hypothesis testing. This happens when one incorrectly rejects Ho or incorrectly does not reject Ho.

58
Q

Correct Decision

A

In hypothesis testing when one makes a conclusion that is consistent with reality. In other words, one makes a correct decision to reject Ho when it is false and to not reject Ho when it is true.

59
Q

Power

A

The probability or correctly rejecting Ho. Typically, we want power to be .80 (i.e., 80%).

60
Q

Beta, 1-Beta

A

The probability of a Type II error.

61
Q

Factorial Design

A

An experimental design in which the influence of more than one factor is studied.