Final Exam Part 2 Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

Two variable bivariate hypothesis tests exmples

A
  • tabular analysis
  • difference of means
  • correlation coefficient and regression
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Tabular Analysis

A
  • Categorical (IV and DV)
  • goal: is difference bw groups statistically significant?
  • How: calculate the chi2 stat and p-value
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Degrees of Freedom

A

max amount of independent values, that have freedom to vary in data sample
- Formula: degrees of freedom=size of data sample-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

critical value

A

point on the test distribution that is compared to the test stat to determine whether to reject the null hypothesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

level of significance

A
  • in a p-value approach
  • level of significance= alpha
    alpha=1-confidence interval
  • probability of incorrectly rejecting the null hypothesis
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Difference of Means

A
  • continuous DV and IV
  • use sample means and SD to make inferences about unobserved pop.
  • are means different across values of independent variable?
  • How: calculate the t-statistic and p-value
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

correlation coefficient

A
  • are they correlated/associated?
  • how do we know they are correlated?
  • measure of the strength of the relationship
  • more of X associated w more/less of Y
  • can use scatter plot or pearsons r regression
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

perfect negative correlation

A

-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

perfect positive correlation

A

1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

no linear relationship bw two variables

A

0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

The pearsons r

A
  • no unambiguous classification rule for the quantity of a linear relationship bw two variables
  • tell us how confident we can be that relationship is different from what we would see if our IV and DV were not related in population
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

examples of non-linear relationships

A
  • quadratic/culvilinear

- cubic/polynomial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

common mistakes w pearsons r

A
  • it is not applicable for non-linear relationships

- it does not equal slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

R-square

A
  • always bw 0-1

- proportion of variation in the DV(y) that is explained in the IV (x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

coefficient in the regression

A

represents mean increase of Y for every additional unit of X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Simple regression model

A

no controls (only X and Y)

17
Q

multiple regression model

A

controls for confounding variables (Z)

- more applicable in the real world, bc realworld is multivariate

18
Q

most causal theories are…

A

bivariate, does X cause Y

19
Q

two-variable regression

A

fitting the best line through a scatter plot of data

20
Q

statistical model

A

slope and y-intercept

21
Q

systematic components of population regression model

A
  • aplha= fixed value, value of Y depends on alpha, so no random value in systematic components
  • when X changes by 1 unit, then value of Y changes by beta units
22
Q

random components of population regression model

A
  • cannot predict values of this part (no systematic patterns)
  • condition or mood
23
Q

names for the estimated stochastic component

A
  • residual: leftover part of Yi, after line was drawn

- sample error mean/pop. error term

24
Q

how to calculate the smallest sum of residual values (ordinary-least- squares)

A
  • add the squared value of each the residuals for each line

- choose line that has the smallest total value–> draw line that minimizes the sum of the squared residuals

25
Q

goodness of fit measures (root-mean squared error)

A
  • measures of the overall fit bw a regression model and the dependent variable
  • quantifies how well the OLS regression we have obtained fits the data
  • r-squared statistic
26
Q

p value and estimated beta

A

how likely it is that we observe this sample slope of estimated beta from the real world if the true (but unobserved) population slope beta is equal to 0

27
Q

two-tailed (non-directional) hypothesis tests

A

most common hypothesis tests about parameters from the OLS regression model

28
Q

in any observational study, how do we control for the effects of other variables?

A

multiple regression is the most common method in social science

29
Q

omitted variables bias

A

bias from the failure to include a variable that belongs in regression model
(result of omitting a variable, Z, that should have been in the model)

30
Q

small bias

A

if either or both components of bias term are close to zero (red area overlapped is small)

31
Q

large bias

A

if both components are likely large (red area overlapped is quite large)

32
Q

positive bias

A

if beta1 and beta2 correlate bw X and Y are positive

33
Q

standardized coefficient

A

remove metric of each variable to make them comparable to one another
- coefficients on a standardized metric

34
Q

unstandardized coefficients

A

coefficients in the table each exist in the native metric of each variable. Normally not comparable