ARMS Week 1 Flashcards

1
Q

Frequentist approach vs Bayesian framework

A

Frequentist:
- Test how well data fit H0
- p-values, confidence intervals, effect sizes, power analysis

Bayesian:
- Probability of hypothesis given the observed data -> prior information
- BFs, prior, posterior, credible intervals

Both can be used for estimation and hypothesis testing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Frequentist estimation

A
  • Emprical research uses collected data to learn from
  • likelihood function
  • probability of an event = frequency it occurs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Bayesian estimation

A
  • prior information about u
  • prior is updated and provides posterior distribution of u
  • conditional probabilities -> P(A given B)
    Pro: accumulating knowlegde
    Con: results depend on choice prior
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Posterior mean or mode

A

The mean or mode of the posterior distribution. Should add up to 1, just as the prior model probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Posterior SD

A

Standard deviation of posterior distribution (comparable to frequentist standard error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Posterior 95% credible interval

A

Providing bounds of the part of posterior with 95% of posterior mass

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Posterior Model Probability (PMP)

A

The probability of the hypothesis after observing the data
Depends on two criteria:
1. How sensible it is, based on the prior
2. How well it fits the new data

  • PMP are relative probabilities. They are updates of prior probabilities with the BF.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Bayesian testing is comparative. What does this mean?

A

Hypotheses are tested against one another, not in isolation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does the Bayes factor say?

A

How much support there is for H1 compared to H0.
BF10 = 10 -> support for H1 is 10 times stronger than for H0.
BF10 = 1 -> support for H1 is as strong as for H0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Probability theory: frequentist vs bayesian

A

Frequentist:
- probability is the relative frequency of events

Bayesian:
- probability is the degree of belief

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

95% Confidence interval vs 95% credible interval

A

Confidence interval 95% (frequentist)
- if we were to repeat this experiment many times and calculate a CI each time, 95% of the intervals will include the true parameter value.

Credible interval 95% (Bayesian)
- there is a 95% probability that the true value is in the credible interval
- a zero present = not significant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Linear regression

A

Lineair association between a dependent and independent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Residual

A

Difference between value and the line in the plot (we can’t explain them).
- we want to minimize residuals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Multiple Lineair regression (MLR)

A

Lineair regression with multiple predictors (independent variables)

Assumptions:
- Dependent variable is continuous (interval/ratio)
- Independent variables are continuous or dichotomous (nominal with two categories)
- Linearity of relations (the L in MLR) -> checked with scatterplots
- No outliers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Dummy variables

A

Has value 0 and 1. Is used in MLR to code data suitable for this approach (interval/ratio).
Number of dummy variables is equal to the groups minus 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

R^2 vs adjusted R^2

A

R^2 -> proportion of explained variance in the sample
Adjusted R^2 -> proportion of explained variance in the population

17
Q

R

A

Correlation coefficient, explains how much the variables correlate with one another

18
Q

B value

A

Unstandardized effect

19
Q

B0

20
Q

Hierarchical MLR

A

Comparing two nested models (multiple formulas) -> one model is a smaller version of the other

21
Q

R^2 changed

A

How much more variance does the second model explain compared to the first?

22
Q

Prior

A

Existing knowlegde before looking at own data

23
Q

Prior model probabilities

A

How likely is each hypothesis before seeing the data.
- add up to 1 -> relative probabilities divided over the hypotheses of interest
For example: H1 = 0,2 and H2 = 0,8 ; or H1 = H2 = H3 = 0,33

24
Q

Standardized residuals

A

Check wether there are outliers in the Y-space.
Rule of thumb -> values must be between -3,3 and +3,3

25
Q

Cook’s Distance

A

With this you can check whether there are outliers within the XY-space.
Rule of thumb -> must be lower than 1

26
Q

Multicollinearity

A

Indicates whether the relationship between two or more independent variables is too strong.

Including overly related variables in your model has three consequences:
1. B is unreliable
2. Limits magnitude of R (correlation between Y and Y^)
3. Importance of individual independent variables can hardly be determined (if at all)

Possible solutions:
- remove variables
- combine variables in scales

27
Q

Variance Inflation Factor (VIF)

A

Used to determine whether multicollinearity is an issue.
Rule of thumb -> VIF > 5 = potential problem, VIF > 10 = problem.

28
Q

Homoscedasticy

A

Spread of residuals must be approximately the same across all values for y. When not: heteroscedasticy

29
Q

B (Beta)

A

Standardized coefficients, can be used to determine the most important predictor of a model. The independent variables with the largest beta is the most important predictor. -> - or + doesn’t matter.

30
Q

Construct validity

A

Extent to which a conceptual variable is accurately measured or manipulated

31
Q

Internal validity

A

Extent to which the research method can eliminate alternative explanations for an effect/relationship (trying to find causal relationship)

32
Q

External validity

A

Extent to which the research results generalize to besides those in the original study

33
Q

Statistical validity

A

Extent to which te results of a statistical analysis are accurately measured and well founded (checking assumptions and report significance)