Hypotheses, Prediction, Scientific Method Flashcards

1
Q

what are hypothesis testing alternatives?

A

Bayesian methods and likelihood techniques

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what are the main characteristics of “new” hypothesis testing alternatives?

A

prior information
several hypotheses
inferences from multiple models
parameters which refer to biological processes/states

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

how can the scientist use his knowledge in these “new” hypothesis testing alternatives?

A

the knowledge of the scientist is used to formulate hypotheses (testable ideas) and select models that represent them

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

how detailed should a model be?

A

this can be objectively assessed depending on the data available to estimate parameters for model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is the probability of observing this data if the null hypothesis was true? (which statistical method is used to answer this question?)

A

we answer that with p-value, which shows the probability of observing current or more extreme values considering that the null hypothesis is true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is missed by a p-value based analysis?

A

does not evaluate if hypothesis/model is strong; as it only evaluates chance of extreme events

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is the likelihood of competing hypotheses considering the data?

A

proportional to likelihood of data given the hypotheses

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

give 3 characteristics of likelihood methods for testing hypotheses

A

competing models can have different parameters data is fixed and hypothesis is variable
relative support of each model given the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is goal and main assumption of maximum likelihood estimations?

A
  • goal: find parameter estimates which give a distribution which will in turn maximize chance of observing data
  • assumption: considering that each observation is independent, then the total probability of observing all of data is the product of observing each data point individually
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how to evaluate likelihood models?

A

by AIC (Akaike’s Information Criterion): the best model loses less information when attempting to approximate reality (lowest AIC value); assessment of a given model depends on performance of other models which are being compared together

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what are Akaike weights for?

A

to account for uncertainty (if we selected these models many times, what is the probability of model x being the best one?)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

compare a traditional hypothesis testing with a likelihood method in terms of how groups are compared

A

instead of comparing means, we compare competing hypotheses in terms of how much support there is in the data for effects of each treatment; can be multivariate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the main difference between bayesian and likelihood methods?

A

pre-existing info are probabilities instead of likelihoods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

describe a Bayesian analysis

A

previous knowledge for each parameter is a requirement; prior = pre-existing datasets; new knowledge from new set of data = previous knowledge modified by what was learned from the new data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what is a unique feature of bayesian methods?

A

output of Bayesian analyses: probabilities of hypotheses (no other method does it)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

describe output for frequentist, likelihood, and bayesian methods

A

frequentist stats say that if experiment is repeated many times, parameter estimate would fall in the 95% confidence interval 95% of the times
bayesian: probabilities of hypotheses (no other method does it)
likelihoods are proportionate to probabilities

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

what is a confidence interval?

A

range of values we are fairly sure our true value lies in

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

confidence, in statistics, is another way to describe probability. exemplify this.

A

if you construct a confidence interval with a 95% confidence level, you are confident that 95 out of 100 times the estimate will fall between the upper and lower values specified by the confidence interval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

what happens with the confidence interval if I am comparing a highly variable group with a uniform group?

A

the confidence interval will be narrower for the uniform group (more certainty)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

which method answers these questions: what is the odd of alternative results? What are the expected values of alternative decisions? Or the frequency distribution of results?

A

these are questions which can only be answered by Bayesian methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

compare likelihood and bayesian methods in terms of parameters

A

likelihood: maximization of possible values across parameters; Bayesian: integration across parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

problem with Bayesian methods: when we don’t know much about parameters; what is the solution?

A

meta-analysis: summarizing pre-existing knowledge

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

what are meta-analysis types in terms of parameter estimation?

A

stimate the value of a parameter based on several studies; several parameters whose values are different in each population (can be visualized in a histogram)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

what is a problem with meta-analyses in terms of how samples are taken?

A

we can’t guarantee that the samples were randomly taken (eg., more productive samples usually are chosen)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

why are ecological data and conclusions different than other fields?

A

natural variability and complex interactions interferes with results, is of interest to us, and makes data analysis a complex task
instead of having a single correct hypothesis, we usually end up with several partially correct hypotheses. Thus, we obtain a solid understanding of a phenomenon by looking at several studies over time instead of looking at a single one (eg., how trophic interactions shape communities)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

why are new statistical methods more adequate for ecological studies?

A

ecological processes are characterized by several forces acting together; each portion of this complex system will have hypotheses that are more or less true. Thus, new statistics (likelihoods and Bayesian probabilities) are more aligned with ecology because they allow us to evaluate several hypotheses in terms of their relative importance to explain a given process

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

what can be the contribution of new statistics to the precautionary principle and for adaptive management of natural resources?

A
  • Precautionary principle: practice is forbidden unless you can prove inexistence of effects; this is questionable: how large should an effect be to justify regulation? New statistics can help understand the probability of competing hypotheses and the size of the effects.
  • Adaptive management of natural resources: improved data (future, after harvest for example) is estimated with Bayesian methods
28
Q

what was the original idea behind p-values?

A

could my results be produced by random chance?

29
Q

what is a null hypothesis?

A

what we want to disprove

30
Q

what is a p-value?

A

what are the chances of getting results at least as extreme as the ones I actually got? This is the p-value (the smallest the value, the greater the chance of the null hypothesis being false)

31
Q

p-value of 0.01 means that there is a…

A

99% chance of value falling in the 99% confidence interval; it summarizes the data given one specific null hypothesis

32
Q

does a p-value of 0.01 mean that there is only 1% of accepted hypothesis being wrong?

A

no, as it does not calculate the odds of an effect being true in the first place

33
Q

describe a problem with p-value

A

deviates attention from effect size analysis

34
Q

what is p-hacking?

A

dropping conditions or doing other manipulations so that significant p-value is achieved

35
Q

what are the recommendations for analyzing data moving forward from traditional p-value based analyses?

A
  • Confidence intervals (relative importance) and effect sizes (magnitude of effect)
  • move to Bayesian methods (probabilities of outcome instead of potential frequency of outcome)
  • try several methods and compare results
  • disclosure written on paper (how was sample size determined? Was data excluded? How were manipulations and measures made?)
  • use an open science framework database to publish results of a first round of experiments, then replicate studies and publish
36
Q

complete the sentence for traditional hypotheses testing: what is the chance of observing current effects (or more extreme)…

A

if the true effect is zero?

37
Q

when we want to falsify a hypothesis, what analysis is appropriate?

A

p-values

38
Q

what is a difficulty of likelihood techniques?

A

how to select candidate set of models to represent hypotheses?

39
Q

what is the problem when analyzing with small populations in a frequentist approach?

A

a small decrease in pop growth, for example, will be likely be non-significant (small statistical power); this would be better analyzed as an exercise of parameter estimation only

40
Q

when are frequentist approaches useful?

A

frequentist approaches are useful for carefully designed experiments which aim to falsify a given hypothesis

41
Q

what are three distinct goals for statistical modeling in ecology?

A

data exploration, inference, and prediction

42
Q

Populations, communities, and ecosystem-level
properties fluctuate through time in response to internal
and external drivers. Describe them.

A

Internal drivers include intraspecific density dependence, demographic stochasticity, interspecific interactions, and food web dynamics. External drivers are typically related to environmental conditions,
and weather is perhaps the most variable.

43
Q

Why is quantifying the relative impacts of internal forcing versus weather on ecological dynamics a core goal of ecology with
new relevance?

A

As we attempt to predict ecological responses to climate change.

44
Q

what is the goal of exploratory studies?

A

describe patterns in the data and generate hypotheses about nature

45
Q

what is a type 1 error? how can it happen during exploratory analysis?

A

making false discoveries. it can happen when trying to indiscriminately combinate many potential covariates (a few of them will be highly correlated by chance)

46
Q

what are the 3 good practices for exploratory analyses?

A

only consider plausible relationships
correct p-values for multiple testing to reduce chance of false discoveries
communicate exploratory nature of analysis (proposed hypotheses will only be valid if confirmed by subsequent independent studies)

47
Q

what is inference analysis?

A

evaluate the strength of evidence in a data set for some statement about nature

48
Q

what is the main difference between exploratory and inferential analysis?

A

exploratory: hypotheses are generated
inference: tests hypotheses based on a priori knowledge of ecological systems; eg. null hypothesis significance testing
we need independent datasets for both of these techniques

49
Q

are conclusions obtained via statistical inference automatically accepted as scientific fact?

A

no, as the reliability of statements is tied to a particular set of data obtained by particular methods under particular conditions; thus, it requires replication and validation across a range of conditions before they are accepted as scientific facts

50
Q

what happens when we are comparing too many models in inferential analyses?

A

the true goal of the analyses may be exploratory or predictive. inferential analyses have a small risk of false discoveries only when a small set of groups (2 or 3) are compared.

51
Q

what is distinguishing feature of modeling for prediction?

A

the need to test predictive models “out of sample”: using independent data that were not used to fit the model. Exploratory modeling can help identify important predictors even if the causal mechanisms are currently not understood, but without validation on independent data, the predictive skill of the model is unknown.

52
Q

what is the difference between predictive and inferential analyses when looking at a linear model?

A

predictive analyses focus on the response variable (which model will best predict y values for new observations of covariates in x?
inferential analyses focus on estimates of the regression coefficients (which ones are positive, negative, non-zero? which covariates are more important?)

53
Q

what are the 2 main components of the AIC (Akaike’s Information Criterion) formula?

A

number of parameters
maximum likelihood estimates for model parameters

54
Q

what are the pitfalls of AIC analyses?

A

Proponents of AIC argue that this approach avoids arbitrary P-value cut-offs (Burnham and Anderson 2002), but in practice researchers have relied on equally arbitrary and less interpretable cut-offs for the
difference in AIC values required to conclude that one model is more supported by the data than another.

55
Q

what is regularization (eg., 2p in AIC equations)? what would happen if it was not there?

A

2p = regulates model complexity by penalizing models with too many parameters
without this penalty, models with the highest number of parameters would always be the chosen as the best ones (more complex models)

56
Q

define regularization.

A

regularization involves selecting a model or estimating parameter values based on a weighted average of a goodness-of-fit measure and a complexity penalty.

57
Q

what is model validation? why is it important?

A

comparing model predictions with observations that were not used to
“train,” or fit, the model. Out-of-sample validation is important because it is easy for a model to reproduce patterns in the training data, but much harder to accurately predict the outcome in another situation, which is what we expect predictive models to do.

58
Q

what is the process of out-of-sample validation like?

A
  • randomly splitting the available data into training and validation sets
  • we then fit candidate models using the training data, use
    each model to make predictions for the validation data,
    and quantify model errors using a summary measure
    such as root mean squared error.
59
Q

what can we do if our dataset is too small to do an out-of-sample validation technique?

A

when data sets are too small for out-of-sample validation,
we can use cross-validation

60
Q

how can we screen for potentially important covariates when doing an exploratory analysis?

A

we can calculate the correlation coefficient between population growth rate and each covariate and establish an arbitrary cut-off value (eg., 0.3)

61
Q

describe the steps for exploratory analysis

A
  • define the research question (eg., which covariates are associated with population growth rates)
  • screen many covariates for potential associations (eg., correlation coefficients with arbitrary threshold)
  • assess statistical evidence more rigorously (linear mixed-models including covariates with high coefficient; followed by likelihood methods comparing full model with reduced versions to exclude covariates which do not change full model results)
  • fit final model and check for inconsistencies (eg., estimates are positive in Nov but negative in Dec?)
  • correct for multiple comparisons (control for expected proportion of of false significant results)
  • consider alternative approaches (sliding climate windows to average weather data from a group of periods instead of analyzing periods individually)
62
Q

describe the steps for inference.

A

formulate competing hypotheses: priori hypotheses based on the
literature or deduced from theory
translate hypotheses into alternative models: eg. null-hypothesis significance testing

63
Q

describe the steps for prediction.

A
  • define the predictive goal: eg., objective is to predict butterfly population size (on log scale) at time t given data on log population size in year t − 1 and information about weather during the period in between the observations
  • choose model selection approach (eg., regularization or AIC?)
  • choose model validation approach (eg., select sub dataset as training data and validation data from dataset)
64
Q

does an understanding of the process translate into improved predictive skill?

A
  • not necessarily, as using covariates from an inference study can lead to loss of predictive accuracy, for example
65
Q

give an example to explain why a given predictive model testing effects of climate on population growth may be weak

A

a given population may be responding to many weak climate signals instead of some stronger ones

66
Q

describe classical errors for each exploratory, inference, and predictive studies.

A

exploratory: overfitting, coming up with fake discoveries
inference: misrepresenting exploratory analyses as inference studies
predictive: not testing predictive power of models in new, independent data

67
Q

while in traditional statistics we are aiming at establishing significant differences among means, what kind of comparisons can we make using new methods (maximum likelihood and bayesian)?

A

because p-values in traditional analysis do not estimate evidence, new techniques are more appropriate for finding in the data the relative support for various competing hypotheses regarding the effects of treatment.