Statistics Flashcards
sample
A sample is selected to represent the population in a research study; helps answer questions about a population.
variable
characteristic or condition that can change or take on different values.
discrete variables
(such as class size) consist of indivisible categories (weird when it is represented by a fraction
continuous variables
(such as time or weight) are infinitely divisible into whatever units a researcher may choose. Also, that can be legitimately measured.
goal of experiment
Goal of an experiment is to demonstrate a cause-and-effect relationship between two variables; that is, to show that changing the value of one variable causes changes to occur in a second variable.
IV and DV
In an experiment, the manipulated variable is called the independent variable and the observed variable is the dependent variable.
non-experimental or quasiexperimental
non-experimental or quasi-experimental, are similar to experiments because they also compare groups of scores. These studies do not use a manipulated variable to differentiate the groups. Instead, the variable that differentiates the groups is usually a pre-existing participant variable (such as male/female) or a time variable (such as before/after).
Similar to correlational research because they simply demonstrate and describe relationships
positively skewed
In a positively skewed distribution, the scores tend to pile up on the left side of the distribution with the tail tapering off to the right.
negatively skewed
In a negatively skewed distribution, the scores tend to pile up on the right side and the tail points to the left.
percentile rank
The percentile rank for a particular X value is the percentage of individuals with scores equal to or less than that X value. When an X value is described by its rank, it is called a percentile.
nominal
name only, categorical; only permit you to determine whether two individuals are the same or different. (i.e. male/female; diagnosis)
ordinal
rank ordered (e.g. height shortest to tallest); tell you the direction of difference between two individuals.
spearman correlation (i.e. class rank)
interval
consistent intervals between numbers but no absolute zero (i.e. IQ); identify the direction and magnitude of a difference
ratio
interval plus absolute zero – height in inches; identify the direction and magnitude of differences and allow ratio comparisons of measurements
reliability
same results with repeated administrations
validity
measures what it says it measures- taps into the construct
standard error of the mean (SEM)
measure of variability; the average expected difference between sample means (i.e. M1 – M2 expected)
confidence interval
a type of interval estimate of a population parameter and is used to indicate the reliability of an estimate. Certain factors may affect the confidence interval size including size of sample, level of confidence, and population variability. A larger sample size normally will lead to a better estimate of the population parameter.
sampling error
The discrepancy between a sample statistic and its population parameter is called sampling error.
central tendency in normal distribution
a statistical measure that determines a single value that accurately describes the center of the distribution and represents the entire distribution of scores. The goal of central tendency is to identify the single value that is the best representative for the entire set of data. Allows researchers to summarize or condense a large set of data into a single value (thus a descriptive statistic).
mean and a mean in skewed data
Mean: the average; most commonly used; requires scores that are numerical values measured on an interval or ratio scale
When a distribution contains a few extreme scores (or is very skewed), the mean will be pulled toward the extremes (displaced toward the tail).
median
Median: If the scores in a distribution are listed in order from smallest to largest, the median is defined as the midpoint of the list; values measured on an ordinal, interval, or ratio scale
Relatively unaffected by extreme scores
mode
the most frequently occurring category or score in the distribution; any scale of measurement: nominal, ordinal, interval, or ratio
what will be equal in a symmetrical distribution
the mean and median will always be equal
variability
goal is to obtain a measure of how spread out the scores are in a distribution; describes how the scores are scattered around that central point. In the context of inferential statistics, variability provides a measure of how accurately any individual score or sample represents the entire population. Measuring distance.
range and interquartile range
largest to smallest
The interquartile range is the distance covered by the middle 50% of the distribution (the difference between Q1 and Q3).
standard deviation
average distance between most scores in the distribution (square root of the variance) – b/w the score and the mean
variance
average squared deviation between most scores in the distribution (it is the standard deviation squared)
negative skew vs positive skew outliers
Positive Skew – extreme outlier(s) in the positive/ high end
Negative Skew – extreme outlier(s) in the negative/ low end
implication of skew for parameters
violates the parametric assumptions needed for the parametric tests (t-tests, anova, pearson correlation)
descriptive statistics
methods for organizing and summarizing data
-mean, median, mode, standard deviation, variance
parameter vs statistic
A descriptive value for a population is called a parameter and a descriptive value for a sample is called a statistic.
frequency distribution
organized tabulation showing exactly how many individuals are located in each category on the scale of measurement. A frequency distribution presents an organized picture of the entire set of scores, and it shows where each individual is located relative to others in the distribution.
regular frequency distribution
when a frequency distribution table lists all of the individual categories (X values) it is called a regular frequency distribution.
grouped frequency distribution
In a grouped frequency distribution, the X column lists groups of scores, called class intervals, rather than individual values (too many different X values).
why are frequency distribution graphs useful
Frequency distribution graphs are useful because they show the entire set of scores. At a glance, you can determine the highest score, the lowest score, where the scores are centered, most common score.
inferential statistics
methods for using sample data to make general conclusions (inferences) about populations
effect size
measure of the strength of a phenomenon
hypothesis testing
general goal of a hypothesis test is to rule out chance (sampling error) as a plausible explanation for the results from a research study. A technique to help determine whether a specific treatment has an effect on the individuals in a population.
null hypothesis
H0, always states that the treatment has no effect (no change, no difference). According to the null hypothesis, the population mean after treatment is the same is it was before treatment.
Test statistic in critical region → reject the null = p<0.05 = there is an effect, difference
Test statistic in the body (outside of the critical region) → fail to reject the null = there is no effect/ no difference
alternative hypothesis
a statement that directly contradicts a null hypothesis by stating that the actual value of a population parameter is less than, greater than, or not equal to the value states in the null hypothesis
alpha level: establishes a criterion, or “cut-off”, for making a decision about the null hypothesis. The alpha level also determines the risk of a Type I error.
Power - probability that the test will reject the null hypothesis when the null hypothesis is false (find an effect when there is one)
-Influenced by: Alpha level; Sample size; Sensitivity of test; Effect size
type 1 error
occurs when the sample data appear to show a treatment effect when, in fact, there is none. In this case the researcher will reject the null hypothesis and falsely conclude that the treatment has an effect.
what causes type 1 error
caused by unusual, unrepresentative samples. Just by chance the researcher selects an extreme sample with the result that the sample falls in the critical region even though the treatment has no effect.
The hypothesis test is structured so that Type I errors are very unlikely; specifically, the probability of a Type I error is equal to the alpha level.
type 2 error
occurs when the sample does not appear to have been affected by the treatment when, in fact, the treatment does have an effect. In this case, the researcher will fail to reject the null hypothesis and falsely conclude that the treatment does not have an effect.
what causes type 2 errors
commonly the result of a very small treatment effect. Although the treatment does have an effect, it is not large enough to show up in the research study.
directional test or one tailed test
A directional test or a one-tailed test includes the directional prediction in the statement of the hypotheses and in the location of the critical region.
what is recommended that the hypothesis test be accompanied by
effect size
We use Cohen’s d as a standardized measure of effect size. Much like a z-score, Cohen’s d measures the size of the mean difference in terms of the standard deviation (Impact of the IV).
what are the three parametric assumptions we made to engage in inferential statistics
1) Independent Observations – random selection, representative sample you can’t just bring friends to the experiment – b/c of bias (not representative)
2) Normally distributed: Populations which samples are selected from are normally distributed (if not normally distributed then can’t run these tests)
3) Homogeneity of variance: populations from which samples are selected have equal variances (if change the experiment then the scores should shift equally)
independent measures between subjects t test
(2 separate groups)
An independent-measures design can be used to test for mean differences between two distinct populations (such as men versus women) or between two different treatment conditions (such as drug versus no-drug).
repeated measures design t test
single group of individuals is obtained and each individual is measured in both of the treatment conditions being compared. Thus, the data consist of two scores for each individual.
related sample t -test, matched subjects design
each individual in one treatment is matched one-to-one with a corresponding individual in the second treatment
ANOVA?
comparing 3 or more treatment conditions; more than 1 IV (factor); more than 2 levels of an IV
what does analysis of variance do in ANOVA
controls the risk of a Type I error in situations where a study is comparing more than two population means
why use post hoc tests
ANOVA simply establishes that differences exist, it does not indicate exactly which treatments are different. Specifically, you must follow the ANOVA with additional tests, called post hoc tests, to determine exactly which treatments are different and which are not.
examples of post hoc tests
The Scheffe test and Tukey=s HSD are examples of post tests. Indicates exactly where the difference is
what does the repeated measure design do
the repeated measures design eliminates individual differences from the between treatments variability because the same subjects are used in every treatment condition
MANOVA
ANOVA with the addition of multiple Dependent variables
ANCOVA
ANOVA with a covariate- a variable that varies systematically with the IV
-covariate: another variable that has a relation to the DV
when are non parametric tests used
used when violate the parametric assumptions (or data is NOT interval or ratio level)
chi square tests
tests the shape of the distribution of nominal data/ categorical/ not a parametric test
goodness of fit in chi square
roughly the same number of subjects in each category? OR does the distribution fit a predetermined distribution (i.e. 40% male and 60% female)
test for independence
similar to correlation in that it looks at the relationship between 2 variables but uses NOMINAL data
Mann Whitney U
analogous to independent measures t-test
Friedman tests
analogous to repeated measures anova
Wilcoxin test
analogous to repeated measures t-test
Kruskal Wallace
analogous to independent measures (one-way) anova
correlation
tests the relationship between 2 variables that occur naturally: relationship only; no cause and effect; determine whether there is a relationship between two variables and to describe the relationship
sign and strength
positive or negative tells nothing about the strength of the relationship but tells about the direction (both go up together or one goes up the other goes down)
-Negative correlation: one variable increases, other decreases
-Positive correlation: one variable increases, the other increases
Strength – between 0 and 1 (closer to 1 = stronger)
restriction of range
can make the relationship seem weaker because you are only getting a small snapshot of the full relationship between the two variables (i.e. looking at the relationship between age and reaction time if you only use 19-22 year olds you won’t find a relationship)
pearson product moment correlation
used for parametric data/ interval or ratio level data with a linear relationship