Statistics Flashcards

1
Q

sample

A

A sample is selected to represent the population in a research study; helps answer questions about a population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

variable

A

characteristic or condition that can change or take on different values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

discrete variables

A

(such as class size) consist of indivisible categories (weird when it is represented by a fraction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

continuous variables

A

(such as time or weight) are infinitely divisible into whatever units a researcher may choose. Also, that can be legitimately measured.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

goal of experiment

A

Goal of an experiment is to demonstrate a cause-and-effect relationship between two variables; that is, to show that changing the value of one variable causes changes to occur in a second variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

IV and DV

A

In an experiment, the manipulated variable is called the independent variable and the observed variable is the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

non-experimental or quasiexperimental

A

non-experimental or quasi-experimental, are similar to experiments because they also compare groups of scores. These studies do not use a manipulated variable to differentiate the groups. Instead, the variable that differentiates the groups is usually a pre-existing participant variable (such as male/female) or a time variable (such as before/after).

Similar to correlational research because they simply demonstrate and describe relationships

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

positively skewed

A

In a positively skewed distribution, the scores tend to pile up on the left side of the distribution with the tail tapering off to the right.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

negatively skewed

A

In a negatively skewed distribution, the scores tend to pile up on the right side and the tail points to the left.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

percentile rank

A

The percentile rank for a particular X value is the percentage of individuals with scores equal to or less than that X value. When an X value is described by its rank, it is called a percentile.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

nominal

A

name only, categorical; only permit you to determine whether two individuals are the same or different. (i.e. male/female; diagnosis)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

ordinal

A

rank ordered (e.g. height shortest to tallest); tell you the direction of difference between two individuals.

spearman correlation (i.e. class rank)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

interval

A

consistent intervals between numbers but no absolute zero (i.e. IQ); identify the direction and magnitude of a difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

ratio

A

interval plus absolute zero – height in inches; identify the direction and magnitude of differences and allow ratio comparisons of measurements

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

reliability

A

same results with repeated administrations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

validity

A

measures what it says it measures- taps into the construct

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

standard error of the mean (SEM)

A

measure of variability; the average expected difference between sample means (i.e. M1 – M2 expected)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

confidence interval

A

a type of interval estimate of a population parameter and is used to indicate the reliability of an estimate. Certain factors may affect the confidence interval size including size of sample, level of confidence, and population variability. A larger sample size normally will lead to a better estimate of the population parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

sampling error

A

The discrepancy between a sample statistic and its population parameter is called sampling error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

central tendency in normal distribution

A

a statistical measure that determines a single value that accurately describes the center of the distribution and represents the entire distribution of scores. The goal of central tendency is to identify the single value that is the best representative for the entire set of data. Allows researchers to summarize or condense a large set of data into a single value (thus a descriptive statistic).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

mean and a mean in skewed data

A

Mean: the average; most commonly used; requires scores that are numerical values measured on an interval or ratio scale

When a distribution contains a few extreme scores (or is very skewed), the mean will be pulled toward the extremes (displaced toward the tail).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

median

A

Median: If the scores in a distribution are listed in order from smallest to largest, the median is defined as the midpoint of the list; values measured on an ordinal, interval, or ratio scale

Relatively unaffected by extreme scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

mode

A

the most frequently occurring category or score in the distribution; any scale of measurement: nominal, ordinal, interval, or ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

what will be equal in a symmetrical distribution

A

the mean and median will always be equal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

variability

A

goal is to obtain a measure of how spread out the scores are in a distribution; describes how the scores are scattered around that central point. In the context of inferential statistics, variability provides a measure of how accurately any individual score or sample represents the entire population. Measuring distance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

range and interquartile range

A

largest to smallest

The interquartile range is the distance covered by the middle 50% of the distribution (the difference between Q1 and Q3).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

standard deviation

A

average distance between most scores in the distribution (square root of the variance) – b/w the score and the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

variance

A

average squared deviation between most scores in the distribution (it is the standard deviation squared)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

negative skew vs positive skew outliers

A

Positive Skew – extreme outlier(s) in the positive/ high end

Negative Skew – extreme outlier(s) in the negative/ low end

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

implication of skew for parameters

A

violates the parametric assumptions needed for the parametric tests (t-tests, anova, pearson correlation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

descriptive statistics

A

methods for organizing and summarizing data
-mean, median, mode, standard deviation, variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

parameter vs statistic

A

A descriptive value for a population is called a parameter and a descriptive value for a sample is called a statistic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

frequency distribution

A

organized tabulation showing exactly how many individuals are located in each category on the scale of measurement. A frequency distribution presents an organized picture of the entire set of scores, and it shows where each individual is located relative to others in the distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

regular frequency distribution

A

when a frequency distribution table lists all of the individual categories (X values) it is called a regular frequency distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

grouped frequency distribution

A

In a grouped frequency distribution, the X column lists groups of scores, called class intervals, rather than individual values (too many different X values).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

why are frequency distribution graphs useful

A

Frequency distribution graphs are useful because they show the entire set of scores. At a glance, you can determine the highest score, the lowest score, where the scores are centered, most common score.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

inferential statistics

A

methods for using sample data to make general conclusions (inferences) about populations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

effect size

A

measure of the strength of a phenomenon

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

hypothesis testing

A

general goal of a hypothesis test is to rule out chance (sampling error) as a plausible explanation for the results from a research study. A technique to help determine whether a specific treatment has an effect on the individuals in a population.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

null hypothesis

A

H0, always states that the treatment has no effect (no change, no difference). According to the null hypothesis, the population mean after treatment is the same is it was before treatment.

Test statistic in critical region → reject the null = p<0.05 = there is an effect, difference

Test statistic in the body (outside of the critical region) → fail to reject the null = there is no effect/ no difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

alternative hypothesis

A

a statement that directly contradicts a null hypothesis by stating that the actual value of a population parameter is less than, greater than, or not equal to the value states in the null hypothesis

alpha level: establishes a criterion, or “cut-off”, for making a decision about the null hypothesis. The alpha level also determines the risk of a Type I error.

Power - probability that the test will reject the null hypothesis when the null hypothesis is false (find an effect when there is one)
-Influenced by: Alpha level; Sample size; Sensitivity of test; Effect size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

type 1 error

A

occurs when the sample data appear to show a treatment effect when, in fact, there is none. In this case the researcher will reject the null hypothesis and falsely conclude that the treatment has an effect.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

what causes type 1 error

A

caused by unusual, unrepresentative samples. Just by chance the researcher selects an extreme sample with the result that the sample falls in the critical region even though the treatment has no effect.

The hypothesis test is structured so that Type I errors are very unlikely; specifically, the probability of a Type I error is equal to the alpha level.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

type 2 error

A

occurs when the sample does not appear to have been affected by the treatment when, in fact, the treatment does have an effect. In this case, the researcher will fail to reject the null hypothesis and falsely conclude that the treatment does not have an effect.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

what causes type 2 errors

A

commonly the result of a very small treatment effect. Although the treatment does have an effect, it is not large enough to show up in the research study.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

directional test or one tailed test

A

A directional test or a one-tailed test includes the directional prediction in the statement of the hypotheses and in the location of the critical region.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

what is recommended that the hypothesis test be accompanied by

A

effect size

We use Cohen’s d as a standardized measure of effect size. Much like a z-score, Cohen’s d measures the size of the mean difference in terms of the standard deviation (Impact of the IV).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

what are the three parametric assumptions we made to engage in inferential statistics

A

1) Independent Observations – random selection, representative sample you can’t just bring friends to the experiment – b/c of bias (not representative)

2) Normally distributed: Populations which samples are selected from are normally distributed (if not normally distributed then can’t run these tests)

3) Homogeneity of variance: populations from which samples are selected have equal variances (if change the experiment then the scores should shift equally)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

independent measures between subjects t test

A

(2 separate groups)

An independent-measures design can be used to test for mean differences between two distinct populations (such as men versus women) or between two different treatment conditions (such as drug versus no-drug).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

repeated measures design t test

A

single group of individuals is obtained and each individual is measured in both of the treatment conditions being compared. Thus, the data consist of two scores for each individual.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

related sample t -test, matched subjects design

A

each individual in one treatment is matched one-to-one with a corresponding individual in the second treatment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

ANOVA?

A

comparing 3 or more treatment conditions; more than 1 IV (factor); more than 2 levels of an IV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

what does analysis of variance do in ANOVA

A

controls the risk of a Type I error in situations where a study is comparing more than two population means

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

why use post hoc tests

A

ANOVA simply establishes that differences exist, it does not indicate exactly which treatments are different. Specifically, you must follow the ANOVA with additional tests, called post hoc tests, to determine exactly which treatments are different and which are not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

examples of post hoc tests

A

The Scheffe test and Tukey=s HSD are examples of post tests. Indicates exactly where the difference is

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

what does the repeated measure design do

A

the repeated measures design eliminates individual differences from the between treatments variability because the same subjects are used in every treatment condition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

MANOVA

A

ANOVA with the addition of multiple Dependent variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

ANCOVA

A

ANOVA with a covariate- a variable that varies systematically with the IV
-covariate: another variable that has a relation to the DV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

when are non parametric tests used

A

used when violate the parametric assumptions (or data is NOT interval or ratio level)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

chi square tests

A

tests the shape of the distribution of nominal data/ categorical/ not a parametric test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

goodness of fit in chi square

A

roughly the same number of subjects in each category? OR does the distribution fit a predetermined distribution (i.e. 40% male and 60% female)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

test for independence

A

similar to correlation in that it looks at the relationship between 2 variables but uses NOMINAL data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

Mann Whitney U

A

analogous to independent measures t-test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

Friedman tests

A

analogous to repeated measures anova

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

Wilcoxin test

A

analogous to repeated measures t-test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

Kruskal Wallace

A

analogous to independent measures (one-way) anova

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

correlation

A

tests the relationship between 2 variables that occur naturally: relationship only; no cause and effect; determine whether there is a relationship between two variables and to describe the relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

sign and strength

A

positive or negative tells nothing about the strength of the relationship but tells about the direction (both go up together or one goes up the other goes down)
-Negative correlation: one variable increases, other decreases
-Positive correlation: one variable increases, the other increases

Strength – between 0 and 1 (closer to 1 = stronger)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

restriction of range

A

can make the relationship seem weaker because you are only getting a small snapshot of the full relationship between the two variables (i.e. looking at the relationship between age and reaction time if you only use 19-22 year olds you won’t find a relationship)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

pearson product moment correlation

A

used for parametric data/ interval or ratio level data with a linear relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

spearman correlation

A

used for non-linear (i.e. inverted U relationship) or when data is ordinal/ ranked

72
Q

regression

A

used for prediction; a statistical technique that related a dependent variable to one or more independent (explanatory) variables. A regression model is able to show whether changes observed in the dependent variable are associated with changes in one or more of the explanatory variables

73
Q

multiple regression

A

same as regression but using multiple variables to make the prediction (i.e. SAT scores, HS GPA, and quality of essay to predict college GPA)

74
Q

path analysis

A

allows for the evaluation of the causal flow of relationships because a priori predictions are made about the correlations

75
Q

logistic regression

A

prediction of group membership (doctor, lawyer, tradesperson) from interval or ratio level predictors (scores usually)

76
Q

discriminant analysis

A

same as Logistic regression but predictors can be any kind of variable (dichotomous, interval, ratio, nominal)

77
Q

statistics

A

branch of mathematics used to summarize, analyze, and interpret a group of numbers or observations

78
Q

research method or scientific method

A

set of systemic techniques used to acquire, modify, and integrate knowledge concerning observable and measurable phenomena

79
Q

experiments

A

use of methods and procedures to make observations in which a researcher fully controls the conditions and experiences of participants by applying three required elements of control (manipulation, randomization, and comparison/control) to isolate cause-and-effect relationships between variables

80
Q

correlations

A

examines the relationship between variables by measuring pairs of score for each individual. This method can determine whether a relationship exists between variables, but it lacks the appropriate controls needed to demonstrate cause and effect.

81
Q

quasi experiments

A

this study does not include a manipulated independent variable and it lack a comparison/control group

82
Q

quasi independent variable

A

preexisting variable that is often a characteristic inherent to an individual, which differentiates the groups or conditions being compared in a research study. Because the levels of the variable are preexisting, it is not possible to randomly assign participants to groups

83
Q

population

A

set of all individuals, items, or data of interest. This is the group about which scientist will generalize

84
Q

sample

A

set of individuals, items, or data selected from a population of interest.

85
Q

independent variable

A

variable that is manipulated in an experiment. This variable remains unchanged (or “independent”) between conditions being observed in an experiment. The specific conditions of the IV are referred to as levels.

86
Q

dependent variable

A

the variable that is measured in each group of a study, and it is believed to change in the presence of the independent variable. It is the “presumed effect.”

87
Q

descriptive statistics

A

procedures used to summarize, organize, and make sense of a set of scores called data. They are typically presented graphically or in tables.

88
Q

inferential statistics

A

procedures used that allow researchers to infer or generalize observations made with samples to the larger population from which they are selected

89
Q

population parameter

A

a characteristic (usually numeric) that describes a population. **Will usually never have this!

90
Q

sample statistic

A

a characteristic (usually numeric) that describes a sample

91
Q

nominal

A

are measurements in which a number is assigned to represent something or someone (a name, has no numerical properties)

92
Q

ordinal

A

measurements that convey order or rank alone (does not tell us space between ranks)

93
Q

interval

A

measurements that have no true zero and are distributed in equal units (equidistant, ex. Likert Scale)

94
Q

ratio

A

measurements that have a true 0 and are distributed in equal units

95
Q

qualitative

A

varies by class. This variable is often represented as a label and describes nonnumeric aspects of phenomena.

96
Q

quantitative

A

varies by amount. This variable is measured numerically and is often collected by measuring or counting.

97
Q

continuous

A

is measured alone a continuum at any place beyond the decimal points, can thus be measures in fractional units (ex. Time)

98
Q

discrete

A

measured in whole units or categories that not distributed along a continuum

99
Q

central tendency

A

statistical measures for locating a single score that is most representative or descriptive of all scores in a distribution

100
Q

mean and when best to use

A

also called the average) is the sum of a set of scores in a distribution, divided by the total number of scores summed.
**Best to use the mean in a normal distribution and with interval and ratio scales. In normal distributions, the mean median, and mode are all the same

101
Q

median and when best to use

A

is the middle value in a distribution of data listed in numeric order (aka the 50th percentile)
**Best to use the median with data that had outliers, is skewed and with ordinal scales

102
Q

mode and when best to use

A

is the value in a data set that occurs most often or most frequently (can have multiple modes)
**Best to use the mode when working with a modal/bimodal distribution (curve sinks in the middle) and with nominal scales

103
Q

normal distribution

A

bell-shaped) a theoretical distribution in which scores are symmetrically distributed above and below the mean, the median, and the mode at the centers of the distribution

104
Q

skewed distribution

A

distribution of scores that includes scores that fall substantially above or below most other scores in a data set

-Positive Skew: a distribution of scores that includes scores that are substantially larger than most other score (tail/flat part will be on the positive/ right side)
-Negative Skew: a distribution of scores that includes scores that are substantially smaller than most other scores (tail/flat part will be on the negative/ left side)

105
Q

bimodal distribution

A

a distribution of scores in which two scores occur most often or most frequently (a high hump where one mode is, a dip in the center (mean/median), and then another high hump where the second mode is, refer to pg. 95 in Privitera book for visual so you don’t have to keep listening to me talk about humps)

106
Q

variability

A

is a measure of the dispersion or spread of scores in a distribution that ranges from 0 to positive infinity

107
Q

range

A

is the difference between the largest value (L) and smallest value (S) in a data set.

108
Q

interquartile range

A

is the range of values between the upper (Q3) and the lower (Q1) quartiles of a data set

Lower Quartile: the median value of the lower half of a data set at the 25th percentile of a distribution

Upper Quartile: is the median value of the upper half of the data set at the 75th percentile of a distribution

109
Q

SS (sum of squares)

A

is the sum of the squared deviations of scores from their mean. The SS is the numerator in the variance formula

110
Q

variance

A

measure of variability for the average squared difference that scores deviate from their mean

111
Q

standard deviation

A

measure of variability for the average distance that scores deviate from their mean. It is calculated by taking the square root of the variance.

112
Q

z score

A

value on the x-axis of a standard normal distribution. The numerical value of a z score specifies the distance or the number of standard deviations that a value is above or below the mean

113
Q

characteristics of normal distribution

A

-Mathematically defined
-Theoretical (no distribution is perfect, it is approximate)
-Mean, median, and mode all in the 50th percentile
-It is symmetrical
-The total area under the curve is equal to 1 or 100%

114
Q

sampling distribution

A

a distribution of samples means or sample variances that could be obtained in samples of a given size from the same population
**Sampling distribution of sampling means is an unbiased estimator and follows the central limit theorem

115
Q

unbiased estimator

A

any sample statistic, such as sample variance when we divide SS by n-1, obtained from a randomly selected samples that equals the value of its respective population parameter, such as population variance, on average

116
Q

central limit theorem

A

explains that regardless of the distribution of scores in a population, the sampling distribution of samples means selected at random from that population will approach the shape of a normal distribution, as the number of samples in the sampling distribution increases

117
Q

standard error of the mean

A

the standard deviation of a sampling distribution of sampling means. It is the standard error or distance that sample mean values deviate from the value of the population mean

118
Q

sampling error

A

the extent to which sample means selected from the same population differ from the mean of sampling distribution of sample means (SDSM)
-The question is: is you randomly select one sample mean from the sample, how close will it be to the SDSM?
-Therefore, sampling error is the measurement of how narrow you SDSM is so a lot of error will mean that when taking a random sample mean, that number will be far off from the SDSM

119
Q

factors that decrease standard error

A

The larger the population standard deviation, the larger the standard error
-As numerator gets bigger, quantity as a whole gets bigger
-The wider the distribution, the less likely you are to get a random sample mean that is close to the SDSM

As sample size increases, standard error decreases
-As denominator gets bigger, quantity gets smaller
-Narrower distribution

bigger population = less error

120
Q

z scores for sample means

A

tells you the probability of sampling a particular mean from a particular population (NEED population standard devation)

121
Q

hypothesis testing

A

a method for testing a claim or hypothesis about a parameter in a population, using data measured in a sample. In this method, we use test a hypothesis by determining the likelihood that a sample statistic would be selected if the hypothesis regarding the population parameter were true
-Goal: to see if difference are significant

122
Q

null hypothesis

A

is a statement about a population parameter, such as the population mean, that is assumed to be true

123
Q

alternative hypothesis

A

a statement that directly contradicts a null hypothesis by stating that the actual value of a population parameter is less than, greater than, or not equal to the value states in the null hypothesis

124
Q

level of significance

A

is a criterion of judgment upon which a decision is made regarding the value stated in a null hypothesis. The criterion is based on the probability of obtaining a static measured in a sample if the value stated in the null hypothesis were true (p value)

125
Q

significance

A

describes a decision made concerning a value stated in the null hypothesis. When the null hypothesis is rejected, we reach significance. When the null hypothesis is retained, we fail to reach significance.

126
Q

type 1 error

A

is the probability of rejecting the null hypothesis that is actually true. Researchers directly control for the probability of committing this type of error (using alpha). So, saying there is significance and rejecting the null when the null was actually true (false positive)

127
Q

alpha level

A

the largest probability of committing a Type I error than we will allow and still decide to reject the null hypothesis (usually .05)

128
Q

type 2 error

A

is the probability of retaining a null hypothesis that is actually false. So, retaining the null and saying your results are not significant when really the alternative hypothesis is true (false negative)

129
Q

power

A

the probability that a randomly selected sample will show that the null hypothesis is false when the null hypothesis is indeed false

130
Q

critical value

A

a cutoff value that define the boundaries beyond which less than 5% of sample means can be obtained if the null hypothesis is true

131
Q

z statistic

A

an inferential statistic used to determine the number of standard deviations in a standard normal distribution that a sample mean deviates form the population mean stated in the null hypothesis
**To do a z-test, you need a POPULATION standard deviation, which is very unlikely. That is why t test are more commonly used.

132
Q

effect size

A

a statistical measure od the size of an effect in a population, which allows researchers to describe how far scores shifted in the population, or the present of variance that can be explained by a given variable (Cohen’s d)

Cohen’s d effects: Small= d < 0.2, Medium= 0.2 < d < 0.8, Large= d > 0.8

133
Q

one tailed

A

alternative hypothesis is stated as specifically greater of less than, region of rejection is on either the upper or lower portion of sampling distribution, makes it easier to reject the null and has greater power BUT is harder to justify (everybody will be like… that’s some whack ass research)

134
Q

two tailed

A

alternative hypothesis is stated as not equal, region of rejection is upper AND lower portion of sampling distribution, ALWAYS USE TWO TAILS

135
Q

t statistic

A

an inferential statistic used to determine the number of standard deviations in a t distribution that a sample mean deviates from the mean value or mean difference stated in the null hypothesis

136
Q

assumptions for t test

A

Normality

Random Sampling

Independence

Equal Variances
-(larger s2 divided by smaller s2, should be less than 2 to show equality)

137
Q

one independent sample t test

A

a statistical procedure used to compare a mean value measured in a sample to known value in the population. It is specifically used to test hypothesis concerning the mean in a single population with an unknown variance (or other arbitrary value).

138
Q

two independent sample t test

A

a statistical procedure used to compare the mean difference between two independent groups. This test is specifically used to test hypotheses concerning the difference between two population means, where the variance in one or both populations is unknown

139
Q

related samples

A

participants are related

140
Q

repeated measures design

A

research design in which the same participants are observed in each treatment

141
Q

pre post test design

A

a types of repeated-measures design in which researchers measure a dependent variable from participants before and after a treatment

142
Q

within subjects design

A

type of repeated-measures design in which researchers observe the same participants across many treatments but not necessarily before and after a treatment

143
Q

matched pairs design and the two ways it can be done

A

research design in which participants are selected and then matched based on common characteristics or traits

Experimentally: ex. measure intelligence and then match the two participants who scored the highest (put one in the drug group and one in the placebo group)

Naturally: ex. testing twins

144
Q

advantages for selecting related samples

A

More practical

Minimizes standard error
-Most important advantage, it reduces standard error by removing individual differences and isolating effect

Increases power

145
Q

disadvantages for selecting related samples

A

Requires individuals to be exposed to more than one treatment
-The 1st test may influence the score of the 2nd test (aka order effects)
-People get tired the more tests you give them

146
Q

one sample z test

A

a statistical procedure used to test hypothesis concerning the mean in a single population with a known variance

147
Q

independent sample

A

a type of sample in which different participants are independently observed one time in each group

148
Q

one independent sample t test

A

only one group of participants that are given treatment and then compared to the population mean or some arbitrary value

149
Q

two independent sample t test

A

participants in each group or sample are unrelated, selected from 2 or more populations, or single population with random different groups, then the groups are compared to each other

150
Q

related or dependent sample

A

participants are related

151
Q

repeated measures design

A

is a research design in which the same participants are observed in each treatment. Two types of repeated-measures designs are the pre-post design and the within-subjects design

152
Q

pre-post design

A

types of design in which researchers measure a dependent variable for participants before (pre) and after (post) a treatment

153
Q

within subject design

A

is a type of repeated-measures design in which researchers observe the same participants across many treatments but not necessarily before and after a treatment

154
Q

matched pairs design

A

a within-subjects research design in which participants are selected and then matched, experimentally or naturally, based on common characteristics or traits

Experimentally: ex. Measure intelligence then match the two participants who scored the highest with each other and so on

Naturally: ex. Match based on genetics (twins)

155
Q

difference score

A

score or value obtained by subtracting one score from another. In a related-samples t test, difference score are obtained prior to computing the test statistic

156
Q

error

A

a test, this refers to any unexplained difference that cannot be attributed to, or caused by, having different treatments. The standard error of the man is used to measure error or unexplained differences in a statistical design.

157
Q

levels of the factor

A

symbolized as K, are the number of groups or different ways in which an independent or quasi-independent variable is observed

158
Q

analysis of variance (ANOVA)

A

statistical procedure used to test hypothesis for one or more factors concerning the variance among two or more group means (k>2), where the variance in one or more populations is unknown

159
Q

one way between subjects ANOVA

A

a statistical procedure used to test hypothesis for one factor with two or more levels concerning the variance among group means. This test is used when different participants are observed at each level of a factor and the variance in any one population is unknown.

160
Q

source of variation

A

variation that can be measures in a study. In the one-way between-subjects ANOVA, there are two sources of variation: variation attributed to difference between group means and variation attributed to error.

161
Q

within groups variation

A

is the variation attributed to mean differences with each group. This source of variation cannot be attributed to or caused by having difference groups and is therefore called error variation (this is bad variation)

162
Q

between groups variation

A

variation attributed to the mean differences between groups (this is good variation). This is where you will see treatment effects.

163
Q

post hoc test

A

statistical procedure computed following a significant ANOVA to determine which pair or pairs of group means significantly differ. This test is only necessary when k >2 because multiple comparisons are needed. When k=2, only one comparison is made because only one pair of group means can be compared.

164
Q

factorial design

A

research design in which participants are observed across the combination of levels of two of more factors

165
Q

two way ANOVA

A

statistical procedure used to test hypotheses concerning the variance of groups created by combining the levels of two factors. This test is used when the variance in any one population is unknown.

166
Q

1 between 1 within design or mixed design

A

research design in which different participants are observed at each level of the between-subjects factor and the same participants are observed across the levels of the within-subjects factor.

167
Q

cell

A

combination of one level from each factor. Each cell is a group in a research study

168
Q

complete factorial design

A

research design in which each level of one factor is combined or crossed with each level of the other factor, with participants observed in each cell or combination of levels

169
Q

main effect

A

source of variation associated with mean differences across the levels of a single factor. In the two-way ANOVA, there are two factors and therefore two main effects: one for Factor A and one for Factor B.

170
Q

interaction

A

source of variation associated with the variance of group means across the combination of levels of two factors. It is a measure of how cell means at each level of one factor change across the levels of a second factor.

171
Q

levels

A

number of different ways a variable is manipulated or measured

172
Q

conditions

A

unique combinations of different levels of different independent variables

173
Q

simple main effects test

A

are hypothesis tests used to analyze a significant interaction by comparing the mean differences or simple main effects of one factor at each level of a second factor

**Main effects and interactions occur independently of each other

174
Q

eye balling

A

add cell means then divide by how many cells, the marginal means will inform you about main effects and the cell means will inform you about interactions

175
Q

one way vs two way anova

A

one way: uses one variable (e.g. average heights of plants grown with different fertilizers)

two way: uses two independent variables, also assesses for interaction between the two independent variables (e.g. worker productivity based on gender and department)