Key Terms part 2 Flashcards
What is Kurtosis?
It is a measure of the degree to which a distribution is “peaked” or flat, in comparison to a normal distribution whose graph is characterized by a bell shaped appearance. This is part of the shape – one of the three characteristics that completely describe any distribution.
What are the three most commonly used measures of central tendency?
Mode, Median and Mean. The goal of measuring central tendency is to describe a distribution of scores by determining a value/s that identifies the center of the distribution.
What statistics do you use when hypothesis testing for t statistics ? What is included in the 4 part process?
1 - state the hypotheses (H0 and H1, including the alpha level, for ex: .05)
2 - set the critical region for z, using the chart
3 - compute the statistics:
First, calculate the standard error
Then, compute the test statistic
4 - compare the value you computed for z to your decision criteria. Make a decision regarding your hypotheses.
How do you report results following hypothesis testing? When are results significant?
If a result is significant, it is very unlikely to occur when the null hypothesis is true. You can reject the null hypothesis.
What is a shape of the t-distribution?
It is usually flatter and more variable than a normal z distribution, because the bottom of the formula (the sample variance or s2) changes from one sample to the next, meaning the estimated standard error also varies. In a z distribution, the bottom of the formula does not change.
T statistics are more variable than z scores. T statistics are flatter and more spread out.
What are the two assumptions regarding hypothesis testing with t statistic?
- The values in the sample must consist of independent observations, where two events (or observations) are independent if the occurrence of the first event has no effect on the probability of the second event.
- The population that is sampled must be normal. This is important with very small samples. This assumption can be violated with larger samples, without affecting the validity of the hypothesis test.
When hypothesis testing with t statistic, when do you reject the null hypothesis?
You reject the null hypothesis when the difference between the data and the hypothesis (numerator) is much greater than expected (denominator) and we obtain a large value for t. We conclude the data is not consistent with the hypothesis and we reject H0.
When hypothesis testing with t statistic, when do you fail to reject HO?
When the difference between the data and the hypothesis is small relative to the standard error, we obtain a t statistic near zero and we fail to reject H0.
What is degrees of freedom?
Df = n-1
It is defined as the number of observation in the data that are free to vary when estimating statistical parameters. (hat example).
What is the independent measures design?
A research design that uses a separate group of participants for each treatment condition. Also called between subjects research design.
What is the repeated measures design?
A research design, also called within subject design, is one in which the dependent variable is measure two or more times for each individual in a single sample. The same group of subjects is used in all treatment conditions.
Main advantage is that it uses exactly the same individuals in all treatment conditions. Sometimes researchers will try to approximate this with a matched-subjects design in which each individual from one sample is matched with an individual in another sample, so that they are equivalent in the variable that the researcher is trying to control.
What is the confidence interval?
It is the range of values that you expect your estimate to fall between a certain percentage of the time if you run your experiment again or re sample the population in the same way.
What does the confidence interval mean for hypothesis testing?
A mean difference of zero is exactly what would be predicted by the null hypothesis if we did a hypothesis test.
What is the independent measures t test?
It compares two groups, to see if there is any difference. All formulas are doubled. You’ll need to compute the standard error for both.
What are the 3 assumptions regarding the independent-measures t-formula?
- The observations within each sample must be independent.
- The two populations from which samples are selected must be normal.
- The two populations from which the samples are selected must have equal variance. <– This is referred to as homogeneity of variance, it is most important when there is a large discrepancy between sample size. SPSS uses Levene’s Test for equality of variances.
What is the Levene’s Test?
Levene’s Test tests the equality of variances between two populations that you are comparing in your independent-measures t formula. If the test is significant (p .05) the variances are not equal and you should use the t statistic computed with equal variances not assumed and then you should use the second row.
What is a constraint of a T-test?
You are just comparing two groups. If you have 3 groups or levels to test, you should use ANOVA, which gives you a bigger, broader picture.
What is ANOVA?
Analysis of variance is a hypothesis testing procedure that is used to evaluate mean differences between two or more treatments/populations. It uses qualitative variables. The major advantage is that it can be used to compare two or more treatments.
ANOVA asks if there are differences between the groups and also looks at variance within each group.
You use the F distribution/statistic with ANOVA.
What is sum of squares?
It is the difference between mean value of the sample and a certain value. It gives you an idea of how much variance you’ll have.
s2 or ss = ss/n-1
What is the difference between an independent variable and a quasi-independent variable?
Independent variable: a research manipulates a variable to create the treatment conditions (ex Treatment A, B, C)
Quasi-Independent variable: a researched uses a non manipulated variable to designate groups (ex occupational status)
What is the factor?
In ANOVA, the variable that designates the groups being compared is called a factor. The conditions or values that make up a factor are called the levels of the factor.
What are the example statistical hypotheses for ANOVA?
If you are comparing three different conditions, the hypotheses would be as follows:
What is the test statistic for ANOVA?
The F ratio/ distribution.
If F is greater than 1, we will reject the null hypothesis.
What is the goal of ANOVA and what are the steps?
The final goal is F ratio.
Step 1 - calculate the sum of squares
2 - compute the mean
3 - compute the F statistic and compare to F distribution chart. If the number is lower than the number we find in the table, we reject the null hypothesis.
What is the logic of ANOVA?
The total variability includes both the between treatment variance and the within-treatment variance.
When looking at the F-ratio, what are you considering?
- Numerator: Degrees of freedom between
- Denominator: degrees of freedom within
When there are no systematic treatment effects, the differences between treatments (in the numerator) are entirely caused by random, unsystematic factors.
If the numerator and denominator roughly equal, the F ratio should be around 1, showing no variance and H0 is true.
In ANOVA the denominator of F ratio is called error term.
When the treatment does have an effect, and there are systematic differences between samples, then the numerator should be noticeably larger.
F values are always positive numbers.
What is degrees of freedom?