Stats and Research Design Flashcards
Alpha
Alpha determines the probability of rejecting the null hypothesis when it is true; i.e., the probability of making a Type I error. The value of alpha is set by the experimenter prior to collecting or analyzing the data. In psychological research, alpha is commonly set at .01 or .05.
Alpha (Level of Significance)
Alpha determines the probability of rejecting the null hypothesis when it is true; i.e., the probability of making a Type I error. The value of alpha is set by the experimenter prior to collecting or analyzing the data. In psychological research, alpha is commonly set at .01 or .05.
Bell Curve Graph (z-scores, t-scores, percentile ranks, standard deviations)
Bell Curve Graph (attached)
BETWEEN-GROUPS DESIGNS
BETWEEN-GROUPS DESIGNS: Between-groups designs are experimental research designs that allow a researcher to assess the effects of the different levels of one or more IVs by administering each level or combination of levels to a different group of subjects.
CENTRAL LIMIT THEOREM
CENTRAL LIMIT THEOREM: The Central Limit Theorem is derived from probability theory and predicts that the sampling distribution of the mean (a) will approach a normal shape as the sample size increases, regardless of the shape of the population distribution of scores; (b) has a mean equal to the population mean; and (c) has a standard deviation equal to the population standard deviation divided by the square root of the sample size. This standard deviation is referred to as the standard error of the mean
Chi–Square Test
(Single–Sample And Multiple–Sample)
The chi–square test is a nonparametric statistical test that is used with nominal data (or data that are being treated as nominal data) – i.e., when the data to be compared are frequencies in each category. The single–sample chi–square test is used when the study includes one variable; the multiple–sample chi–square test when it includes two or more variables. (When counting variables for the chi–square test, independent and dependent variables are both included.)
Cluster Analysis
Cluster analysis is a multivariate technique that is used to group people or objects into a smaller number of mutually exclusive and exhaustive subgroups (clusters) based on their similarities – i.e., to group people or objects so that the identified subgroups have within–group homogeneity and between–group heterogeneity.
CORRELATION COEFFICIENT (PEARSON r, SPEARMAN RHO, POINT BISERIAL, BISERIAL, ETA)
CORRELATION COEFFICIENT (PEARSON r, SPEARMEN RHO, POINT BISERIAL, BISERIAL, ETA):
The correlation coefficient is a numerical index of the relationship (degree of association) between variables. The magnitude of the coefficient indicates the strength of the relationship; its sign indicates the direction (positive or negative).
The Pearson r is used when data on both variables represent a continuous scale,
Spearman rho is used when both variables are ranks, the point biserial coefficient is used when one variable is a true dichotomy and the other is continuous, the biserial coefficient is used when one variable is an artificial dichotomy and the other is continuous, and eta is used when the variables are both continuous and have a nonlinear relationship.
Cross–Validation/Shrinkage
Cross–validation refers to validating a correlation coefficient (e.g., a criterion–related validity coefficient) on a new sample. Because the same chance factors operating in the original sample are not operating in the subsequent sample, the correlation coefficient tends to shrink” on cross–validation. In terms of the multiple correlation coefficient (R)
Discriminant Function Analysis
Discriminant function analysis is the appropriate multivariate technique when two or more continuous predictors will be used to predict or estimate a person’s status on a single discrete (nominal) criterion.
Effect Size
An effect size is measure of the magnitude of the relationship between independent and dependent variables and is useful for interpreting the relationship’s clinical or practical significance (e.g., for comparing the clinical effectiveness of two or more treatments). Several methods are used to calculate an effect size including Cohen’s d (which indicates the difference between two groups in terms of standard deviation units) and eta squared (which indicates the percent of variance in the dependent variable that is accounted for by variance in the independent variable).
Effect Size
An effect size is measure of the magnitude of the relationship between independent and dependent variables and is useful for interpreting the relationship’s clinical or practical significance (e.g., for comparing the clinical effectiveness of two or more treatments). Several methods are used to calculate an effect size including Cohen’s d (which indicates the difference between two groups in terms of standard deviation units) and eta squared (which indicates the percent of variance in the dependent variable that is accounted for by variance in the independent variable).
Experimental Research
(True and Quasi–Experimental)
research involves conducting an empirical study to test hypotheses about the relationships between independent and dependent variables. A true experimental study permits greater control over experimental conditions, and its hallmark” is random assignment to groups. A quasi–experimental study permits less control.”
Experimentwise Error Rate
The experimentwise error rate (also known as the familywise error rate) is the probability of making a Type I error. As the number of statistical comparisons in a study increases, the experimentwise error rate increases.
External Validity (Pretest Sensitization, Reactivity, Multiple Treatment Interference)
External validity refers to the degree to which a study’s results can be generalized to other people, settings, conditions, etc. Threats include pretest sensitization (which occurs when pretesting affects how subjects react to the treatment), reactivity (which occurs when subjects respond differently to a treatment because they know they are participating in a research study), and multiple treatment interference (which occurs when subjects receive more than one level of an IV). Counterbalancing can be used to control multiple treatment interference and involves administering different levels of the IV to different groups of subjects in a different order.
Factorial ANOVA
The factorial ANOVA is the appropriate statistical test when a study includes two or more IVs (i.e., when the study has used a factorial design) and a single DV that is measured on an interval or ratio scale. It is also referred to as a two–way ANOVA, three–way ANOVA, etc., with the words two” and “three” referring to the number of IVs.”
Factorial Design (Main And Interaction Effects)
Factorial designs are research designs that include two or more factors” (independent variables). They permit the analysis of main and interaction effects: A main effect is the effect of a single IV on the DV
Independent and Dependent Variables
The independent variable (IV) is the variable that is believed to have an effect on the dependent variable and is varied or manipulated by the researcher in an experimental research study. Each independent variable in a study must have at least two levels. The dependent variable (DV) is the variable that is believed to be affected by the independent variable and is observed and measured.
Internal Validity
(Maturation, History, Statistical Regression, Selection)
Internal validity refers to the degree to which a research study allows an investigator to conclude that observed variability in a dependent variable is due to the independent variable rather than to other factors. Maturation is one threat to internal validity and occurs when a physical or psychological process or event occurs as the result of the passage of time (e.g., increasing fatigue, decreasing motivation) and has a systematic effect on subjects’ status on the DV. History is a threat when an event that is external to the research study affects subjects’ performance on the DV in a systematic way. Statistical regression is a threat when subjects are selected to participate because of their extreme status on the DV or a measure that correlates with the DV and refers to the tendency of extreme scores to regress to the mean” on retesting. Selection threatens internal validity when groups differ at the beginning of the study because of the way subjects were assigned to groups and is a potential threat whenever subjects are not randomly assigned to groups.”