Concepts Flashcards
Fixed effects
ALL treatment conditions of interest are included in an experiment. Fixed effects cannot be generalized beyond treatment conditions
Random effects
Experiment contains only a random sample of possible treatment conditions; can be generalized beyond treatment conditions in the experiment (provided representative treatment conditions)
Repeated measures design
Different treatment conditions utilize the same organism (in human experiments, it means all participants are subjected to every level of an independent variable), and so the resulting data are related (i.e. within-subject design)
Between-subjects deisgn
Experimental design in which different treatment conditions utilize different organisms (in human experiments, it means participants are only subjected to one level of an independent variable), and so resulting data are independent
Order effects
Potential confounding effect in which observed differences are due to the order of treatments as opposed to the treatments themselves
t-statistic
Student’s t is a test statistic with a known probability distribution (the t-distribution). In regression, it is used to test whether a regression coefficient, b, is significantly different from zero; in experimental work, it is used to test whether differences between two means are significantly different from zero
Independent t-test
Test using the t-statistic that established whether two means collected from independent samples differ significantly
Dependent t-test
test using the t-statistic that establishes whether two means collected from the same sample (or related observations) differ significantly
Chi-square distribution
probability distribution of the sum of squares of several normally distributed variables. It is used to 1) test hypotheses about categorical data and 2) test the fit of models to the observed data
Chi-square test
Can apply to any test statistic having a chi-square distribution, it generally refers to a pearson’s chi-square test of the independence of 2 categorical variables. Essentially it tests whether 2 categorical variables forming a contingency table (table representing cross-classification of 2 or more categorical variables) are associated
Probability distribution
a curve describing an idealized frequency distribution of a particular variable from which it is possible to ascertain the probability with which specific values of that variable will occur. For categorical variables, it is a formula yielding the probability with which each category occurs
Sum of squares (sum of squared errors)
Estimate of the total variability (spread) of a set of data. First the deviance (the difference between observed value of a variable and the mean) for each score is calculated, and then this value is squared. The SS is the sum of these squared deviances
Test statistic
a statistic (a quantity that describes a sample) for which we know how frequently different values occur. The observed value of such a statistic is used to test hypotheses.
bi (regression coefficient)
unstandardized regression coefficient. Indicates the strength of the relationship between a given predictor, i, and an outcome in the units of measurement of the predictor. It is the change in the outcome associated with a unit change in the predictor.
Bi (regression coefficient)
The standardized regression coefficient. Indicates the strength of the relationship between a given predictor, i, and an outcome in a standardized form. It is the change in the outcome (in standard deviations) associated with one standard deviation change in the predictor
Hypothesis
a prediction about the state of the world
Null hypothesis
Reverse of alternative hypothesis that your prediction is wrong and the predicted effect doesn’t exist
Alternative hypothesis
Prediction that there will be an effect (i.e., that your experimental manipulation will have some effect or that certain variables will relate to each other)
Sampling distribution
the probability distribution of a statistic; the distribution of possible values of a given statistic that we could expect to get from a given population; a distribution of statistics obtained by selecting all possible samples of a specific size from a population
Residual
the difference between the value a model predicts and the value observed in the data on which the model is based. When the residual is calculated for each observation, the resulting collection is called the “residuals”
Residual sum of squares
a measure of the variability that cannot be explained by the model fitted to the data. it is the total squared deviance between the observations predicted by whatever model is fitted to the data
Linear model
Model that is based upon a straight line
Model (explained) sum of squares
A measure of the total amount of variability for which a model can account. It is the (Total Sum of Squares - Residual Sum of Squares)
Fit
The degree to which a statistical model (a formalization of relationships between variables in the form of mathematical equations) is an accurate representation of some observed data
Simple regression
a linear model in which one variable/outcome is predicted from a single predictor variable that takes the form of Yi = (b0 + b1X1) + Ei, where Y is the outcome variable, X is the predictor, b1 is the regression coefficient associated with the predictor and b0 is the value of the outcome when the predictor is zero
F-ratio
a test statistic with a known probability distribution (the F-distribution). It is the ratio of the average variability in the data that a given model can explain to the average variability unexplained by the same model. It is used to test the overall fit of the model in simple/multiple regression, and to test for overall differences between group means in experiments
ANOVA
Analysis of variance; procedure that uses the F-ratio to test the overall fit of a linear model. In experimental research, this linear model tends to be defined in terms of group means, and the resulting ANOVA is therefore a test of whether group means differ