Research Design and Statistics Flashcards
What is the sequence of the scientific method?
- Form a hypothesis
- Operationally define the hypothesis: what will be measured to show results?
- Collect & analyze data
- Disseminate results
Independent Variable: Define
The variable that is manipulated by researchers
The variable that is thought to impact the dependent variable
Dependent Variable: define
The outcome variable
What is hypothesized to change as a result of the IV
Predictor & Criterion Variables: Define
**Predictor: **Essentially the same as IV, but it can’t be manipulated
E.g. gender, age
Criterion: essentially the dependent variable
This is for correlational research
Can a variable have levels in a study?
Yes, especially the independent variable
E.g. Male & Female could be levels of the predictor variable
No treatment/Med Only/Combined treatment could be levels of the IV for treatment group
Factorial Designs
These have multiple IV’s
E.g. 1 IV is treatment; 2nd IV is type of schizophrenia
If you look at the effects of all levels on each other, it becomes a factorial design
What gives a study Internal Validity?
If you can determine a causal relationship between the IV and DV
No/limited effects of extraneous variables
Internal Validity in Multiple Group Studies: what impacts it?
The groups must be comparable to control for extraneous/confounding factors
Internal Validity: History
What is it? Any external event that affects scores on the dependent variable
Example: learning environment between groups is different, w/ one being superior
Internal Validity: Maturation
What is it? an internal change that occurs in subjects while the experiment is in progress
Example: time may lead to intellectual development, or fatigue, boredom, hunger may impact it
Internal Validity: Testing
What is it? practice effects
Example: take an EPPP sample test, attend a course, and then retake an exam to see if the course helped improve. but it may be just knowing what to expect on the test
Internal Validity: Instrumentation
What is it? changes in DV scores that are due to the measuring instrument changing
Example: raters may gain more experience over time. This is why we need highly reliable measuring instruments
Internal Validity: What is Statistical Regression?
What is it? extreme scores tend to fall closer to the mean upon re-testing
Example: if you test severe rated depression people, just by nature they are likely to report as less depressed next time regardless of any IV
Internal Validity: Selection
What is it? Pre-existing subject factors that account for scores on DV
Example: Classroom A students may simply just be smarter than Classroom B students, so regardless of different interventions they will score better
Internal Validity: Differential Validity
What is it? drop out is inevitable, so if you have 2 diff groups and there are differences in the type of people who drop out from each group, it can affect int. validity
Example: studying a new SSRI, some people may experience a worsening of depression/SI while on it, and they drop out. Because they dropped out, the med may appear to have been more helpful than it truly is
Internal Validity: Experimenter Bias
What is it? researchers preconceived bias impacts how they interact with subjects, which impacts the subjects scores
AKA: experimental expectancy effect, rosenthal effect, pygmalion effect
Example: experimenter unintentionally communicates expectations to subject
Prevention: double-blind technique
Protecting Internal Validity: Random Assignment
Each person has equal chance of ending up in a particular group
Protecting Internal Validity: Matching
What is it? ID subjects who are matched on an expected confounding variable, and then randomly assign them to treatment/control group
Ensures that both groups have equal proportion of the confounding variable
Protecting Internal Validity: Blocking
What is it? make the confounding variable another IV to determine to what extent it may be impacting the DV
Allows you to separate the effects of a variable and see interactions
Protecting Internal Validity: Holding the Extraneous Variable Constant
What is it? only use subjects who match the same on the extraneous variable
Problem: results not generalizable to other groups
Protecting Internal Validity: Analysis of Covariance
What is it? a stat strategy that adjusts DV scores so that subjects are equalized in terms of status on extraneous variables
Pitfall: only effective for extraneous variables that have been identified by the researchers
External Validity: Define
The degree to which results of a study can be generalized to other settings, times, people
Threats to External Validity: Interaction between Selection & Treatment
What is it? effects of a treatmetn don’t generalize to other target populations
Example: may work with college students, but not non-college students
Threats to External Validity: Interaction between History & Treatment
What is it? effects of treatment don’t generalize beyond setting and/or time period the experiment was done in
Threats to External Validity: Interaction between Testing & Treatment
What is it? pre-tests may sensitize the subjects to the purpose of the research study
AKA: Pretest sensitization
Example: pre-test before a film designed to reduce racism. The group who viewed the film may be primed and more motivated to pay attention to the film, as opposed to those who may watch the film without a pretest
Threats to External Validity: Demand Characteristics
What are they? cues in the research setting that may tip subjects off to the hypothesis
People pleasers may act in ways to confirm the hypothesis, while others may act to disprove it
Threats to External Validity: Hawthorne Effect
What is it? research subjects may behave differently simply because they are participating in research
Threats to External Validity: Order Effects
AKA Carryover effects & Multiple Treatment Interference
What is it? DV is impacted by other aspects of the study
Example: subjects get three treatments, always in the same order. Last treatment may show the best results, but there’s no way of knowing if it’s just from that treatment, or from impacts of the previous two
Stratified Random Sampling
Protecting External Validity
Take a random sample from subgroups of a population
Example: random sample of different age groups
Cluster Sample
Protecting External Validity
The unit of sampling is a naturally occurring group of individuals
Example: residents of a city
Naturalistic Research
Protecting External Validity
Behaviour is observed and recorded in its natural setting
Reduces many external validity concerns, but has no internal validity
What is analogue research?
Protecting External Validity
Results of lab studies are used to draw conclusions about real-world phenomenon
E.g. Milgram’s obedience studies
Single and Double-Blind Research
Protecting External Validity
Single Blind: subjects don’t know what group they are in
Double Blind: neither subjects or research know what group they are in
Reduce demand characteristics, researcher bias and hawthorne effect
Counterbalancing
Protecting External Validity
Controls for order effects by ensuring variables are received in different order
Latin Square Design: order the administration of variables so that each appears only once in each position
True Experimental Research
Subjects randomly assigned to groups
Groups receive different levels of manipulated variable
Greatest for internal validity
Quasi Experimental Research
When to use? when random assignment is not possible
Example: studying a learning program that is being introduced to all grade 1 classes
Next best for internal validity
Correlational Research
What is it used for?
Does it have Internal Validity?
Internal Validity? correlational research has none
Use for? Prediction, esp. for variables that can’t be manipulated
Developmental Research: 3 types
**Goal: **Assessing variables over time
Longitudinal: same people studied over long time
* Pitfall: underestimate changes, bc its often those who drop out that have the most significant changes
Cross-Sectional: different groups of subjects, divided by age, are assessed at same time
* Pitfall: cohort effects lead to overestimation of differences (e.g. may not account for an aid a different generation had, which is responsible for helping memory)
Cross-Sequential: combines the two. Samples of diff groups are assessed more than once
Time-Series Design
What is it?
What are the benefits?
Take multiple measurements over time (e.g. multiple pretest/posttest) to assess effects of IV
Benefits: controls for threats to internal validity. You can add a control group to help with history effects
Example: smoking reduction program in school. degree of post test results can indicate if it was a confounding factor or a result of the program
Single Subjects Design
Can be one subject, or multiple that are treated as one group
Used for: behaviour modification research
Dependent variable measured multiple times during phases of the study (phase 1-no treatment/phase 2-treatment)
Single Subject Design: AB Design
Single baseline and single treatment phase
Phase 1: collect data on frequency of behaviour before treatment
Phase 2: give treatment, collect data on if it reduced behaviour
Single Subject Design: Reversal (Withdrawal)
Benefits: controls for extraneous factors, which AB does not
What does it do? give treatment, withdraw treatment and reassess, and then provide treatment again. If behaviour continues again without treatment, the effect was likely due to treatment
Types:
ABA: baseline -> treatment -> withdraw
ABAB: baseline -> treatment -> withdraw -> treatment
Multiple Baseline Design
When to use?
Types of baselines to use
Used when: reversal not possible for ethical reasons
It doesn’t involve withdrawal of treatment
Treatment applied sequentially
Multiple Baseline Across Behaviours: start with one behaviour, then use same treatment for another
Multiple Baseline Across Settings: home, school
Multiple Baseline Across Subjects: try treatment on another subject
Qualitative Research: Surveys
Types
Risks/Benefits
Cons: many threats to validity
Pros: can try to ensure random sample
Types: personal interviews, telephone surveys, mail surveys
Qualitative Research: Case Studies
Con: lack internal and external validity
Pro: thorough on one person
Useful as pilot studies that can ID variables to be studied in a more systematic manner
Qualitative Research: Protocol Analysis
What is it? research involving the collection and analysis of verbatim reports
Example: subject thinks aloud while doing something, which is then analyzed to look for themes/concepts evident as the subject performed the task
Scales of Measurement: Nominal Data
Unordered categories, none of which are higher than the others
E.g. male/female
Scales of Measurement: Ordinal Data
Provides info about the ordering of categories, but not specifics
E.g. agree, strongly agree, neutral, etc.
Scales of Measurement: Interval Data
Numbers are scaled at equal distances, but the scale has no absolute zero point
e.g. IQ scores, temperature
Multiplication or division not possible, but addition and subtraction are
Scales of Measurement: Ratio Data
Identical to interval, but they have an absolute zero
E.g. dollar amounts, time, distance, height, weight, frequency of behaviours per hr
What does a Frequency Distribution provide? How are they displayed?
A summary of a set of data
tables, bar graphs, histograms
Normal Distribution
Symmetrical, half scores above mean and half below
Most scores are close to mean
Skewed Distributions
May happen with ceiling/floor effects
Negatively Skewed: has a tail on the left. Indicates easy test
Positively Skewed: has a tail on the right. Indicates difficult test
Measures of Central Tendency: the mean
Arithmetic average
Add all values and divide by n
Con: sensitive to extreme values
Measures of Central Tendency: the median
What is it? The middle value of data when ordered from lowest to highest (Md)
Odd groups: literally the middle number
Even groups: mean of the two middle numbers
Pros: not as affected by extreme scores, so good for skewed distributions
Measures of Central Tendency: the mode
What is it? the most frequent value in a set of numbers
May have multiple modes (bimodal/multimodal)
Relationship between the Mean, Median & Mode
Normal Distribution: all equal
Positively Skewed Distribution: mean higher than median, median higher than mode
Negatively Skewed Distribution: mean is less than median, mode is more than median
Measures of Variability: the range
What is it? the difference between the highest and lowest scores
Cons: impacted by extremes, so doesn’t give accurate representation of the distribution
Measures of Variability: The Variance
What is it? The average of the squared differences of each observation from the mean
For me: Get the mean. How far is each score from the mean? Square that distance, and then add them all up. Take an average of the sum. This is variance.
What to know?
1. measure of variability of distribution
2. many stat tests use it in formulas
3. It’s equal to the square of the SD