2 Flashcards
- What is DV?
The proposed effect, outcome variable, measure not manipulated in exp
- Null Hypothesis Signifiance Tesitng (NHST) computes proability of
Probability of null hypothesis being true (referred as p-value) and computing statistic and how likely statistic has that value by chance along
- What is misconceptions of NHST?
A significant result means effect is important, a non-sig result means null hypothesis is true and significant result means null hypothesis is false
- Effect size can help with NHST issue, what is effect size?
Effect size is quantiative measure of magnitude of experimental effect, larger effect size the stronger relationship between 2 variables and can be used to compare studies on basis on effect size
- How to calculate effect size using Cohen’s D
Mean 1 minus mean 2 divided by standard devation
- What distribution do you need for parametric tests?
Normal distribution which can be described by mean (central tendency) and SD (dispersion)
- Is mean good measure of central tendency?
Mean can be misleading measure of central tendency in skewed distributions as it can be greatly influenced by extreme scores
- Aside from mean as central tendency, median and mode can be used where: - (2)
Median is unaffected with extreme scores and used with ordinal, interval and raio data
Mode only used with nominal data and greatly subject to sample fluctuations and many distributions have more than one mode
- What happens in positively skewed distributions?
Mean is greater than median which is greater than mode (left skewed)
- What happens in negatively skewed distributions?
The mode is greater than median which is greater than mean (right skewed)
- What are tests of normality dependent on?
Sample size, and if you have a massive sample size have normality tests significant even when data is visually normally distributed (usually trust visual plot)
- Two ways a distribution can deviate from normal
Lack of symmetry (skew) and pointiness (kurotosis)
- What does kurotosis tell you?
It tells you how much of our data lies around tails of histogram and helps us to identify when outliers may be present in the data.
- What is difference between parametric and non-parametric tests? - 4
Parametric tests assume specific distributions, like the normal distribution, and require adherence to certain statistical assumptions, such as homogeneity of variances and independence of observations.
They tend to be more powerful when these assumptions are met, making them suitable for analyzing data that closely aligns with their requirements, such as interval or ratio data.
On the other hand, non-parametric tests make fewer distributional assumptions, making them robust and applicable to a wider range of data types, including ordinal or skewed data.
While non-parametric tests are generally less powerful than their parametric counterparts when assumptions are met, they provide reliable results in situations where assumptions are violated or when dealing with non-normally distributed data. These differences in assumptions and robustness make each type of test valuable in different research contexts
- Non-parametric equivalent of correlation
Spearman’s Rho or Kendall Tau
- If skewness values between -1 and 1 then but…
Its good … but is below -1 then negatively skewed and above 1 then positively skewed
- If skew and kurotsis is 0 then tell you your data has a
Normal distribution
- If your kurotsis value between -2 and 2 then but…
All good but..if less than -2 then platykurtic and above 2 then leptokurtic
- Correlationa study does not rule out presence of third variable (tertium quid) which can be ruled out using
RCTs to even out confounding variable between groups
- What is variance?
Average squared deviation of each number from its mean (SD squared)
- To get SD from variance you:
Square root variance
- What is central limit therom?
States sampling distribution of mean approaches normal distribution as sample sixe sizes and especially case for sample sizes over 30
- What is a type 1 error?
False positive so think there is a sig effect but there isn’t = alpha
- Wha is type II error?
False negative so much variance unaccounted for by the model so thinking there is sig effect but there is not = beta
- Acceptable level of type I error is:
Alpha level of 0.05 and alpha level means probability of making type 1 error
- Non-parametirc equivalent of multti-way repeated ANOVA
Loglinear analysis
- Non parametric equivalent of one-way repeated ANOVA
Friedman’s ANOVA
- Non-parametric equivalent of one-way independent ANOVA
Kruskall Wallis
- We accept results as true (accept H1) when
Less proability of test statistic happening by chance e.g., p less than 0.05 means low proability of obtaining at least extreme results given H0 is true
- Acceptable level of Type II error is probability of
Beta level (often 0.2)
- R = 0.1, d = 0.2 (small effect) mean
Effect explains 1% of variance
- R = 0.3, D = 0.5 (medium effect)
Effect accounts 9% of total variance
- R = 0.5, d = 0.8 (large effect)
Effect accounts for 25% of total variance
- Non-parametric equivalent of paired t-tests
Wilcoxon signed rank test compare two dependent groups of scores compute differences between socres of two condition, note sign (pos or negative), rank differences and sum ranks for positive and negattive
- Non-parametric equivalent of independent t-tests
Mann Whitney or Wilcoxon rank sum (rank data irrespective of group and see one group has higher ranking of scores than other) compare two independent groups of scores
- What is log-linear? - (2)
Extension of chi-square as investigate relationship between more than 2 categorical variables
Assum independence and expected counts greater than 5 and goodness of fit is non sig in loglinear tests hypothesis that frequencies predicted by model sig different from actual want to be non-sig
- In chi-sqaure: - 7
One categorical DV and one categorical IV with different participants at each predictor level
Compare observed frequencies from data with frequencies expected if no relation between two variables
Tests if there is a relationship between 2 categorical variables
Assumption like data values from simple random sample of interest, for each combination of levels of 2 variables need at least expected counts greater than 5 if not met then use Fisher’s exact test
Expected counts = comun total * row total divided by grand total
We calculate chi-square test of calculate difference from actual vs expected for each combination of 2 variables and square difference and divideexpected by expected value of combination and add all values
DF = (r-1)(c-1)
- What is power? - (3)
Proability of ability of test to find an effect assuming there is one in population
It is 1 minus beta or probability of making type II error
To increase power of study, you can increase sample size , probability of type 1 erorr or alpha be bigger so more power
- What is correlation?
Correlation measures strength and direction of linear relationship of 2 variables
- What is assumptions of correlation?
IV and DV continous
- What is covariance? - (5)
It gathers info if one variable covary with other so interested if one variable deviates from mean then maybe other variable deviate from mean similar way
It is calculated by (x – mean of x) (y-mean of y) divided by N-1 and add all the values
Positive covariance says as one variable deviates from mean, other variable deviates in same direction, negative is difference in direction
Problem with covariance is dependent on units and we standardise by dividng by product of SD of both variables sigma(x-mean of x)(y – mean of y)/ (N-1) sxsy (formula of pearson correlation)
Can’t say predictor uniquely explain outcome and its coefificents can be exhcangable
- What is assumptions of correlation? - (2)
Assumptions of correlation include linearity, homoscedasticity (constant variance), and bivariate normality;
correlation typically involves two continuous variables (IVs and DVs) and assumes a linear relationship between them.
- Note on biserial and point-biserial - (2)
Point biserial =discrete dichotomy whether someone is pregnant
Biserial = variable is continuous dichotomoy like passing or failing exam like someone may just pass whether some fail and some will excel
- What is regression?
Regression analyses relationship between DV and one or more IV and estimate impact of IV(s) on DV and different to correlaiton as adds a constant bo
- Assumptions of simple regression - 9
1 DV continuous and 1 continous predictor
Predictors have non-zero variance
Independence = All values of outcome come from different person
Linearity
Homoscedasticity = each value of predictor variance of error term is constant
Both linearity and homodesasticity tested of scatterplot of ZPRES against ZPRED
Independent errors = errors should be uncorrelated – durbin Watson
Normally distributed e
Multiple regression = no co linearity where two predictors correlated strongly with each other check via VIF and tolerance
- Closer SSM to SST then?
Better model can account for the data and smaller SSR must be
- What is SST?
Difference between observed data and mean value of Y
- What is SSM?
It is the sum of squares attributable to regression model which measures how much variation in DV explained by IV
What is SSR?
It is differences between observed values of DV and values predicted by regression model = measures unexplained variation in DV not accounted for by IV
- What does R^2 represent? - 4
It is the coefficient of determination and gives how well predictor explain variance in outcome
Gives overall fit of model
Adjusted R^2 is how well R^2 generalises to population and generally more smaller as adjust statistic based on number of IVs
R^2 explains variance SSM to total variance of SST
- What is mixed effect ANOVA? - (3
A mixture of between and within
Many IVs have measured and some measured with diff pps whereas other use same pps
Need at least two IVs for mixed design that are categorical and one 1 DV continuous
- What does F-raito? - (2)
Used in regression and ANOVA to assess overall sig of IVs in model and calculated by dividng MSM by SSR
If model is good large portion of variance MSM as compared to what is left – MSR
- In multiple regression it has:
2 or more predictors that are continuous
- What is hierarchical regression? - (2)
In this we are seeing if one model explains significantly more variance than other
As a general rule, predictors selected based on past work and experimenter decides what order in which predictors entered in model -> typically known predictors added first and new predictors add second
- In VIF and tolerance for MR:
VIF above 10 and tolerance (reciprocal of VIF) below 0.2 = issue with multicollinearity
- In Durbin-Watson :
- (2)
Closer to 2 then assumption met
If value less than 1 or greater than 2 than raises alarm bells
- What is a t-test?
A t-test is a statistical test used to compare the means of two groups and determine if there is a significant difference between them, taking into account the variability within each group
- In two-way Independent ANOVA, break SSM to:
SSA (variance explained by 1st IV) and SSB (variance explained by second IV) and its interaction (SSA * SSB)
- What is difference between partial and eta-squared?
This differs from eta squared in that it looks not at the proportion of total variance that a variable explains, but at the proportion of variance that a variable explains that is not explained by other variables in the analysis
- How is eta-squared calculated?
SSM/SST
- What are the 3 types of t-tests?
One-samples t-test (compare mean of sample data to known value), paired t-test and independent t-test
- What is assumptions of t-tests? - (7)
One DV continuous and one predictor categorical (interval) and two levels
Same (paried-t-test) and different (independent) for each predictor level
Normal distribution via frequency histogram, P-plot (Straight line) and non sig normality tests
No sig outliers
Homogenity of variances via Levene’s test (independent)
Independence – no relationship between groups (independent since scores coming from different people)
Related samples (paired-t-test)
- What is formula of calculating t statistic?
T = observed difference between sample means subtracted between expected differences between population mean (if H0 is true, typically 0) divided by estimate of standard error of difference between 2 sample mean
- Guidelines for interpreting partial eta-sqaured
0.01 = small effect
0.06 = medium effect
0.14 = large effect
- In one-way ANOVA,
Between = SSM
Within = SSR
- How to calculate paired- t-tests? - (2)
T = Mean differences between samples (D line) subtracted by differences in sample means expect to find between population means (uD) which is divided by standard error of differences (standard deviation of differences divided by square root of N)
If D line is greater than standard error of differences then differences we observed in sample is not chance result and caused by experimental manipulation
- How to calculate standard error?
SD / square root of N
- How to calculate Cohen’s D?
Mean of group 1 (i.e., control group )minus mean of group 2 divided by SD of control group (can be pooled SD but control group SD common used)
Independent vs paired
With independent-tests, different participants participate in different condition, pair of score differ not just of experimental manipulation but of other sources of variance like IQ etc.. whereas this individual differences is eliminated in paired t-tests
What is ANOVA? - 5
ANOVA = Analysis of variance
It is used to compare whether there is significant differences between three or more groups of their means
It provides F-statistic and associated p-values to assess sig of differences
But does not say where differences lie with F ratio = follow up with planned contrasts or post-hoc tests since omnibus test
Can be used to reduce familywise error (error on statistical tests on same data) and false positive by not conducting t-tests to compare every group
- What is assumption of one-way ANOVA? - 8
1 DV continuous, 1 predictor variable (one-way ANOVA but 2 predictor variables in two-way independent and two-way repeated ANOVA) that is categorical and more than 2 levels (ANCOVA)
Different participant in each predictor (one-way ANOVA + ANCOVA) same (one-way repeated ANOVA, two-repeated ANOVA).
No sig outliers (ANCOVA)
Normal distribution of DV for each level of categorical IV (ANCOVA)
Homogenity of variance (independent design, levene’s test not sig, one-way ANOVA + two-way independent ANOVA)
Independent of covariate (covariate as DV and IV as IV should be non-sig p-value) and treatment effect and homogeneity of regression slopes (- covariate has same relationship of outcome regardless of level of IV via scatterplot and if interaction between covariate and IV non-sig ANCOVA)
Spherecity (equality of variances of the differences between treatment levels , via non sig Maulchy’s test which is dependent on sample size – one and two-way repeated ANOVA, if sig then use corrections to produce valid F ratio such as green house geisser [conservative] or huynh feldt correction)
Independent of covariate and homogeneity as well as spherecity in mixed ANOVA
In repeated ANOVA,
SSB between, SSW = SSM and SSR
- When should Sikdmak correction used?
When concerned about loss of power associated with Bonferroni corrected values
- What happens in ANCOVA (Analyse of covariance)? - (3)
Compare Many means without increasing chance of type 1 error
Assume different participant in each predictor level
Measure covariates and include them in analysis of variance which reduce SSR (within group variance) and eliminate confounds
Bonferroni correction - (2)
Ensures type 1 error is 0.05 and reduces type 1 error (conservative) but also lacks statistical power (probability of type II error of false negative increase) so increase chance of missing genuine difference
Alpha level divided by number of comparison
Slide 1 - Intro
My name is Gitanjali Sharma. I will be talking about my UG dissertation project called: Predicting Speech in Noise Ability from Non-speech stimuli.
Slide 2 - 6 - SiN
SiN is the ability of how well individuals can hear speech in background noise.
For example, listening to a friend in a busy pub.
This ability varies with age and hearing loss.
The question is how to measure this ability? It is typically measured through SiN tests that use a sample of speech (like a word or sentence) with noise added to it.
The task is to recognize the word/sentence.
For example:
Slide 3 - (3) - Limitations
The disadvantage of these tests is that the speech content is typically recorded in a specific language (typically English) and thus give undue performance advantage to fluent/native speakers of the language and these tests can not be employed to non-fluent/non-native speakers as well as children whose language skills are not fully developed.
To overcome this limitation, we can ask the question of whether non-speech auditory stimuli can be used to predict the performance on SiN tests.
In the project, two non-speech stimuli were tested: 1) SFG (stochastic figure-ground) and, 2) Harmonic roving SFG (HRSFG).
Slide 4 - Stimul - (5)
The SFG stimuli consist of a pattern of coherent tones, which is perceived as a figure, against a background of non-coherent tones and sounds like this
The HRSFG is different from SFG in two aspects:
Frequencies in figure is harmonically related and,
The frequencies changing with time
These two features are more speech like
Slide 5 - Hypotheses and Method - (3)
This study was conducted in sound-proof booth with 54 participants in which they completed a hearing test (called pure-tone audiometry) as well as computer tasks such as a sentence in noise task, pattern discrimination task using HRSFG and figure discrimination task using SFG.
We expected that SFG and HRSFG performances will be individually correlated with SiN test performance
Additionally, HRSFG will explain the performance on SiN tests better than SFG
Slide 6 - Non-speech stimuli tasks - (2)
In each trial of figure discrimination task, participants heard two examples of SFG stimuli and had to say whether one of the two had a gap in the figure portion
In each trial of pattern discrimination task, participants heard two HRSFG stimuli and had to say whether they have a same or different pattern
Slide 7 - Results - (2)
Here are the results as you can see on the left panel is correlation between SFG and SiN thresholds and on the right is between SiN and HRSFG thresholds which both show significant correlations.
Using hierarchical regression, it was observed that HRSFG increased the explained variance by 15% in SiN tests as compared to SFG.
Slide 8 - (4)
The results show that non-speech auditory figure ground stimuli such as our SFG and HRSFG can predict SiN test performance
Furthermore, stimuli that captures aspect of speech such as harmonicity and time varying frequency, HRSFG, predict the performance better than static SFG stimuli
These findings support the hypothesis
Future research needs to explore whether there is a shared brain regions for processing non-speech stimuli like stochastic figure ground (SFG) and harmonic SFG as well as SiN stimuli used in SiN tests