Chapter 9 Flashcards

1
Q

Simple vs Composite Hypothesis

A
  • A simple hypothesis completely specifies the probability distribution
  • A composite hypothesis does not fully specify the distribution;
    For example, H₀: μ = μ₀ is a simple hypothesis, whereas H₀: μ > μ₀ is a composite hypothesis.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Type I & Type II errors & Powers

A
  1. Type I error: rejecting H₀ when it is true.
  2. Type II error: failing to reject H₀ when it is false.
  3. The Power of the test is thee probability of rejecting H₀ when it is false (1-β). For simple hypotheses the power is a single value, but for composite hypotheses it is a function being defined at all points in the alternative hypothesis.
    * The level of significance of the test, denoted α, is the probability of committing a Type I error
    * The probability of committing a Type II error is denoted β
    * An ideal test would minimize both α and β, but in practice, this is not fully achievable.

In Binary Classifications
(a common situation in healthcare as well as in machine learning contexts) Here, rather than gathering a data sample consisting of multiple observations to assess whether a (population-level) hypothesis holds, a decision is required for each individual observation.
In a medical context, the classification is into healthy and diseased based on a binary test result. In these contexts:
A Type I error, known as a false positive, occurs when a healthy individual receives a positive test result; and
A Type II error, known as a false negative, occurs when a diseased individual tests negative for the disease.
Sensitivity=Power (β)=No of True Positives/No of actual positives
Specificity=1-α = No of True Negatives/No of actual negatives

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Neyman-Pearson Lemma

A

If C is a critical region of size α and there exists a constant k such that L₀ / L₁ ≤ k inside C and L₀ / L₁ ≥ k outside C, then C is the most powerful critical region of size α for testing the simple hypothesis H₀: θ = θ₀ against H₁: θ = θ₁.

A Neyman-Pearson test rejects H₀ if:

(Likelihood under H₀) / (Likelihood under H₁) < critical value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

LIkelihood ratio test

A

The critical region (and test statistic) for the test is determined by setting an upper bound on the ratio (max L₀ / max L), where max L₀ is the maximum likelihood under the null hypothesis and max L is the overall maximum likelihood across all possible parameter values.

The likelihood ratio test rejects H₀ if:

(max likelihood under H₀) / (max likelihood under H₀ ∪ H₁) < critical value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Non-parametric test

A

a test which doesn’t make any distributional assumptions about the data or the test statistic. These tests are called non-parametric tests.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Chi Squared goodness of fit test

A

Chi-squared goodness-of-fit tests are used for categorical or count data.

The test compares observed frequencies with expected frequencies under a specified hypothesis, using the test statistic:

Σ ((Oᵢ - Eᵢ)² / Eᵢ),
where Oᵢ and Eᵢ are the observed and expected frequencies for category i.

Under the null hypothesis, this statistic follows an approximate chi-squared (χ²) distribution.

The degrees of freedom (df) are calculated as:

df = (number of categories - 1) - (number of estimated parameters).
The chi-squared approximation is more accurate when expected frequencies are not too small. A common rule is that all expected frequencies should be at least 5. If necessary, categories can be combined to meet this condition.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Contingency Table

A

Very similar to chi squared test
A contingency table is a two-way table used to classify sample items according to two categorical variables.

The goal is to test whether the two classification criteria are independent.

Null Hypothesis: H₀: The two classification criteria are independent.
The expected frequency for any cell is calculated as:

(Row total × Column total) / Table total
If the two criteria are independent, the expected frequency for cell (i,j) is:

Eᵢⱼ = (row totalᵢ × column totalⱼ) / table total.
The degrees of freedom (df) for a table with r rows and c columns is:

df = (r - 1)(c - 1).
This method is commonly used in tests for differences among two or more population proportions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly