Topics 15-19 Flashcards
Probability of independent, dependent and mutually exclusive events, formulae
The probability of an independent event is unaffected by the occurrence of other events, but the probability of a dependent event is changed by the occurrence of another event.
- Events A and B are independent if and only if: P(A | B) = P(A), or equivalently, P(B | A) = P(B)
- The probability that at least one of two events will occur is P(A or B) = P(A) + P(B) — P(AB).
- For mutually exclusive events, P(A or B) = P(A) + P(B), since P(AB) = 0.
Joint probability of two events
The joint probability of two events, P(AB), is the probability that they will both occur.
P(AB) = P(A | B) x P(B)
For independent events, P(A | B) = P(A), so that
P(AB) = P(A) xP(B)
Inferential statistics, defenition
Inferential statistics pertain to the procedures used to make forecasts, estimates, or judgments about a large set of data on the basis of the statistical characteristics of a smaller set (a sample).
Mode, definition
The mode is the value that occurs most frequently in a data set. A data set may have more than one mode or even no mode.
When a distribution has one value that appears most frequently, it is said to be unimodal. When a set of data has two or three values that occur most frequently, it is said to be bimodal or trimodal.
Geometric mean comparing to arithmetic mean
The geometric mean is always less than or equal to the arithmetic mean, and the difference increases as the dispersion of the observations increases.
The only time the arithmetic and geometric means are equal is when there is no variability in the observations (i. e., all observations are equal).
Expected value and some its properties
The expected value is the weighted average of the possible outcomes of a random variable, where the weights are the probabilities that the outcomes will occur.
Some properties:
- If X and Y are independent random variables, then E(XY) = E(X) x E(Y)
- If X and Y are NOT independent, then E(XY) ≠ E(X) x E(Y)
- E(XY) = E(X)*E(Y) + covariance(X,Y)
- If X is a random variable, then E(X2) ≠ [E(X)]2
Variance and some of its properties
Variance is defined as:
Var(X) = E [(X - μ)2]
Some properties:
- Var(X) = E[(X - μ)2] = E(X2) - [E(X)]2 where μ = E(X)
- If a and c are constants, then:
Var(aX + c) = a2 x Var(X) - If X and Y are independent random variables, then:
Var(X + Y) = Var(X) + Var(Y)
Var(X - Y) = Var(X) + Var(Y)
Covariance and some of its properties
Covariance is the expected value of the product of the deviations of the two random variables from their respective expected values.
Since we will be mostly concerned with the covariance of asset returns, the following formula has been written in terms of the covariance of the return of asset i, Ri, and the return of asset j, Rj
Cov(Ri,Rj) = E{[Ri - E(Ri)] [Rj - E(Rj)]}
This equation simplifies to:
Cov(Ri,Rj) = E(Ri*Rj) - E(Ri)xE(Rj)
Some properties:
- If a, b, c, and d are constants, then:
Cov(a + bX, c + dY) = b x d x Cov(X,Y) - Cov(Z, aX + bY) = a Cov(Z,X) + b Cov(Z,Y)
- If X and Y are NOT independent, then:
Var(X + Y) = Var(X) + Var(Y) + 2 x Cov(X,Y)
Var(X - Y) = Var(X) + Var(Y) - 2 x Cov(X,Y)
Correlation coefficient, definition
To make the covariance of two random variables easier to interpret, it may be divided by the product of the random variables’ standard deviations. The resulting value is called the correlation coefficient, or simply, correlation.
The relationship between covariances, standard deviations, and correlations can be seen in the following expression for the correlation of the returns for asset i and j:
Central moments (skewness, kurtosis)
Central moments are measured relative to the mean (i.e., central around the mean). The k-th central moment is defined as:
E(R-μ)k = Σi=1n pi(Ri - μ)k
where: pi is probability of event i, Ri is return associated with event i.
1. Since central moments are measured relative to the mean, the first central moment equals zero and is, therefore, not typically used.
2. The second central moment is the variance of the distribution, which measures the dispersion of data.
3. The third central moment measures the departure from symmetry in the distribution. This moment will equal zero for a symmetric distribution (such as the normal distribution).
The skewness statistic is the standardized third central moment. Skewness (sometimes called relative skewness) refers to the extent to which the distribution of data is not symmetric around its mean. It is calculated as:
skewness = E[(R - μ)3]/σ3
- The fourth central moment measures the degree of clustering in the distribution.
The kurtosis statistic is the standardized fourth central moment of the distribution. Kurtosis refers to the degree of peakedness or clustering in the data distribution and is calculated as:
kurtosis = E[(R - μ)<span>4</span>]/σ4
Kurtosis for the normal distribution equals 3. Therefore, the excess kurtosis for any distribution equals:
excess kurtosis = kurtosis - 3
Effect of Skewness on Mean, Median, and Mode
The key to remembering how measures of central tendency are
affected by skewed data is to recognize that skew affects the mean more than the median and mode, and the mean is “pulled” in the direction of the skew.
Properties of kurtosis
Kurtosis is a measure of the degree to which a distribution is more or less “peaked” than a normal distribution. Leptokurtic describes a distribution that is more peaked than a normal distribution, whereas platykurtic refers to a distribution that is less peaked (or flatter) than a normal distribution. A distribution is mesokurtic if it has the same kurtosis as a normal distribution.
A distribution is said to exhibit excess kurtosis if it has either more or less kurtosis than the normal distribution: the Poisson distribution has mean and variance equal to lambda, λ, and its excess kurtosis is elegantly given 1/λ such that it always has (slightly) heavy tails; the student’s t also has slightly heavy tails with excess kurtosis given by df/(df-2) when df > 4.
Coskewness and Cokurtosis
Previously, we identified moments and central moments for mean and variance. In a similar fashion, we can identify cross central moments for the concept of covariance. The third cross central moment is known as coskewness and the fourth cross central moment is known as cokurtosis.
Desirable statistical properties of an estimator, The Best Linear Unbiased Estimator
There are certain statistical properties that make some estimates more desirable than others. These desirable properties of an estimator are unbiasedness, efficiency, consistency, and linearity.
- An unbiased estimator is one for which the expected value of the estimator is equal to the parameter you are trying to estimate. For example, because the expected value of the sample mean is equal to the population mean [E(x̅)=μ], the sample mean is an unbiased estimator of the population mean.
- An unbiased estimator is also efficient if the variance of its sampling distribution is smaller than all the other unbiased estimators of the parameter you are trying to estimate. The sample mean, for example, is an unbiased and efficient estimator of the population mean.
- A consistent estimator is one for which the accuracy of the parameter estimate increases as the sample size increases. As the sample size increases, the sampling distribution bunches more closely around the population mean.
- A point estimate is a linear estimator when it can be used as a linear function of sample data.
If the estimator is the best available (i.e., has the minimum variance), exhibits linearity, and is unbiased, it is said to be the best linear unbiased estimator (BLUE).
** Desirable statistical properties of an estimator include unbiasedness (sign of estimation error is random), efficiency (lower sampling error than any other unbiased estimator), consistency (variance of sampling error decreases with sample size), and linearity (used as a linear function of sample data).
Parametric and nonparametric distributions
Probability distributions are classified into two categories: parametric and nonparametric.
Parametric distributions, such as a normal distribution, can be described by using a mathematical function. These types of distributions make it easier to draw conclusions about the data; however, they also make restrictive assumptions, which are not necessarily supported by real-world patterns.
Nonparametric distributions, such as a historical distribution, cannot be described by using a mathematical function. Instead of making restrictive assumptions, these types of distributions fit the data perfectly; however, without generalizing the data, it can be difficult for a researcher to draw any conclusions.
Uniform distribution, its mean and variance
Bernoulli distribution
A Bernoulli distributed random variable only has two possible outcomes. The outcomes can be defined as either a “success” or a “failure.” The probability of success, p, may be denoted with the value “1” and the probability of failure, 1 - p , may be denoted with the value “0”.
Variance = p*(1-p)
Binominal distribution
A binomial random variable may be defined as the number of “successes” in a given number of trials, whereby the outcome can be either “success” or “failure.”
Binomial tends to normal as n increases. It is a discrete distribution
Expected Value and Variance of a Binomial Random Variable
For a given series of n trials, the expected number of successes, or E(X), is given by the following formula:
expected value of X = E(X) = np
The intuition is straightforward; if we perform n trials and the probability of success on each trial is p , we expect np successes.
The variance of a binomial random variable is given by:
variance of X = np(l — p) = npq
The Poisson Distribution
Poisson tends to normal as lambda increases. It is a discrete distribution
Normal distribution and its density function
Confidence Intervals for a Normal Distribution