Chapter 5: Joint Probability Distributions Flashcards
Joint Probability Mass Function for Discrete Variables
Joint Probability Mass Function for Continuous Variables
Joint Probability Density Function for Continuous Variables
Example of Joint Probability Density
Marginal Densities in example

Independent Random Variables
Independent Random Variables Explained
- The definition says that two variables are independent if their joint pmf or pdf is the product of the two marginal pmf’s or pdf’s.
- Intuitively, independence says that knowing the value of one of the variables does not provide additional information about what the value of the other variable might be.
- Independence of X and Y requires that every entry in the joint probability table be the product of the corresponding row and column marginal probabilities.
- Independence of two random variables is most useful when the description of the experiment under study suggests that X and Y have no effect on one another.
- Then once the marginal pmf’s or pdf’s have been specified, the joint pmf or pdf is simply the product of the two marginal functions. It follows that
Conditional Distributions
Expected Value
Covariance
- Thus, the covariance of and provides a measure of the degree to which two variables X and Y tend to “move together”:
- a positive covariance indicates that the deviations of X and Y and from their respective means tend to have the same sign;
- a negative covariance indicates that deviations of X and Y from their respective means tend to have opposite signs.
Correlation

Correlation Coefficient

Properties of the Correlation Coefficient

Joint PMF of Two Discrete Random Variables
Marginal PMF of Two Discrete RVs
Example: Joint Probability Distribution of 2 Discrete RVs
Example: Joint Probability Distribution of 2 Discrete RVs (contd.)
Finding Marginal Distributions from Previous Example
Two Continuous Random Variables
If A is the two-dimensional rectangle
If A is the two-dimensional rectangle
Example 3
Example 3 contd.
Marginal PDF of Two Continuous RVs
Joint Probability Distributions - Expected Values
Conditional Distributions
Conditional PDF and PMF
Example 12
Example 12: Conditional Probability Function
- The probability that the walk-up facility is busy at most half the time given that X = .8 is then
•The expected proportion of time that the walk-up facility is busy given that X = .8 (a conditional expectation) is,
Statistics and Their Distributions
Statistics and Their Distributions (contd.)
Example 19
Example 19 contd.
Example 19 - Table 5.1
- Notice first that the 10 observations in any particular sample are all different from those in any other sample.
- Second, the 6 values of the sample mean are all different from one another, as are the 6 values of the sample median and the 6 values of the sample standard deviation.
- The same is true of the sample 10% trimmed means, sample fourth spreads, and so on.
- Furthermore, the value of the sample mean from any particular sample can be regarded as a point estimate (“point” because it is a single number, corresponding to a single point on the number line) of the population mean mew, whose value is known to be 4.4311.
- None of the estimates from these six samples is identical to what is being estimated.
- The estimates from the second and sixth samples are much too large, whereas the fifth sample gives a substantial underestimate.
- Similarly, the sample standard deviation gives a point estimate of the population standard deviation.
- All 6 of the resulting estimates are in error by at least a small amount
Example 19 Summary
In summary,
- the values of the individual sample observations vary from sample to sampleèso will, in general, the value of any quantity computed from sample data; and
- the value of a sample characteristic used as an estimate of the corresponding population characteristic will virtually never coincide with what is being estimated.
Statistics and Their Distributions
- A statistic is any quantity whose value can be calculated from sample data.
- Prior to obtaining data, there is uncertainty as to what value of any particular statistic will result.
- Therefore, a statistic is a random variable and will be denoted by a uppercase letter; a lowercase letter is used to represent the calculated or observed value of the statistic.
Statistics and Their Distributions (contd.)
Random Samples:
The Probability Distribution of any particular statistics depends on:
- the population distribution
- the sample size n
- the method of sampling
Simple Random Sample
- Advantage of simple random sampling method:
- probability distribution of any statistic can be more easily obtained than for any other sampling method
Deriving the Sampling distribution of a statistic:
Two general methods for obtaining information about a statistic’s sampling distribution:
- Using calculations based on probability rules
- By carrying out a simulation experiment.
Deriving a Sampling Distribution: Probability Rules
- Probability rules can be used to obtain the distribution of a statistic provided that :
- it is a “fairly simple” function of the Xi’s and either there are relatively few different X values in the population ;
OR
* the population distribution has a “nice” form.
Typical procedure (see Example 20):
(i) construct all possible samples of the desired sample size,
(ii) calculate the value of the statistic of interest for each sample and associated probability
(iii) construct the pmf of the statistic of interest.
Deriving a Sampling Distribution: Simulation Experiments
- This method is usually used when a derivation via probability rules is too difficult or complicated to be carried out.
- Such an experiment is virtually always done with the aid of a computer.
Deriving a Sampling Distribution: Simulation Experiments
Simulation Experiment: Density curve of Normal rv X with mean=8.25 and s.d. = 0.75

Simulation Experiment Histograms
Notable features of the histograms showing the sampling distribution of the mean:
- Shape –>All reasonably normal
- Center–> Each histogram approximately centered at mew= 8.25.
- Spread–>The larger the value of n, the more concentrated is the sampling distribution about the mean value. This is why the histograms for n = 20 and n = 30 are based on narrower class intervals than those for the two smaller sample sizes.
Simulation Experiment Contd.
The Distribution of the Sample Mean
Properties of Sample Mean and Sample Sum
Example 25
Example 25 contd.
The Central Limit Theorem
Central Limit Theorem (contd.)
Convergence of means from U[-1,1] to a normal shape
Histograms of raw means of samples from U[-1,1]
Apply Central Limit Theorem
Example 26
Implications of CLT
Linear Combinations and their means
Variances of linear combinations
The Difference between Random Variables
The Case of Normal Random Variables