Week 3: Joint Distributions, Independent Random Variables, Conditional Distribution, Transformations of Random Variables Flashcards
Joint Distributions
Joint distributions describe the probability distribution of multiple random variables simultaneously, mapping outcomes in the probability space to a space combining all target spaces.
This allows for the study of their interdependence and marginal probabilities.
When would Binomial Random Variable be a joint distribution and when would it not be
Univariate Binomial Random Variable (not joint distribution): X ~ Binomial (n , p) , single test consisting of n independent trials
Multivariate (joint distribution): 2 tests - X1 ~ Binomial (n1 , p1), X2 ~ Binomial (n2 , p2). Ex: successes in sales in region A and successes in sales in region B
For which type of variable does one use Joint PMF
Discrete (die, coins, yes/no)
How to calculate Joint PMF Values:
Ex: Two tests - Biased coin toss (60% heads), 3 sided die (33% for 1,2,3)
P(X1 = Heads, X2=1) = 0.6(1/3)=0.2
P(X1=Heads, X2=2) = 0.6(1/3)=0.2
…..
P(X2=Tails, X2=3) = 0.4*(1/3)=0.133
Sum of all these probabilities = 1
Calculate the marginal PMF here of X1:
Two tests - Biased coin toss (60% heads), 3 sided die (33% for 1,2,3)
P(X1 = Heads, X2=1) = 0.6(1/3)=0.2
P(X1=Heads, X2=2) = 0.6(1/3)=0.2
…..
P(X2=Tails, X2=3) = 0.4*(1/3)=0.133
Marginal PMF of X1:
px1(Heads)=(0.2+0.2+0.2)=0.6
px2(Tails)=(0.133+0.133+0.133)=0.4
For which type of variable does one use Joint PDF
Continuous (weight, height, etc)
How to calculate the Joint PDF values
fxx(x1,x2) * (Difference between full range and specified range X1) * (Difference between full range and specified range X1)
ex: (1/2) (1-0.5)(1-0.5), where 1/2 is the value of fxx(x1,x2) specified in PDF function for a range.
Integrate x+y between 0 and 1:
f(x) = Integral(1,0) (x+y) dy = xy + (y^2)/2 = x + 1/2 (I get this by plugging 1 and 0 into the previous equation for y and subtracting results)
f(y) = Integral(1,0) (x+y) dx = (x^2)/2 + xy = 1/2 + y
From f(x), if I continue to integrate by dx on x + 0.5,
((x^2)/2 + x/2) is the result. I plug in 1 and 0. Subtracting x as 1 - x as 0 equations for this and the result is equal to 1.
1 is indicative of the coverage of the full sample space.
How to test if random variables are independent for joint PDF
f(x,y) = f(x) * f(y) if independence is true.
f(x,y) = integral integral (equation) dy dx
f(x) = integral (equation) dy
f(y) = integral (equation) dx
Condition Distribution for Joint PMF Formula
Pxy(X|Y) = Pxy(X,Y) / Py(Y)
Use of Uniform distribution in creating random samples for distributions
I can take value between 0 and 1 from a uniform distribution and plug it into a quantile function (inverse) of any distribution type (ex Exponential) and the output of that will be a random sample from that desired distribution.
How to execute change of variables formula for discrete random variables (univariate)
Obtain random sample from original distribution (ex: P(X =1)=0.2, P(X=2)=0.5, P(X=3)=0.3)
Consider another distribution y=g(x) that is based on original distribution.
Ex: g(x) is evenness of variable X, g(0) is even and g(1) is odd.
g(1) = P(X=1) + P(X=3) = 0.5
How to execute change of variables formula for continuous random variables (univariate)
PDF fy(y) = fx(x) * (dx/dy g^-1(y))
Ex:
fx(x) is a uniform distribution over [0,2], fx(x) = 1/2 = y (P(X<=1)=0.5)
y=g(x)=X^2 for y in [0,4], transform that into it’s quartile, x = √y.
Take g^-1(y)=x and differentiate: dx/dy (√y) = 1 / (2√y)
PDF = fx(x) * (dx/dy g^-1(y)) = (1/2) * (1 / (2√y))
PDF = (1/2) * (1 / (2√y)) = 1 / (4√y) for y in [0,4]
How to find the PDF of X if I know PDF Y, Y=g(X) (related to X), and g is strictly monotonic (always increasing or always decreasing) and continuously differentiable.
fx(x) = fy(g(x)) * |g’(x)|
For discrete independent random variables X & Y, the probability mass function (PMF) of their sum Z=X+Y is calculated by summing over all possible values x in the domain of X, using the formula:
pz(z) = Sum (px(x) * py(z-x)) , where p is a PMF