Final: Reasoning under Uncertainty Flashcards
(L23) What is a random variable, domain and its characteristics?
- Variable that an agent can be uncertain about its value
- dom(X): set of values that X can take
- Mutually exclusive and exhaustive
(L23) What is the probability distribution of a random variable?
P on a random variable X is a function dom(X) → [0, 1] such that for som value x → P(X = x)
(L23) Give examples of random variables?
(Discrete) Roll a dice,
dom = 1, …, 6
Probability distribution: even across all values of the dice
(Boolean) Raining outside
(L23) w╞ X=x
variable X = x in world w
(L23) What are the semantics of probability?
- µ(w): probability of world w
- \sum w \in W of µ(w) = 1
- P(f)= \ sum w⊧ f of µ(w) = 1 (sum the probabilities of the worlds where f is true)
(L23) Calculate the probability of proposition f given u(w) for the set of possible worlds
P(f)= \ sum w⊧ f of µ(w) = 1 (sum the probabilities of the worlds where f is true)
(L23) Define a joint probability distribution
- Their joint distribution is a probability distribution over the variables’ Cartesian product
- Sum of entries across the whole table is 1
(L24) Can you compute the probability distribution for each variable?
YES, by marginalization
(L24) Can you compute the probability distribution for any combination of variables?
YES, by marginalization we can compute the distribution over any subset of variables
(L24) Can you update these probabilities if you know something about the variables?
YES
(L24) DOWNSIDE of the joint probability distribution (JDC)
Size is exponential in the number of variables
(L24) What is normalization?
Rule out worlds and divide by p (new known information)
(L24) Prove the formula to compute conditional probability P(h | e) = P(h, e)/ P(e)
p.16
(L24) Derive the Product Rule, Chain Rule and Bayes’ Rule
p.17
(L24) Two ways to use conditional probability for inference
- Causal knowledge (from cause to evidence) P(evidence| cause hypothesis)
- Evidential reasoning (from evidence to cause) P(hypothesis| evidence)
(L25) Define Marginal Independence + interpretation
For all x_i in dom(X) and all y_k dom(Y), P(X= x_i | Y= y_k ) = P(X= x_i) –> P(X, Y) = P(X) * P(Y)
- New evidence Y (or X) does not affect current belief in X (or Y)
Joint probability Distribution (JPD) requiring 2^4 = 16 entries is reduced to two smaller ones (2^3 = 8 and 2).
(L25) Define Conditional Independence + interpretation
For all x_i \in dom(X), y_k \in dom(Y), z_m \in dom(Z),
P(X = x_i | Y = y_k, Z = z_m) = P(X = x_i | Z = z_m)
Value of Y doesn’t affect the belief in the value of X, given the value of Z.
(L25) Space complexity of P(X1, …, Xn), with n independent random variables
Be independency, P(X1, …, Xn) = P(X1) * … * P(Nn)
O(d^n) → O(n*d) space complexity:
(L25) Two equivalent statements:
Given catch is conditionally independent of toothache given cavity.
- P(tooth | catch, cavity) = P(tooth | cavity)
2. P(tooth, catch | cavity) = P(tooth | cavity) * P(catch | cavity)
(L25) How many values do we store in JPD, CPT?
Joint Probability Distributions (JPD): has O(d^n) values
Since all of them sum to 1, we store O(d^n) - 1 values
Conditional Probability Table (CPT): has O(d^n) values
Each row sums to 1, we store all but 1 per row.