lecture 3: intro to probability and statistics Flashcards

1
Q

what is statistical causality/causation

A

one thing will directly cause the other, i.e. cause and effect
note: gold standard to determine causality is specific experimental design + randomised studies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is correlation

A

any statistical relationship whether causal or not, indicates a predictive relationship that can be exploited in practice

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is Simpson’s paradox

A

a phenomenon in which a trend appears in several groups of data but disappears/reverses when these groups are combined

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what is a random variable

A

its possible values are numerical outcomes of a random phenomenon, eg. roll dice

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what are the two types of random variables (RV)

A

discrete and continuous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is a discrete RV (DRV)

A

it takes on only a countable number of values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is the probability distribution of a DRV

A

probability mass function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is a continuous RV (CRV)

A

it takes on an infinite number of possible values in some interval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is the probability distribution of a CRV

A

probability density function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

what are the 3 most important statistics of a RV

A

expectation, variance and standard deviation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what is the main difference in formulas between DRV and CRV

A

for CRV, the formulas are integrals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Bayes’ rule

A
P(y⎮X) 
P(y) P(X⎮y)
= -------------------
P(X)
P(y) P(X⎮y)
= ------------------
∑y P(y) P(X⎮y)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

parameter estimation look at lecture 3 page 27 to 32

A

maximum likelihood estimation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what is parametric machine learning

A

a learning model that summarises data with a set of parameters of fixed size
involves 2 steps:
1. select a from for the function eg. normal distribution
2. learn the coefficients for the function from the training data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what is non-parametric learning

A

algorithms that do not make strong assumptions about the form of the mapping function, good when you don’t have a lot of data and no prior knowledge

How well did you know this?
1
Not at all
2
3
4
5
Perfectly