Chapter 1 Flashcards

1
Q

If there are p variables, how many pairwise scatterplots can be produced? How implications does this have?(1)

A

p*(p-1)/2
Means scatterplots arent best for large p.
Heat maps can be used if data is not categorical as a replacement with yellow and white indication higher values and reds and oranges for lower values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does covariance indicate?(1)

A

Positive indicates that one variables value larger (or smaller) from its mean generally means larger or smaller for the other variable. Conversely, negative covariance indicates larger values than mean will produce smaller in the other variable (and vice versa).
Note 1/n-1 in the formula gives an unbiased estimate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the sample mean vector of the data matrix X?(1)

A

1/nX^T1n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does pre or post multiplying by a vector of ones do?(1)

A

Has the effect of calculating the column sums or row sums respectively.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the identity mstrix?

A

Zeros everywhere 1 diagonals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the centering matrix (Hn)? What does this do?(2)

A

In-1/n1n^T1n.
Pre-multiplying X by the centering matrix Hn has the effect of subtracting the appropriate sample mean from each element. Therefore the centred data matrix has a sample mean vector of 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What lis the sample covariance matrix of the data matrix X?(1)

A

(1/n-1)X^THn*X
NEED TO LEARN PROOF!!
Note that is a symmetric and semi-positive definite-learn proof.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are 2 properties of the centering matrix?(2)

A

LEARN PROOFS
Symmetric Hn^T=Hn
and idempotent Hn^2=Hn (Idempotence means that multiple applications of a particular operation do not change the result. In other words, if we try to center the
centering matrix we are left with the centering matrix.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Calculated rij for correlation matrix.(1)

A

rij = sij/sisj

where si = √sii is the sample standard deviation of the ith variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does R = Ip mean?(1)

A

Variables are uncorrelated as correlation matrix shows variables only correlate with one another.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How would you calculate the sample correlation matrix R from sample covariance matrix S?(1)

A

D^-1SD^-1.

Where D is diagonal matrix with standard deviations of the variables.

Thud R is positive semi-definite from S also being psd.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are 2 single measures of multivariate scatter?(2)

A

Generalised variance=Det(S)

Total variation=Tr(S) (sum of diagonals of S)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is a linear functional?(1)

A

q=1 for transformation f(x)=a^Tx where a is a vector length p Rp–>R

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is an affine transformation?(1)

A

Linear transformation combined with a shift in location

Rp–>Rqbf(x)=Ax+b for A qxp b vector length b.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a 2d projection?(1)

A

A linear transformation of x y=AX where A=(ej^T,ek^T)^T and ej is the length-p vector with 1 in element j and 0 everywhere else
-A selection of a pair of variables for a scatterplot matrix plot.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a 2d rotation?(1)

A

Rotation of the point x=(x1,x2)^T anticlockwise through the angle theta
Linear transformation y=Ax, y=(y1,y2)^T,
A(costheta,-sintheta,
sintheta,costheta) ie 2x2 matrix

17
Q

What us the matrix formulation of affine transformations?What are the mean and covariance matrix?(1,2)

A

Y=XA^T + 1nb^T
ybar=Axbar+b, Sy=ASA^T
Learn proof for mean and cov!!
Note Sy is symmetric due to S being symmetric and also semi-positive definite

18
Q

What is the spectral (eigen)decomposition?(1)

A

L/\L^T
/=diag(lambda1,lambda2…lambdap) diagonal matrix of eigenvalues of A (lambda1>lambda2>…>lambdap)
L=matrix of normalised eigenvectors (i.e lj^Tlj=1 corresponding to lambdaj)

19
Q

What is the orthogonal matrix?(1)

A

LL^T=Ip=L^TL i.e. li^Tli=1 and li^Tlj=0 for i/=j

20
Q

How do we calculate the roots lambda?What about eigenvectors?(1)

A

|A-lambda*In|=0

Av=lambdav

21
Q

How to calculate matrix square root? Under what conditions does this apply? Can this be generalised?(2)

A

If A is a symmetric positive semi-definite matrix with spectral decomposition L/\L^T then the matrix square root of A is:
A^(1/2)=L/\^(1/2)L^T
Where /\^(1/2) is the diagonal matrix with square roots of lambda
Yes it can be for A^alpha replace sqrt with power of alpha in the above where alpha is any real number

Note: Cholesky usually cheaper in computation if only require solution to GG^T=A where A=LL^T L is a lower triangle matrix.

22
Q

What is the scaling (standardisation) transformation?(1)

A

Y=(X-1nxbar^T)*D^-1

This pulls all variables onto a common scale but maintains the correlation between them.

23
Q

What is the Mahalanobis transformation?What is its purpose?(1)

A
Aims to place variables on a common scale AND remove correlation between them.
y=S^(-1/2)
Where S(-1/2)=L/\^(-1/2)L^T
The transformation is a p-dimensional linear transformation yr=Axr+b with transformation matrix A=S^(-1/2) and b=-S(-1/2)xbar, each transformed variable has 0 mean, unit variance and uncorrelated variables.