Quantitative methods Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

cross sectional data

A

> many observations of variables (subset)
same time period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

time series data

A

> many observations
different time periods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

panel data

A

> different time periods
many observations for each time period
combo of cross sectional and panel data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

strong positive corr

A

steep positive line

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

most appropriate functional form of regression by inspecting the residuals

A

want residuals to be random

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

permissionless distributed ledger technology (DLT) networks

A

> No centralised place of authority exists
all users i(nodes) within the network have a matching copy of the blockchain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

DLT that could facilitate the ownership of physical assets

A

Tokenization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Tokenization

A

> representing ownership rights to physical assets e.g. real estate
creating a single digital record of ownership to verify ownership title and authenticity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

application of DLT management

A

> cryptocurrencies
tokenization
compliance
post-trade clearing
settlement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

type of asset manager making use of fintech in investment decision making

A

> quants
fundamental assets mngrs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

data processing methods

A
  1. capture
  2. curate
  3. storage
  4. search
  5. transfer
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

fintech

A

technological innovation in the design and delivery of financial services and products

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is fintech

A

> analysis of large databases (traditional , non-traditional data)
analytical tools (AI for complex non-linear relationships)
automated trading (algorithms - lower costs, anonymity, liquidity)
automated advice (robo-advisers - may not incorporate whole information in their recommendations)
financial record keeping (DLT)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Big data characteristics

A

volume
velocity (real-time)
variety (structured, semi-structured and unstructured data)
veracity (important for inference or prediction, credibility and reliability of various data sources)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

sources of big data

A

finanicla markets
businesses
governments
individuals
sensors
internet of things

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

main sources of alternative data

A

businesses
individuals
sensors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

types of machine learning

A
  • supervised learning (inputs and outputs labelled, local market performance)
  • unsupervised learning (no data labelled, grouping of firms into peer groups based on characteristics)
  • deep learning (multi stage non linear data to identify patterns, supervised + unsupervised ML approaches)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Determinants of Interest Rates

A

r = Real risk-free interest rate + Inflation premium + Default risk premium + Liquidity premium + Maturity premium.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

1 + nominal risk-free rate

A

(1 + real risk-free rate)(1 + inflation premium)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

increased sensitivity of the market value of debt to a change in market interest rates as maturity is extended

A

maturity premium

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

defined benefit pension plans and retirement annuities

A

over the life of a beneficiary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

MWRR & TWRR

A

1) cash flows where inflows = outflows
2) HPR : (change in value of share + dividend)/initial value
annualised compounding rate of growth

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

r annual

A

(1+r weekly)^52 -1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

gross return

A

excl : mngmnt , taxes , custodial fees
incl : trading expenses

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

net return large vs small fund

A

small fund at disadvantage due to fixed administration costs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

return on leverage portfolio

A

R_p + (V_d/V_e)(R_p - r_d)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

cash flows associated with fixed income

A

> discount e.g. zero coupon bond (FV-PV)
periodic interest e.g. bonds w coupons
level payments : pay price + pay cash flows at intervals both interest and principal ( amortizing loans)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

ordinary annuity

A

r(PV) / (1-(1+r)^(-t))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

forward P/E

A

payout / (r-g)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

trailing P/E

A

(p*(1+g))/(r-g)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

(1+spot rate) ^n

A

(1+spot rate) ^(n-i) * (1+ forward)^(n-i)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

IRP

A

> spot FX * IR = forward FX
continuous compounding

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

percentile

A

(n+1)*(y/100)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

mean absolute deviation

A

> dispersion
(sum abs(x-xavg))/n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

sample target semi-deviation formula

A

((SUM_(x<=B)(X-B)^2)/(n-1))^(1/2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

coefficient of variation

A

sample st dev / sample mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

skewness

A

positive:
> small losses and likely
> profits large and unlikely
> invesotrs prefer distribution with large freq of unuasally large payoffs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

kurtosis

A

observations/ distribution in its tails than normal distrib
> platykurtic (thin tails, flat peak)
> mesokurtic (normal distr)
> leptokurtic (fat tails, tall peak)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

high kurtosis

A

higher chance of extrmees in tails

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

> platykurtic (thin tails, flat peak)
mesokurtic (normal distr)
leptokurtic (fat tails, tall peak)

A
  1. kurotsis < 3 , excess kurotsis -ve
  2. kurtosis = 3, excess kurtosis 0
  3. kurotsis > 3, excess kurotsis +ve
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

spurious correl

A

> chance rel
mix of two variables divided by third induce correl
rel of two var between third have correl

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

updated probability

A

(prob of new info given event / unconditional prob of new info) * prior prob of event

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

p(event|info)

A

[P(info|event)/P(info)]*P(event)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

P(F|E)

A

P(F)P(E|F)/[P(F)P(E|F)+P(Fnot)*P(E|Fnot)]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

odds for event

A

P(E)/[(1-P(E)]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

odds against event

A

[(1-P(E)]/P(E)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

Empirical

A

> Probability - relative frequency
historical data
Does not vary from person to person
objective probabilities

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

A priori

A

> Probability - logical analysis or reasoning
Does not vary from person to person
Objective probabilities

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

Subjective

A

> Probability - personal or subjective judgment
No particular reference to historical data
used in investment decisions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

A&B mutually exclusive and exhaustive events

A

P(C) = P(CA)+P(CB)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

P(B or C) (non-mutually exclusive events)

A

P(B or C) = P(B) + P(C) – P(B and C)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

P(B C)Dependent events

A

P(B C) = P(B) x P(C| B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

P(C) unconditional probability

A

P(C) = P(B) x P(C given B) + P(Bnot) x P(C given Bnot) = P(C and B) + P(C and Bnot)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

No. of ways the k tasks can be done

A

= ( n1)( n2 )( )….(nk )

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

Combination (binomial) formula

A

seq does not matter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

cov

A

P * (r-E(r_a))(r-E(r_b))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

shortfall risk

A

return below min level

(E(R_p)- R_l) / sigma_p

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

Roy’s safety-first criterion

A
  • Optimal portfolio: minimizes the probability that portfolio returns fall below a specified level
  • If returns are normally distributed, optimal portfolio maximizes safety-first ratio
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

Measuring and controlling financial risk

A
  • Stress testing and scenario analysis
  • Value-at-Risk (VaR) - value of losses expected over a specified time period at a given level of probability
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

Bootstrapping

A

> no knowledge of population
sample of size n
Unlike CLT that considers all samples of size n from the population - samples of size n from the known sample that also has size n
Each data item in our known sample can appear once or more or not at all in each
resample (due to replacement)
computer simulation to mimic the process of CLT : randomly drawn sample as if population
Easy to perform but only provides statistical estimates not exact results

61
Q

Resampling

A

repeatedly draws samples from one observed sample to make statistical inferences about
population parameters.

62
Q

Monte Carlo Simulation

A

> large number of random samples : represent the role of risk in the system

> specified probability distribution

e.g. pension assets with reference to pension liabilities

> Produces a frequency distribution for changes in portfolio value

> Tool for valuing complex securities

63
Q

Limitations of Monte Carlo simulation

A
  • Complement to analytical methods
  • Only provides statistical estimates, not exact results
  • Analytical methods provide more insight to cause-and-effect relationships
64
Q

Historical simulation

A
  • Sample from a historical record of returns or other underlying variables
  • Underlying rationale is that the historic record provides the best evidence
    of distributions
  • Limited by the actual events in the historic record used
  • Does not lend itself to ‘what if’ analysis like Monte Carlo simulation
65
Q

sampling error

A

diff be/een statistic and estimated parameter

66
Q

Stratified random sampling

A
  • divided into strata
  • simple random samples taken from each
    e.g. bond indices
  • Guarantees population subdivisions are represented
67
Q

Cluster sampling

A
  • divided into clusters – mini-representation of the entire population
    -certain clusters chosen as a whole using simple random sampling
  • If all members in each sample cluster are sampled: one-stage cluster sampling
  • If a subsample is randomly selected from each selected cluster : twostage cluster sampling
  • time-efficient and cost-efficient but the cluster might be less representative of the population
68
Q

Convenience sampling

A

Might be used for a pilot study before testing a large-scale and more representative
sample

69
Q

Judgmental sampling

A

Sample could be affected by the bias of the researcher

70
Q

Properties of Central Limit Theorem

A
  • Assuming any type of distribution and a large sample
  • Distribution of sample mean is approximately normal
  • Mean of the distribution of sample mean will be equal to population mean
  • Variance of distribution of sample mean equals population variance divided by the
    sample size
71
Q

Jackknife

A

> no knowledge of what the population looks like
sample of size n which is assumed to be a good representation of the population
unlike bootstrapping items are not replaced
bootstrapping we have B resamples but with jackknife we have n resamples such that resample sizes are n, n-1, n-2, n-3,……, 3, 2, 1
For a sample of size n, jackknife resampling usually requires n repetitions. In contrast, with bootstrap resampling, we are left to determine how many repetitions are appropriate
used to reduce the bias of an estimator and to find the standard error and confidence interval of an estimator

72
Q

Bootstrapping and Jackknife

A
  • Jackknife tends to produce similar results for each run whereas bootstrapping usually gives different results because resamples are drawn randomly
  • Both can be used to find the standard error or construct confidence intervals for
    the statistic of other population parameters
    > such as the median which could not be done using the Central Limit Theorem.
73
Q

Bernoulli and Binomial properties

A

mean : p , var: p(1-p)
mean : np , var: np(1-p)

74
Q

Discrete and continuous uniform distribution (random # for Monte Carlo sim)

A

f(x) = 1/#X
f(x) = #/(b-a)

75
Q

multivariate distribution pairwise corr

A

> n*(n-1)/2
feature for the multivariate normal distr

76
Q

99%, 95%, 68%, 90%

A

+-2.58
+-1.96
+-1
+- 1.65

77
Q

t-distr

A

n-1 df
as t large n>30 approaches normal distri
> fatter tails and less peak to normal curve

78
Q

students t and chi squared distr

A

> asymmetrical and bounded below by 0
family of dsitributions
chi square (1)
F(2) numeration and denominator df
as n tends to infty the probability density functions becomes more bell curved

79
Q

properties of an estimator

A

unbiased - sample mean = population mean
effcient - no other estimator has a sampling distribution with smaller variance
consistent - improves w sample size increase

80
Q

Point estimate is not likely to equal population parameter in any given sample

A

CI

81
Q

Confidence intervals

A

Point estimate +/- (Reliability factor (z_(a/2))x Standard error (sigma/(n)^(1/2))

82
Q

increase in reliability e.g. from 90% - 95%

A

wider CI

83
Q
  • If the population’s standard deviation is not known
A

t-stat (sigma >1)

84
Q

Normal distribution with a
known variance

A

sample < 30 - z-stat
sample > 30 - z-stat

85
Q

Normal distribution with
an unknown variance

A

sample < 30 - t-stat
sample > 30 - t-stat or z-stat

86
Q

Non-normal distribution
with a known variance

A

sample < 30 - N/A
sample > 30 - z-stat

87
Q

Non-normal distribution
with unknown variance

A

sample < 30 - N/A
sample > 30 - t-stat or z-stat

88
Q

What affects the width of the confidence interval

A
  • Choice of statistic (z or t)
  • Choice of degree of confidence
  • Choice of sample size
  • Larger sample size decreases width
  • Larger sample size reduces standard error
  • Big sample means t-calcs closer to z-calcs
  • Same for at least 30 observations
89
Q

Problems with
larger sample
size

A

cost
cross- poulation data

90
Q

Two-sided (or two-tailed) hypothesis test

A

Not equal to alternative hypothesis
* H0 : ϴ = ϴ0 versus Ha : ϴ ≠ ϴ0

91
Q

One-sided hypothesis test

A
  • A greater than alternative hypothesis
  • H0 : ϴ ≤ ϴ0 versus Ha : ϴ > ϴ0
  • A less than alternative hypothesis
  • H0 : ϴ ≥ ϴ0 versus Ha : ϴ < ϴ0
92
Q

t-stat z-score

A

(mean - estimated mean) / standard error

93
Q

2-tail or 1-tail significance level

A

subtract 0.3 from 2 tail for z-stat

94
Q

Type II error (β) + Type I error (α)

A

accept false null + reject true null

95
Q

Decrease in significance level (incr in confidence levels)

A

Reduces Type I error, but increases chances of Type II error

96
Q

Reduce both Type I and Type II errors

A
  • Increase sample size
97
Q

Power of a test

A
  • Probability of correctly rejecting H0 when it is false
  • 1-β
98
Q

Type I error

A

false discovery rate
BH number adjusted p − value = α*(Rank of i /Number of tests) — compare p -value w BH - reject null if p value less

99
Q

t-stat > critical value

A

rej H0

100
Q

Test the difference between two population means

A
  1. State the hypotheses
    - Null hypothesis is stated as
    H0: μd = 0
    - I.e. there is no difference in the populations’ mean daily returns (var unknowns but assumed equal)
  2. Identify the appropriate test statistic and its probability distribution
    - t-test statistic and t-distribution
  3. Specify the significance level
    - 5% significance level
  4. State the decision rule
    - If the test statistic > critical value, reject the null hypothesis
101
Q

Test of a single variance (if sample var known can test for population var)

A

chi-squared distributed with n-1 degrees of
freedom
two-tailed because distrib not symmetrical
chi^2_(n-1)=((n-1)s^2)/sigma^2_0

102
Q

Hypothesis Tests Concerning the Variance
Assumptions (chi square)

A
  • Normally distributed population
  • Random sample
  • Chi-square test is sensitive to violations of its assumptions
103
Q

Testing the equality of variances of two variances

A
  • Using sample variances to determine whether the population var are equal
  • F-distribution
  • Asymmetrical and bounded by zero
  • one-tailed
  • Calculation of F test statistic
    F = s^2/ s^2
    ≥ 1 as larger sample variance is numerator
    > df: n-1 / n-1
104
Q

Parametric tests

A
  • assumptions about the distribution of the population
  • E.g., z-test, t-test, chi-square test, or F-test
105
Q

Non-parametric tests are used in four situations

A
  1. Data does not meet distributional assumptions
    -not normally distributed + small sample
  2. OUTliers that affect a parametric statistic (the mean) but not a nonparametric statistic (the median)
  3. Data is given in ranks
  4. Characteristics being tested is not a population parameter
106
Q

Tests concerning a single mean

A

Parametric:
t-distributed test
z-distributed test

Non-Parametric:
Wilcoxon signed-rank
test

107
Q

Tests concerning
differences between
means

A

Parametric:
t-distributed test

Non-Parametric:
Mann-Whitney U test
(Wilcoxon rank sum test)

108
Q

Tests concerning mean differences (paired
comparison tests)

A

A paired comparisons test is appropriate to test the mean differences of two samples believed to be dependent.

Parametric:
t-distributed test

Non-Parametric:
Wilcoxon signed-rank test
Sign test

109
Q

Testing the significance of a correlation coefficient

A

both variables are distributed normally
parametric test
t tables (two-tailed p/2) and n-2 degrees of freedom:

t= r(n-2)^(1/2) / (1-r^2)^(1/2)

110
Q

As n increases we are more likely to reject a false NULL:Testing the significance of a correlation coefficient

A
  1. Degrees of freedom increases and critical statistic falls
  2. Numerator increases and test statistic rises
111
Q

The 3 Rank Correlation Coefficient

A

nonnormal distrbution
nonparemtric test
1. Rank observations on X from largest to smallest assigning 1 to the largest, 2 to the second, etc. Do the same for Y.
2. Calculate the difference, di, between the ranks for each pair of observations and square answer

=1 - (6*sum(d^2)/n(n^2-1))

sample size is large (n>30) we can conduct a t-test : df: n-2

= r((n-2)^(1/2))/(1-r^2)^(1/2)

112
Q

Ordinary Least Squares Regression

A

The estimated intercept, b0, and slope, b1, are such that the sum of the squared vertical distances from the observations to the fitted line is minimized.

113
Q

covariance

A

sum((x-xbar)(y-ybar))/(n-1)

114
Q

slope coefficient

A

covariance(x,y)/var(x)

115
Q

intercept

A

b0bar = Ybar - b1Xbar

116
Q

Assumptions of the Simple Linear Regression Model

A
  1. Linear relationship – might need transformation to make linear
  2. Independent variable is not random – assume expected values of independent
    variable are correct
  3. Variance of error term is same across all observations (homoskedasticity)
  4. Independence – The observations, pairs of Y’s and X’s, are independent of one another. error terms are uncorrelated (no serial correlation) across observations
  5. Error terms normally distributed
117
Q

SST =

A

SSR + SSE

118
Q

Total Variation

A

sum(y-ybar)^2

Sum of the squared differences between the actual value of the dependent variable and the mean value of the dependent variable

119
Q

Explained Variation

A

sum(yhat-ybar)^2

Sum of the squared differences between the predicted value of the dependent variable based on the regression line and the mean value of the dependent variable.

120
Q

Unexplained Variation

A

sum(y-yhat)^2

Sum of the squared differences between the actual value of the dependent variable and the predicted value of the dependent variable based on the regression line

121
Q

Coefficient of Determination – R2

A

> SSE =0 and RSS = TSS - perfect fit
percentage variation in the dependent variable explained by movements in the independent variable

R^2 = RSS / TSS or (1-(SSE/TSS))

r = sign of b1*(R^2)^(1/2)

122
Q

Regression : DF, SS, MS

A

k =1 indep var (measures the number of independent var)
sum(yhat-ybar)^2
MSR = SSR / DF

123
Q

Residual : DF, SS, MS

A

n-k-1
sum(y-yhat)^2
MSE = SSE/ DF

124
Q

standard error of the estimate (SEE)

A

MSE^(1/2)

SSE / n-2

125
Q

ANOVA

A

F-distributed Test Statistic
H0: b0=b1=…=0
F-test= (SSR / k) / SSE / (n-k-1) = MSR / MSE
> df = k , df = n-k-1

126
Q

Hypothesis Test of the Slope Coefficient

A
  • H0: b1 = 0
    tcalc = (bhat - b)/SE
    SE = (MSE)^(1/2) / (SUM(X-Xbar)^2)^(1/2)

or HO: b <= 0
or H0: b=1

127
Q

Hypothesis Test of the Intercept

A

H0: b0 = specified value
tcalc = (bhat0- b0) / SE
df= n-k-1
SE = SEE * (1/N + Xbar^2/sum(x-xbar)^2)^(1/2)

128
Q

Level of Significance and p-Values

A
  • Smallest level of significance at which the null hypothesis can be rejected
  • Smaller the p-value, stronger the evidence against the null hypothesis
  • The smaller the p-value, the smaller the chance of making a Type I error (rejecting the null when, in fact, it is true), but increases the chance of making a Type II error (failing to reject the null when, in fact, it is false)
129
Q

Prediction (confidence) intervals on the dependent variable

A

Y =Ŷf±tc*sf

s^2f= SEE^2[1+1/N+(Xf-Xbar)^2/(n-1)s^2x]

for y predicted need to plug value into linear equation

130
Q

Log-lin model

A

Slope coefficient represents the relative change in the dependent variable for an absolute change in the independent variable

131
Q

Lin-log model

A

Slope coefficient gives the absolute change in the dependent variable for a relative change in the independent variable

132
Q

Log-log (double-log) model

A

Slope coefficient gives the relative change in the dependent variable for a relative change in the independent variable and is useful for calculating elasticities

133
Q

hedged portfolio using
long underlying and short calls
to find c0

A

V0 = hS0 - c0
V1 +/- = hS1+/- -c1+/-
because we are hedged
V1+ = V1-
h (hedge ratio) = (c1+ - c1-) / (S1+ - S1-)
return = V1+ / V0 = V1- / V0 = 1+ R

hS0 - c0 = V1+ / (1+R)

134
Q

a parameter

A

refers to any descriptive measure of a population characteristic

135
Q

normal distribution z-score rejection points
Two-sided
One-sided

A

10%
1.645
1.28

5%
1.96
1.645

1%
2.58
2.33

136
Q

quintiles
1st
2nd
3rd
4th
5th

A

1st = 1/5
2nd = 2/5 etc

e.g. want 3rd quintile
4/5*(n+1) and n 10 then = 8.8
so answer be/teen postion 8 and 9
set numbers into ascending order
and interpolate e.g.
X8 + (8.8 − 8) × (X9 − X8)

137
Q

When working backward from the nodes on a binomial tree diagram, the analyst is most likely attempting to calculate:

A

In a tree diagram, a problem is worked backward to formulate an expected value as of today

138
Q

when the test statistic > critical statistic

A

reject H0
> the correl coefficient is statistically significant

remember : when two -tailed test the p-value / 2

139
Q

Ln(1+ discrete return)

A

= continuous return

140
Q

null hypothesis must always include the

A

equal sign

141
Q

a test of independence using a nonparametric test statistic that is chi-square distributed

A

χ2=∑^m_i=(Oij−Eij)^2/Eij

> m = the number of cells in the table, which is the number of groups in the first class multiplied by the number of groups in the second class;
Oij = the number of observations in each cell of row i and column j (i.e., observed frequency); and
Eij = the expected number of observations in each cell of row i and column j, assuming independence (i.e., expected frequency).
(r − 1)(c − 1) degrees of freedom, where r is the number of rows and c is the number of columns

Eij=(Total row i)×(Total column j)/ Overall total

Standardized residual=Oij−Eij/ √ Eij

142
Q

ML model that has been overfitted is not able

A

to accurately predict outcomes using a different dataset and might be too complex
‘overtrained’
> treating true parameters as if they are noise is most likely a result of underfitting the mode

143
Q

correlation coeff

A

(sign of b) sqrt (RSS)

144
Q

cash return on assets

A

= (Cash flow from operations/Average total assets)

145
Q

F-distributed test statistic to test whether the ? in a regression are equal to zero,

A

slopes
H0: b1 = 0. Ha: b1 ≠ 0

146
Q

Arithmetic mean x Harmonic mean =

A

Geometric mean^2

147
Q

Arithmetic mean

A

≥ Geometric mean ≥ Harmonic mean

148
Q
A