CFA Level 2 Flashcards - 1

You may prefer our related Brainscape-certified flashcards:
1
Q

Adjusted R2

A

Goodness-of-fit measure that adjusts the coefficient of determination, R2, for the number of independent variables in the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Akaike’s information criterion (AIC)

A

A statistic used to compare sets of independent variables for explaining a dependent variable. It is preferred for finding the model that is best suited for prediction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Analysis of variance (ANOVA)

A

The analysis that breaks the total variability of a dataset (such as observations on the dependent variable in a regression) into components representing different sources of variation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Breusch–Godfrey (BG) test

A

A test used to detect autocorrelated residuals up to a predesignated order of the lagged residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Breusch–Pagan (BP) test

A

A test for the presence of heteroskedasticity in a regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Coefficient of determination

A

The percentage of the variation of the dependent variable that is explained by the independent variables. Also referred to as the R-squared or R2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Conditional heteroskedasticity

A

A condition in which the variance of residuals of a regression are correlated with the value of the independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Cook’s distance

A

A metric for identifying influential data points. Also known as Cook’s D (Di).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Dummy variable

A

An independent variable that takes on a value of either 1 or 0, depending on a specified condition. Also known as an indicator variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Durbin–Watson (DW) test

A

A test for the presence of first-order serial correlation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

First-order serial correlation

A

The correlation of residuals with residuals adjacent in time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

General linear F-test

A

A test statistic used to assess the goodness of fit for an entire regression model, so it tests all independent variables in the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Heteroskedastic

A

When the variance of the residuals differs across observations in a regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

High-leverage point

A

An observation of an independent variable that has an extreme value and is potentially influential.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Influence plot

A

A visual that shows, for all observations, studentized residuals on the y-axis, leverage on the x-axis, and Cook’s D as circles whose size is proportional to the degree of influence of the given observation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Influential observation

A

An observation in a statistical analysis whose inclusion may significantly alter regression results.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Interaction term

A

A term that combines two or more variables and represents their joint influence on the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Intercept dummy

A

An indicator variable that allows a single regression model to estimate two lines of best fit, each with differing intercepts, depending on whether the dummy takes a value of 1 or 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Joint test of hypotheses

A

The test of hypotheses that specify values for two or more independent variables in the hypotheses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Leverage

A

A measure for identifying a potentially influential high-leverage point.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Likelihood ratio (LR) test

A

A method to assess the fit of logistic regression models and is based on the log-likelihood metric that describes the model’s fit to the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Log odds

A

The natural log of the odds of an event or characteristic happening. Also known as the logit function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Logistic regression (logit)

A

A regression in which the dependent variable uses a logistic transformation of the event probability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Logistic transformation

A

The log of the probability of an occurrence of an event or characteristic divided by the probability of the event or characteristic not occurring.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

The log of the probability of an occurrence of an event or characteristic divided by the probability of the event or characteristic not occurring.

A

A method that estimates values for the intercept and slope coefficients in a logistic regression that make the data in the regression sample most likely.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

Model specification

A

The set of independent variables included in a model and the model’s functional form.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Multicollinearity

A

When two or more independent variables are highly correlated with one another or are approximately linearly related.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Multiple linear regression

A

Modeling and estimation method that uses two or more independent variables to describe the variation of the dependent variable. Also referred to as multiple regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

Negative serial correlation

A

A situation in which residuals are negatively related to other residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

Nested models

A

Models in which one regression model has a subset of the independent variables of another regression model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

Normal Q-Q plot

A

A visual used to compare the distribution of the residuals from a regression to a theoretical normal distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

Omitted variable bias

A

Bias resulting from the omission of an important independent variable from a regression model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Outlier

A

An observation that has an extreme value of the dependent variable and is potentially influential.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Overfitting

A

Situation in which the model has too many independent variables relative to the number of observations in the sample, such that the coefficients on the independent variables represent noise rather than relationships with the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

Partial regression coefficient

A

Coefficient that describes the effect of a one-unit change in the independent variable on the dependent variable, holding all other independent variables constant. Also known as partial slope coefficient.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

Positive serial correlation

A

A situation in which residuals are positively related to other residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Qualitative dependent variable

A

A dependent variable that is discrete (binary). Also known as a categorical dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

Restricted model

A

A regression model with a subset of the complete set of independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Robust standard errors

A

Method for correcting residuals for conditional heteroskedasticity. Also known as heteroskedasticity-consistent standard errors or White-corrected standard errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

Scatterplot matrix

A

A visualization technique that shows the scatterplots between different sets of variables, often with the histogram for each variable on the diagonal. Also referred to as a pairs plot.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

Schwarz’s Bayesian information criterion (BIC or SBC)

A

A statistic used to compare sets of independent variables for explaining a dependent variable. It is preferred for finding the model with the best goodness of fit.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

Serial correlation

A

A condition found most often in time series in which residuals are correlated across observations. Also known as autocorrelation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

Serial-correlation consistent standard errors

A

Method for correcting serial correlation. Also known as serial correlation and heteroskedasticity adjusted standard errors, Newey–West standard errors, and robust standard errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

Slope dummy

A

An indicator variable that allows a single regression model to estimate two lines of best fit, each with differing slopes, depending on whether the dummy takes a value of 1 or 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

Studentized residual

A

A t-distributed statistic that is used to detect outliers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

Unconditional heteroskedasticity

A

When heteroskedasticity of the error variance is not correlated with the regression’s independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

Unrestricted model

A

A regression model with the complete set of independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

Variance inflation factor (VIF)

A

A statistic that quantifies the degree of multicollinearity in a model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

Autocorrelations

A

The correlations of a time series with its own past values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

Autoregressive model (AR)

A

A time series regressed on its own past values in which the independent variable is a lagged value of the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

Chain rule of forecasting

A

A forecasting process in which the next period’s value as predicted by the forecasting equation is substituted into the right-hand side of the equation to give a predicted value two periods ahead.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

Cointegrated

A

Describes two time series that have a long-term financial or economic relationship such that they do not diverge from each other without bound in the long run.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

Covariance stationary

A

Describes a time series when its expected value and variance are constant and finite in all periods and when its covariance with itself for a fixed number of periods in the past or future is constant and finite in all periods.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

Error autocorrelations

A

The autocorrelations of the error term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

First-differencing

A

A transformation that subtracts the value of the time series in period t – 1 from its value in period t.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

Heteroskedasticity

A

The property of having a nonconstant variance; refers to an error term with the property that its variance differs across observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

Homoskedasticity

A

The property of having a constant variance; refers to an error term that is constant across observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

In-sample forecast errors

A

The residuals from a fitted time-series model within the sample period used to fit the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

kth-order autocorrelation

A

The correlation between observations in a time series separated by k periods.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

Linear trend

A

A trend in which the dependent variable changes at a constant rate with time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

Log-linear model

A

With reference to time-series models, a model in which the growth rate of the time series as a function of time is constant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

Mean reversion

A

The tendency of a time series to fall when its level is above its mean and rise when its level is below its mean; a mean-reverting time series tends to return to its long-term mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

n-Period moving average

A

The average of the current and immediately prior n – 1 values of a time series.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

Out-of-sample forecast errors

A

The differences between actual and predicted values of time series outside the sample period used to fit the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

Random walk

A

A time series in which the value of the series in one period is the value of the series in the previous period plus an unpredictable random error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

Regime

A

With reference to a time series, the underlying model generating the times series.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

Residual autocorrelations

A

The sample autocorrelations of the residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

Root mean squared error (RMSE)

A

The square root of the average squared forecast error; used to compare the out-of-sample forecasting performance of forecasting models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

Seasonality

A

A characteristic of a time series in which the data experience regular and predictable periodic changes; for example, fan sales are highest during the summer months.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

Time series

A

A set of observations on a variable’s outcomes in different time periods.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

Trend

A

A long-term pattern of movement in a particular direction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
72
Q

Unit root

A

A time series that is not covariance stationary is said to have a unit root.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
73
Q

Activation function

A

A functional part of a neural network’s node that transforms the total net input received into the final output of the node. The activation function operates like a light dimmer switch that decreases or increases the strength of the input.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
74
Q

Agglomerative clustering

A

A bottom-up hierarchical clustering method that begins with each observation being treated as its own cluster. The algorithm finds the two closest clusters, based on some measure of distance (similarity), and combines them into one new larger cluster. This process is repeated iteratively until all observations are clumped into a single large cluster.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
75
Q

Backward propagation

A

The process of adjusting weights in a neural network, to reduce total error of the network, by moving backward through the network’s layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
76
Q

Base error

A

Model error due to randomness in the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
77
Q

Bias error

A

Describes the degree to which a model fits the training data. Algorithms with erroneous assumptions produce high bias error with poor approximation, causing underfitting and high in-sample error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
78
Q

Bootstrap aggregating (or bagging)

A

A technique whereby the original training dataset is used to generate n new training datasets or bags of data. Each new bag of data is generated by random sampling with replacement from the initial training set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
79
Q

Centroid

A

The center of a cluster formed using the k-means clustering algorithm.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
80
Q

Classification and regression tree

A

A supervised machine learning technique that can be applied to predict either a categorical target variable, producing a classification tree, or a continuous target variable, producing a regression tree. CART is commonly applied to binary classification or regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
81
Q

Cluster

A

A subset of observations from a dataset such that all the observations within the same cluster are deemed “similar.”

82
Q

Clustering

A

The sorting of observations into groups (clusters) such that observations in the same cluster are more similar to each other than they are to observations in other clusters.

83
Q

Complexity

A

A term referring to the number of features, parameters, or branches in a model and to whether the model is linear or non-linear (non-linear is more complex).

84
Q

Composite variable

A

A variable that combines two or more variables that are statistically strongly related to each other.

85
Q

Cross-validation

A

A technique for estimating out-of-sample error directly by determining the error in validation samples.

86
Q

Deep learning

A

Machine learning using neural networks with many hidden layers.

87
Q

Deep neural networks

A

Neural networks with many hidden layers—at least 2 but potentially more than 20—that have proven successful across a wide range of artificial intelligence applications.

88
Q

Dendrogram

A

A type of tree diagram used for visualizing a hierarchical cluster analysis; it highlights the hierarchical relationships among the clusters.

89
Q

Dimension reduction

A

A set of techniques for reducing the number of features in a dataset while retaining variation across observations to preserve the information contained in that variation.

90
Q

Divisive clustering

A

A top-down hierarchical clustering method that starts with all observations belonging to a single large cluster. The observations are then divided into two clusters based on some measure of distance (similarity). The algorithm then progressively partitions the intermediate clusters into smaller ones until each cluster contains only one observation.

91
Q

Eigenvalue

A

A measure that gives the proportion of total variance in the initial dataset that is explained by each eigenvector.

92
Q

Eigenvector

A

A vector that defines new mutually uncorrelated composite variables that are linear combinations of the original features.

93
Q

Ensemble learning

A

A technique of combining the predictions from a collection of models to achieve a more accurate prediction.

94
Q

Ensemble method

A

The method of combining multiple learning algorithms, as in ensemble learning.

95
Q

Features

A

The independent variables (X’s) in a labeled dataset.

96
Q

Fitting curve

A

A curve which shows in- and out-of-sample error rates (E in and E out) on the y-axis plotted against model complexity on the x-axis.

97
Q

Forward propagation

A

The process of adjusting weights in a neural network, to reduce total error of the network, by moving forward through the network’s layers.

98
Q

Generalize

A

When a model retains its explanatory power when predicting out-of-sample (i.e., using new data).

99
Q

Hierarchical clustering

A

An iterative unsupervised learning procedure used for building a hierarchy of clusters.

100
Q

Holdout samples

A

Data samples that are not used to train a model.

101
Q

Hyperparameter

A

A parameter whose value must be set by the researcher before learning begins.

102
Q

Information gain

A

A metric which quantifies the amount of information that the feature holds about the response. Information gain can be regarded as a form of non-linear correlation between Y and X.

103
Q

K-fold cross-validation

A

A technique in which data (excluding test sample and fresh data) are shuffled randomly and then are divided into k equal sub-samples, with k – 1 samples used as training samples and one sample, the kth, used as a validation sample.

104
Q

K-means

A

A clustering algorithm that repeatedly partitions observations into a fixed number, k, of non-overlapping clusters.

105
Q

K-nearest neighbor

A

A supervised learning technique that classifies a new observation by finding similarities (“nearness”) between this new observation and the existing data.

106
Q

LASSO

A

Least absolute shrinkage and selection operator is a type of penalized regression which involves minimizing the sum of the absolute values of the regression coefficients. LASSO can also be used for regularization in neural networks.

107
Q

Labeled dataset

A

A dataset that contains matched sets of observed inputs or features (X’s) and the associated output or target (Y).

108
Q

Learning curve

A

A curve that plots the accuracy rate (= 1 – error rate) in the validation or test samples (i.e., out-of-sample) against the amount of data in the training sample, which is thus useful for describing under- and overfitting as a function of bias and variance errors.

109
Q

Learning rate

A

A parameter that affects the magnitude of adjustments in the weights in a neural network.

110
Q

Linear classifier

A

A binary classifier that makes its classification decision based on a linear combination of the features of each data point.

111
Q

Majority-vote classifier

A

A classifier that assigns to a new data point the predicted label with the most votes (i.e., occurrences).

112
Q

Neural networks

A

Computer programs based on how our own brains learn and process information.

113
Q

Penalized regression

A

A regression that includes a constraint such that the regression coefficients are chosen to minimize the sum of squared residuals plus a penalty term that increases in size with the number of included features.

114
Q

Principal components analysis (PCA)

A

An unsupervised ML technique used to transform highly correlated features of data into a few main, uncorrelated composite variables.

115
Q

Projection error

A

The vertical (perpendicular) distance between a data point and a given principal component.

116
Q

Pruning

A

A regularization technique used in CART to reduce the size of the classification or regression tree—by pruning, or removing, sections of the tree that provide little classifying power.

117
Q

Random forest classifier

A

A collection of a large number of decision trees trained via a bagging method.

118
Q

Regularization

A

A term that describes methods for reducing statistical variability in high-dimensional data estimation problems.

119
Q

Reinforcement learning

A

Machine learning in which a computer learns from interacting with itself or data generated by the same algorithm.

120
Q

Scree plots

A

A plot that shows the proportion of total variance in the data explained by each principal component.

121
Q

Soft margin classification

A

An adaptation in the support vector machine algorithm which adds a penalty to the objective function for observations in the training set that are misclassified.

122
Q

Summation operator

A

A functional part of a neural network’s node that multiplies each input value received by a weight and sums the weighted values to form the total net input, which is then passed to the activation function.

123
Q

Supervised learning

A

A machine learning approach that makes use of labeled training data.

124
Q

Support vector machine

A

A linear classifier that determines the hyperplane that optimally separates the observations into two sets of data points.

125
Q

A linear classifier that determines the hyperplane that optimally separates the observations into two sets of data points.

A

In machine learning, the dependent variable (Y) in a labeled dataset; the company in a merger or acquisition that is being acquired.

126
Q

Test sample

A

A data sample that is used to test a model’s ability to predict well on new data.

127
Q

Training sample

A

A data sample that is used to train a model.

128
Q

Unsupervised learning

A

A machine learning approach that does not make use of labeled training data.

129
Q

Validation sample

A

A data sample that is used to validate and tune a model.

130
Q

Variance error

A

Describes how much a model’s results change in response to new data from validation and test samples. Unstable models pick up noise and produce high variance error, causing overfitting and high out-of-sample error.

131
Q

Accuracy

A

The percentage of correctly predicted classes out of total predictions. It is an overall performance metric in classification problems.

132
Q

Application programming interface (API)

A

A set of well-defined methods of communication between various software components and typically used for accessing external data.

133
Q

Bag-of-words (BOW)

A

A collection of a distinct set of tokens from all the texts in a sample dataset. BOW does not capture the position or sequence of words present in the text.

134
Q

Ceiling analysis

A

A systematic process of evaluating different components in the pipeline of model building. It helps to understand what part of the pipeline can potentially improve in performance by further tuning.

135
Q

Collection frequency (CF)

A

The number of times a given word appears in the whole corpus (i.e., collection of sentences) divided by the total number of words in the corpus.

136
Q

Confusion matrix

A

A grid used for error analysis in classification problems, it presents values for four evaluation metrics including true positive (TP), false positive (FP), true negative (TN), and false negative (FN).

137
Q

Corpus

A

A collection of text data in any form, including list, matrix, or data table forms.

138
Q

Data preparation (cleansing)

A

The process of examining, identifying, and mitigating (i.e., cleansing) errors in raw data.

139
Q

Data wrangling (preprocessing)

A

This task performs transformations and critical processing steps on cleansed data to make the data ready for ML model training (i.e., preprocessing), and includes dealing with outliers, extracting useful variables from existing data points, and scaling the data.

140
Q

Document frequency (DF)

A

The number of documents (texts) that contain a particular token divided by the total number of documents. It is the simplest feature selection method and often performs well when many thousands of tokens are present.

141
Q

Document term matrix (DTM)

A

A matrix where each row belongs to a document (or text file), and each column represents a token (or term). The number of rows is equal to the number of documents (or text files) in a sample text dataset. The number of columns is equal to the number of tokens from the BOW built using all the documents in the sample dataset. The cells typically contain the counts of the number of times a token is present in each document.

142
Q

Exploratory data analysis (EDA)

A

The preliminary step in data exploration, where graphs, charts, and other visualizations (heat maps and word clouds) as well as quantitative methods (descriptive statistics and central tendency measures) are used to observe and summarize data.

143
Q

F1 score

A

The harmonic mean of precision and recall. F1 score is a more appropriate overall performance metric (than accuracy) when there is unequal class distribution in the dataset and it is necessary to measure the equilibrium of precision and recall.

144
Q

Feature engineering

A

A process of creating new features by changing or transforming existing features.

145
Q

Feature selection

A

A process whereby only pertinent features from the dataset are selected for model training. Selecting fewer features decreases model complexity and training time.

146
Q

Frequency analysis

A

The process of quantifying how important tokens are in a sentence and in the corpus as a whole. It helps in filtering unnecessary tokens (or features).

147
Q

Grid search

A

A method of systematically training a model by using various combinations of hyperparameter values, cross validating each model, and determining which combination of hyperparameter values ensures the best model performance.

148
Q

Ground truth

A

The known outcome (i.e., target variable) of each observation in a labelled dataset.

149
Q

Metadata

A

Data that describes and gives information about other data.

150
Q

Mutual information

A

Measures how much information is contributed by a token to a class of texts. MI will be 0 if the token’s distribution in all text classes is the same. MI approaches 1 as the token in any one class tends to occur more often in only that particular class of text.

151
Q

N-grams

A

A representation of word sequences. The length of a sequence varies from 1 to n. When one word is used, it is a unigram; a two-word sequence is a bigram; and a 3-word sequence is a trigram; and so on.

152
Q

Name entity recognition

A

An algorithm that analyzes individual tokens and their surrounding semantics while referring to its dictionary to tag an object class to the token.

153
Q

One hot encoding

A

The process by which categorical variables are converted into binary form (0 or 1) for machine reading. It is one of the most common methods for handling categorical features in text data.

154
Q

Parts of speech

A

An algorithm that uses language structure and dictionaries to tag every token in the text with a corresponding part of speech (i.e., noun, verb, adjective, proper noun, etc.).

155
Q

Precision

A

In error analysis for classification problems it is ratio of correctly predicted positive classes to all predicted positive classes. Precision is useful in situations where the cost of false positives (FP), or Type I error, is high.

156
Q

Readme files

A

Text files provided with raw data that contain information related to a data file. They are useful for understanding the data and how they can be interpreted correctly.

157
Q

Recall

A

Also known as sensitivity, in error analysis for classification problems it is the ratio of correctly predicted positive classes to all actual positive classes. Recall is useful in situations where the cost of false negatives (FN), or Type II error, is high.

158
Q

Regular expression (regex)

A

A series of texts that contains characters in a particular order. Regex is used to search for patterns of interest in a given text.

159
Q

Scaling

A

The process of adjusting the range of a feature by shifting and changing the scale of the data. Two of the most common ways of scaling are normalization and standardization.

160
Q

Sentence length

A

The number of characters, including spaces, in a sentence.

161
Q

Term frequency (TF)

A

Ratio of the number of times a given token occurs in all the texts in the dataset to the total number of tokens in the dataset.

162
Q

Token

A

The equivalent of a word (or sometimes a character).

163
Q

Tokenization

A

The process of representing ownership rights to physical assets on a blockchain or distributed ledger.

164
Q

Trimming

A

Also called truncation, it is the process of removing extreme values and outliers from a dataset.

165
Q

Web spidering (scraping or crawling) programs

A

Programs that extract raw content from a source, typically web pages.

166
Q

Winsorization

A

The process of replacing extreme values and outliers in a dataset with the maximum (for large value outliers) and minimum (for small value outliers) values of data points that are not outliers.

167
Q

Absolute version of PPP

A

An extension of the law of one price whereby the prices of goods and services will not differ internationally once exchange rates are considered.

168
Q

Covered interest rate parity

A

The relationship among the spot exchange rate, the forward exchange rate, and the interest rates in two currencies that ensures that the return on a hedged (i.e., covered) foreign risk-free investment is the same as the return on a domestic risk-free investment. Also called interest rate parity.

169
Q

Ex ante version of PPP

A

The hypothesis that expected changes in the spot exchange rate are equal to expected differences in national inflation rates. An extension of relative purchasing power parity to expected future changes in the exchange rate.

170
Q

FX carry trade

A

An investment strategy that involves taking long positions in high-yield currencies and short positions in low-yield currencies.

171
Q

Forward rate parity

A

The proposition that the forward exchange rate is an unbiased predictor of the future spot exchange rate.

172
Q

International Fisher effect

A

The proposition that nominal interest rate differentials across currencies are determined by expected inflation differentials.

173
Q

Portfolio balance approach

A

A theory of exchange rate determination that emphasizes the portfolio investment decisions of global investors and the requirement that global investors willingly hold all outstanding securities denominated in each currency at prevailing prices and exchange rates.

174
Q

Real interest rate parity

A

The proposition that real interest rates will converge to the same level across different markets.

175
Q

Relative version of PPP

A

The hypothesis that changes in (nominal) exchange rates over time are equal to national inflation rate differentials.

176
Q

Triangular arbitrage

A

An arbitrage transaction involving three currencies that attempts to exploit inconsistencies among pairwise exchange rates.

177
Q

Uncovered interest rate parity

A

The proposition that the expected return on an uncovered (i.e., unhedged) foreign currency (risk-free) investment should equal the return on a comparable domestic currency investment.

178
Q

Absolute convergence

A

The idea that developing countries, regardless of their particular characteristics, will eventually catch up with the developed countries and match them in per capita output.

179
Q

Capital deepening

A

An increase in the capital-to-labor ratio.

180
Q

Club convergence

A

The idea that only rich and middle-income countries sharing a set of favorable attributes (i.e., are members of the “club”) will converge to the income level of the richest countries.

181
Q

Cobb–Douglas production function

A

A function of the form Y = Kα L1–α relating output (Y) to labor (L) and capital (K) inputs.

182
Q

Conditional convergence

A

The idea that convergence of per capita income is conditional on the countries having the same savings rate, population growth rate, and production function.

183
Q

Constant returns to scale

A

The condition that if all inputs into the production process are increased by a given percentage, then output rises by that same percentage.

184
Q

Diminishing marginal productivity

A

When each additional unit of an input, keeping the other inputs unchanged, increases output by a smaller increment.

185
Q

Dutch disease

A

A situation in which currency appreciation driven by strong export demand for resources makes other segments of the economy (particularly manufacturing) globally uncompetitive.

186
Q

Growth accounting equation

A

The production function written in the form of growth rates. For the basic Cobb–Douglas production function, it states that the growth rate of output equals the rate of technological change plus α multiplied by the growth rate of capital plus (1 – α) multiplied by the growth rate of labor.

187
Q

Labor force

A

Everyone of working age (ages 16 to 64) who either is employed or is available for work but not working.

188
Q

Labor force participation rate

A

The percentage of the working age population that is in the labor force.

189
Q

Labor productivity

A

The quantity of goods and services (real GDP) that a worker can produce in one hour of work.

190
Q

Labor productivity growth accounting equation

A

States that potential GDP growth equals the growth rate of the labor input plus the growth rate of labor productivity.

191
Q

Network externalities

A

The impact that users of a good, a service, or a technology have on other users of that product; it can be positive (e.g., a critical mass of users makes a product more useful) or negative (e.g., congestion makes the product less useful).

192
Q

Non-convergence trap

A

A situation in which a country remains relatively poor, or even falls further behind, because it fails to implement necessary institutional reforms and/or adopt leading technologies.

193
Q

Non-renewable resources

A

Finite resources that are depleted once they are consumed; oil and coal are examples.

194
Q

Potential GDP

A

The maximum amount of output an economy can sustainably produce without inducing an increase in the inflation rate. The output level that corresponds to full employment with consistent wage and price expectations.

195
Q

Purchasing power parity (PPP)

A

The idea that exchange rates move to equalize the purchasing power of different currencies.

196
Q

Renewable resources

A

Resources that can be replenished, such as a forest.

197
Q

Rental price of capital

A

The cost per unit of time to rent a unit of capital.

198
Q

Steady-state rate of growth

A

The constant growth rate of output (or output per capita) that can or will be sustained indefinitely once it is reached. Key ratios, such as the capital–output ratio, are constant on the steady-state growth path.

199
Q

Total factor productivity (TFP)

A

A multiplicative scale factor that reflects the general level of productivity or technology in the economy. Changes in total factor productivity generate proportional changes in output for any input combination.

200
Q

Administrative regulations or administrative law

A

Rules issued by government agencies or other regulators.