Test 2 Flashcards
(376 cards)
What is a paired t-test?
A paired t-test is a statistical test that compares the means of two related samples to determine if there is a significant difference between them.
Why use a paired t-test to compare classification models?
A paired t-test can be used to determine if there is a statistically significant difference in performance between two classification models trained and evaluated on the same data.
What data is needed to perform a paired t-test on two classification models?
To perform a paired t-test, you need the accuracy (or other performance metric) of each model on a set of test instances. The accuracies of the two models on each test instance form a pair.
How are the paired differences calculated?
For each test instance, calculate the difference between the accuracy of Model A and Model B. This gives you a set of paired differences.
What are the null and alternative hypotheses in this case?
Null hypothesis: The mean difference in accuracy between the two models is zero.
Alternative hypothesis: The mean difference in accuracy between the two models is not zero.
How do you draw a conclusion from the t-test results?
If the p-value of the t-test is less than the chosen significance level (e.g., 0.05), reject the null hypothesis and conclude there is a significant difference between the models’ performance. Otherwise, fail to reject the null hypothesis.
What does weighting instances mean?
Weighting instances means assigning different levels of importance or influence to each data point in a dataset during training of a machine learning model.
Why might you want to weight instances differently?
You might want to weight instances differently to:
Compensate for class imbalance
Emphasize certain instances that are more representative or important
Reduce the impact of noisy or less reliable instances
How can you simulate instance weights through data manipulation?
You can simulate instance weights by duplicating instances in the dataset. Instances with higher weights are duplicated more times than instances with lower weights.
What is the effect of duplicating instances?
Duplicating instances effectively increases their weight because the duplicated instances are seen more often during training. The model will be more influenced by the characteristics of the duplicated instances.
How do you determine the number of times to duplicate an instance?
The number of times an instance is duplicated should be proportional to its desired weight. For example, if Instance A has a weight of 2 and Instance B has a weight of 1, you would duplicate Instance A twice and Instance B once.
What are the potential drawbacks of simulating instance weights through duplication?
Duplicating instances increases the size of the dataset, which can increase training time and memory requirements.
Duplication may not be as precise as directly incorporating instance weights into the learning algorithm.
Some learning algorithms may have built-in mechanisms for handling instance weights, making duplication unnecessary.
What is instance weighting?
Instance weighting is the process of assigning different levels of importance or influence to individual instances (data points) in a dataset during the training of a machine learning model.
What is class imbalance?
Class imbalance refers to a situation where one class (or some classes) in a dataset has significantly fewer instances compared to the other class(es).
How can instance weighting help with class imbalance?
Instance weighting can help with class imbalance by assigning higher weights to instances from the minority class(es). This effectively increases their influence during training, helping the model to better learn the characteristics of the underrepresented class(es).
What are noisy instances?
Noisy instances are data points that contain errors, inconsistencies, or outliers that deviate significantly from the general pattern of the data.
How can instance weighting help with noisy instances?
Instance weighting can help with noisy instances by assigning lower weights to these instances. This reduces their influence during training, minimizing their impact on the learned model and potentially improving the model’s generalization performance.
What are some other reasons for weighting instances differently?
Domain knowledge: Experts may assign higher weights to instances that are known to be more representative or important based on their domain understanding.
Instance difficulty: Instances that are harder to classify or predict correctly may be assigned higher weights to encourage the model to focus more on these challenging cases.
Data collection bias: If certain instances are overrepresented due to data collection bias, they may be assigned lower weights to mitigate their disproportionate influence on the model.
What is a multilayer perceptron (MLP)?
A multilayer perceptron is a type of feedforward artificial neural network that consists of an input layer, one or more hidden layers, and an output layer. Each layer is composed of multiple interconnected nodes or neurons.
What are the key components of a node in an MLP?
Each node in an MLP has:
Input connections from nodes in the previous layer
An activation function that transforms the weighted sum of inputs
Output connections to nodes in the next layer
How are the inputs to a node weighted?
Each input to a node is multiplied by a corresponding weight value. These weights determine the strength and importance of the connections between nodes in adjacent layers.
What is the weighted sum of inputs for a node?
The weighted sum of inputs for a node is calculated by summing the products of each input value and its corresponding weight. This sum represents the total input signal to the node before applying the activation function.
What is the purpose of the activation function in a node?
The activation function in a node introduces non-linearity into the network, enabling it to learn and represent complex patterns. It transforms the weighted sum of inputs into an output signal that is passed to the next layer.
What are some common activation functions used in MLPs?
Some common activation functions used in MLPs include:
Sigmoid (logistic) function
Hyperbolic tangent (tanh) function
Rectified Linear Unit (ReLU) function