Extra Flashcards
Why do we want non-polynomial functions
because polynomial functions have multiple low or high points on the graph = multiple answers.
If we have an error of 500 in regression how do we correct that?
Gradient descent will use backpropergation to adjust the theta values to minimize the output by 500
Why do we square cost functions in regression
we want to avoid some data in the sum function to cancel out other values as they might be negative. (when adding to the sum function)
Is the XOR function linear or non-linear?
It is of course, non-linear it.
It returns a binary result, 1 or 0, dependent on whether
What is key and value in a dictionary in terms of Markov Chains?
In Markov Chains the key is the word you want to generate the following word for, and the value is a list of possible words that can follow.
What is a decision tree?
It is a flow-chart like model, where each node is a test on an attribute. e.g. left is yes and right is no.
It is used to make decisions.
Supervised learning as you deal with decisions that are predetermined and you know the cases.
What is the random forest algorithm
It is a classification and regression algorithm for learning. It is a supervised learning method.
What is K-nearest neighbour?
it is a classification algorithm that assumes that values with similar values are similar
When a Deep Neural Network model learn, what is it essentially the method of how it learns?
It minimizes which weight and biases minimize a certain cost function (sum of error)
What is backpropagation
Backpropagation is how we adjust our weights after we have identified how much error we have with the cost functions.
If your model is 100 % exact on your training set what does that imply?
It implies that something is wrong and that you need to look into why
What can you do with your model if your model is predicting too precisely to the training set?
Think in terms of nodes.
You can remove some layers, causing the DNN to have less of a chance of getting the exact weights.
What can you do if you want to make it harder for your model to predict too closely to the training set
You can add regularization which is basically taking the sum of the weights and dividing that with the cost function. Then whenever the model changes the weights to get a better prediction the change in the weights are not accounted for making it less precise.
Why dont you want your model to predict 100 % precise
You dont want it to not be able to predict outside of the training set. so outside with new values. It can have learned the errors as features also