Lecture 3 Flashcards
What are the building blocks of AI?
addition and multiplication
There are two sides to the “is AI intelligent” debate, what are these two sides?
It’s just math (so no) vs. look at the performance (so yes)
aka how ai works vs what it can do
What is the Xor problem?
Very simple models cannot learn the logical function of either this or that
What is the solution to the Xor problem?
more complicated neural networks
What is a neural network?
A neural network is really just stacked logistic regression (and intermediate steps)
Look at figure
Why do neural networks work better than logistic regression?
because they can do multiple things at once, intermediate steps
(i.e. use genre as a determinant as the example in the lecture aka consider other factors that have an effect)
What is the trade-off of using neural networks vs. logistic regression?
Neural networks need large amounts of data to work
The intermediate steps that are a part of neural networks, do they have to be put in?
No the model predicts these intermediate predictors themself, based on prior data
What is the universal approximation theorem?
Any function that looks like predictor <> outcome can be captured with neural networks
Does an AI like chatGPT work on this neural network model?
No, but it is one of the building blocks
How do language models work (simply)?
They predict the next word (dependent on the data they trained with)
Why do neural networks not work for language models?
They disregard the order of the words
What is the solution for neural networks not working for language models?
Transformer models
What are transformer models used for?
Language based models, but nowadays most AI have this as their basis (non-language included)
there are two sides to the transformer models, why?
The left side is to transform the input words into numbers and the right side is to put the numbers back into words