Uda Flashcards
Let’s … a bit.
sidetrack
he does not let himself get … by trends.
sidetracked
LDA assumes that the every chunk of text we feed into it will contain words that are somehow related. Therefore choosing the right corpus of data is …
crucial
Remove very rare and very … words.
common
Therefore, common words like “the” and “for”, which appear in many documents, will be .. down. Words that appear frequently in a single document will be scaled up.
scaled
A … but effective device from the 1840’s, rather like a thermometer. It works, but a human must estimate its readings.
primitive
… dependencies: dependencies over time
Temporal
You will notice that in these videos I use subscripts as well as … as a numeric notation for the weight matrix.
superscript
Even when knowing which annotation will get the most focus, it’s interesting to see how … softmax makes the end score become.
array:
[ 927., 397., 148., 929.]
softmax:
[ 0.11920292, 7.9471515e-232, 5.7661442e-340, 0.88079708]
drastic
You could skip the embedding step, and feed in the one-hot encoded vectors directly to the recurrent layer(s). This may reduce the complexity of the model and make it easier to train, but the quality of translation may suffer as one-hot encoded vectors cannot … similarities and differences between words.
exploit
If you would like to … your knowledge even more, go over the following tutorial.
deepen
Words that you would expect to see more often in positive reviews – like “amazing” – have a ratio greater than 1. The more … a word is toward positive, the farther from 1 its positive-to-negative ratio will be.
skewed
There are still a couple of problems to … out before we use the bigram probability dictionary to calculate the probabilities of new sentences
sort
This diet is … in vitamin B.
deficient
Some possible combinations may not exist in our probability dictionary but are still possible. We don’t want to multiply in a probability of 0 just because our original corpus was …
deficient