Final Review Flashcards
casewrite()
writes a frozen video frame to a file.
caseread() –
reads a frozen video frame from a file. The file also contains the “supervisor’s” choice of a correct move.
readweights()
reads in the weights data from a file. This is the “brain.” There are two weights arrays formed. One is 3 dimensional and the second is 2 dimensional.
Writeweights()
records the current weights files (the “brain”) to a file.
applynet()
uses the input node values and the weights arrays to calculate the firing values at the hidden nodes and then at the output nodes.
backpropagate()
where the learning happens. Working backward from output values to hidden values to input values and the current weights, the network choice for an output node and the correct output node value from the supervisor, this function calculates and applies changes to the weights to reduce the overall network error.
steer()
send the correct commands to the robot motors to steer the robot and go forward
normalize()
adjusts input values to produce a consistent contrast to reduce the differences caused by different light levels
sigmoid() brief
our “threshold” function to make all hidden and output node firing values be between 0 and 1.
Capture functions
normalize() and casewrite()
capture
cam on, robot off; record case w/ steering move
Initialize functions
writeweights()
Initialize
builds starting wts file; creates brain that knows nothing
Train functions
ApplyNet(), Backpropogate(), readwts, writewts, caseread, sigmoit, normalize
train
read in current wts and training, use applynet to find firing vals at hidden and output nodes, use supervisor’s input and do backprop() to correct all wts
RobotRun functions
readwts(), normalize(), applynet(), sigmoid(), steer()
RobotRun
runs robot; neural network driving
how to apply steering move
after applynet finds all the hidden and then output node values, we choose that output node with the highest value. (-2, -1, 0, 1, 2)
tic tac toe recursion
minimax is double recursion with min and max calling each other;
supervised learning
learns by comp calc w/supervisors; hardest part is getting training cases
sigmoid funct and x
1/(1+e^-x);
sigmoid role
use instead of square wave for differentiability; is used to connect every layer to the next layer. For each, the system needs an array of real number weights.
The final layer is called the output layer. Each node represents an action for the agent. The agent chooses that action with the largest firing value.