Neural network models Flashcards
What is a feedforward neural network?
data flows in one direction from input to output through hidden layers comprised of units/neurons each connected to everything in the previous layer via weighted connections
What is a perceptron?
single layer feedforward neural network: takes multiple outputs, computes weighted sum (multiply each input by weight), produces singular output
What is gradient decent?
optimisation algorithm to minimise error/loss, adjusts weights and other parameters in direction that leads to greatest decrease in the error, like decending down a hill by taking small steps in steepest direction
Explain parametric versus non-parametric models
parametric are a fixed size, fixed number of inputs and outputs, learn set of parameters to map inputs onto outputs. non-parametric models grow as number of inputs and outputs grow, encodes values like a table or big list
Explain linear versus non-linear problems/networks
linear problems need straight line to solve, non-linear need a curve. linear activation functions output directly proportional to input but non-linear doesn’t have to be. E.g. step function like light switch but sigmoid like volume knob
What is translation invariance?
ability to recognise objects regardless of spatial location, scale or orientation, achieved in visual system and CNNs by heirarchal processing, spatial pooling and feature detection
Why do regular neural networks lack translation invariance? Why are CNNs better?
FNNs are fully connected and lack weight sharing - only learn patterns independently at each spatial location. CNNs have local connectivity and weight sharing - convolutional layers where neuron connected to local region of input and filters applied across different locations
What is a CNN?
type of FNN specialised for processing visual data/images, distinguished by addition of convolutional hidden layers - local filters convolve across image matrix to prodice feature map, layers build upon features of last to extract more complex features in heirarchal structure
Explain dimensionality
number of features in input data, number of neurons in each layer. more dimensionality = more complex representations but also more computational costs. activation functions can increase by introduce non-linearity, pooling reduces
Explain adversial networks
networks trained to find images that CNNs classify incorrectly, incrementally adjusts image until it maximally resembles image of different class but without losing its class label
Why are CNNs a good model of the ventral but not the dorsal stream?
can classify objects like the “what” pathway but are usuallly for static images and can’t do motion like “where” pathway and don’t understand what objects are like “how” pathway.
Explain spatial and temporal heirarchies
how processing organised in brain- spatial goes from simple features like lines up to complex objects, temporal goes from shorter to longer temporal windows. CNNs have spatial but not temporal
Why are CNNs a good model for the visual system?
image processing, heirarchal feature extraction, local filters analogue for retinal receptive fields, translation invariance, multiclass classification from probability distrubution of classes, pooling to reduce dimensionality
What are the limitations of CNNs as a model for the visual system?
CNNs feedforward only but visual system has feedback loops from higher cognitve areas, still struggle with generalisation as limtied to training data, vulnerable to adverserial methods, only spatial not temporal processing, only ventral not dorsal
Explain delay preiod activity
activity in area of what trying to remember when trying to remember it - cells fire persistently in spatially selective fashion thought to be substrate for short-term integration/working memory. seen in dorsolateral prefrontal cortex of macaque