Week 3: feedforward visual processing Flashcards
Why do we study vision networks to understand DCNN?
- image processing is a common application of artificial deep networks
- the early visual system is the best understood system in the human brain
- early visual responses are a common application of DCNN as simulations of neural processing
Where do responses to specific edge orientations emerge?
V1, the primary visual cortex. first area processing vision in the brain
How is the first step of visual processing different in biological networks and artificial networks?
In biological network, contrast is initially computed in an orientation-independent filter in the retinal ganglion cell. Artificial DCNNs often skip this and go directly to edge orientation.
What happens in V1?
Orientation-selective responses are computed in V1 by operations comparing the outputs of retinal ganglion cells
How are neurons processing the image grouped?
Neurons processing the same part of the visual field are grouped, and neurons with similar orientation preferences are grouped, this grouping is at a very small scale.
How do neurons have different orientation preferences?
Neurons have orientation preferences which gradually change across the cortex
How do orientation preferences form feature maps?
The orientation columns form further feature maps which are squeezed into the same 2D cortical surface
What are in the feature maps at V1?
In V1 there is a large complex set of feature maps with each feature represented at all spatial positions. Includes colour, eye, spatial frequency, orientation, motion direction.
What happens after V1 in terms of processing an image?
Form (object recognition) and motion (motion and space) information are processed separately in different areas of the brain. There are multiple branching hierarchies performing different tasks
How does the hierarchy of the brain differ to that of a neural network?
Lots of brain areas sample from V1 creating a web of connections. Artificial networks use a linear hierarchy
What happens as you go up the hierarchy V1-V4… in the brain?
As you go up, the areas have a larger representation of the central visual field, and respond to increasingly complex features
What happens to the spatial integration of the image as you move up the hierarchy?
Spatial relationships between image locations are maintained
How does V1 represent the different areas in the visual field?
V1 strongly over-represents the central visual field compared to the peripheral parts of the field
Which parts of the brain focus on object recognition and spatial perception?
Object recognition: ventral stream, temporal lobe
Spatial perception/action planning: dorsal stream, pariatal lobe
How do later visual field maps sample from earlier visual field maps?
Later visual field maps sample from approximately constant cortical areas of earlier visual field maps, regardless of the visual position represented