Week 3: feedforward visual processing Flashcards

1
Q

Why do we study vision networks to understand DCNN?

A
  • image processing is a common application of artificial deep networks
  • the early visual system is the best understood system in the human brain
  • early visual responses are a common application of DCNN as simulations of neural processing
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Where do responses to specific edge orientations emerge?

A

V1, the primary visual cortex. first area processing vision in the brain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How is the first step of visual processing different in biological networks and artificial networks?

A

In biological network, contrast is initially computed in an orientation-independent filter in the retinal ganglion cell. Artificial DCNNs often skip this and go directly to edge orientation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What happens in V1?

A

Orientation-selective responses are computed in V1 by operations comparing the outputs of retinal ganglion cells

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How are neurons processing the image grouped?

A

Neurons processing the same part of the visual field are grouped, and neurons with similar orientation preferences are grouped, this grouping is at a very small scale.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do neurons have different orientation preferences?

A

Neurons have orientation preferences which gradually change across the cortex

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How do orientation preferences form feature maps?

A

The orientation columns form further feature maps which are squeezed into the same 2D cortical surface

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are in the feature maps at V1?

A

In V1 there is a large complex set of feature maps with each feature represented at all spatial positions. Includes colour, eye, spatial frequency, orientation, motion direction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What happens after V1 in terms of processing an image?

A

Form (object recognition) and motion (motion and space) information are processed separately in different areas of the brain. There are multiple branching hierarchies performing different tasks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does the hierarchy of the brain differ to that of a neural network?

A

Lots of brain areas sample from V1 creating a web of connections. Artificial networks use a linear hierarchy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What happens as you go up the hierarchy V1-V4… in the brain?

A

As you go up, the areas have a larger representation of the central visual field, and respond to increasingly complex features

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What happens to the spatial integration of the image as you move up the hierarchy?

A

Spatial relationships between image locations are maintained

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How does V1 represent the different areas in the visual field?

A

V1 strongly over-represents the central visual field compared to the peripheral parts of the field

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Which parts of the brain focus on object recognition and spatial perception?

A

Object recognition: ventral stream, temporal lobe

Spatial perception/action planning: dorsal stream, pariatal lobe

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How do later visual field maps sample from earlier visual field maps?

A

Later visual field maps sample from approximately constant cortical areas of earlier visual field maps, regardless of the visual position represented

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does an artificial network filter represent a sample from the visual field map and how is it different?

A

Later visual field maps sample from constant cortical areas of earlier visual field maps. this is represented by the fixed size of artificial network filters
Differences:
-the inputs in biological networks over-represent central vision
-the input in biological network is neurons rather than images

17
Q

How do features transform as they move up through the layers of the hierarchy?

A

Transformations find commonly-seen patterns in activity of earlier layers. this has been difficult for humans to recognise

18
Q

What types of computations are later stages in the hierarchy doing?

A

Later stages are likely doing the same computations as earlier stages, but from more abstracted inputs

19
Q

Why do we use DCNNs to simulate feature transformations?

A

It is hard for humans to think about the transformations and representations of later layers in the network. DCNNs are useful for experimenting and testing hypotheses on how these transformations work

20
Q

How do mid-level representations seem to be optimised?

A

Mid-level representation appears to be optimised to allow subsequent transformation to support object recognition

21
Q

What is distributed encoding?

A

Object identity is not reflected in the activity of a single neuron but the pattern of activity in a larger population of neurons

22
Q

What are the advantages of distributed encoding?

A
  • allows some cell death without representation failing (graceful degradation)
  • allows new patterns to be stored without new cells, a fixed group of cells can store a variable number of objects
  • it is consistent with measured cell properties -ie there are rarely all or nothing responses
23
Q

What is the disadvantage of distributed encoding?

A

It is harder for humans to understand

24
Q

What are face-selective neurons?

A

Found later in the ventral stream, there are cells that respond more strongly to specific faces regardless of the image used

25
Q

How do artifical DCNNs relate to face-selective neurons?

A

Artificial DCNNs produce similar results to the brain as later network layers closely resemble responses of face-selective cells

26
Q

What are Deepfakes?

A

Videos where a DCNN is used to map ones person’s face onto another. first show the network training videos of the target face, then the network maps the features of the face and their movements onto the source face. artificial DCNNs can convincingly manipulate facial identity

27
Q

What are object-selective areas? what type of objects are commonly studied

A

There are brain areas that respond to many classes of objects
often studied: faces, places, words and tools

28
Q

How does the brain process semantic content?

A

Responses in the brain show that processing semantic content produces similar results to processing visual objects which suggests similar processes are involved

29
Q

How does the brain change its process when given a task to identify humans?

A

Recording sites throughout the brain change their object selectivity and start responding more to humans. A face-selective area could become a car-selective area if given a car identification task

30
Q

How does the brain process visual content in dreams?

A

Similarly to content seen when awake. The responses in the brain are similar when dreaming all the way back to the early image representation in V1

31
Q

How do responses change when we focus on something? object vs spatial responses?

A

Responses are drawn towards attended content, more neurons will respond to the attended area of the image

  • object responses are drawn towards attended object
  • spatial responses drawn towards attended locations