exam 1 Flashcards
inattentional blindness
failing to see objects when attention is focused elsewhere (e.g. missing a gorilla in a video)
change blindness
failing to notice changes in a scene due to attention limitations
what are ensemble statistics in the nonselective pathways
the nonselective pathways extracts the overall scene properties (e.g. color, texture, layout)
what is attentional extinction (due to parietal damage)
inability to notice stimuli on the neglected side when competing stimuli are present
what is attentional neglect (due to parietal damage)
ignoring one side of space (often left) after right parietal damage
how do neurons implement attention
- enhancement
- sharper tuning
- altered tuning
enhancement (neurons implementing attention)
stronger response to attended stimuli
sharper tuning (neurons implementing attention)
more precise focus on relevant features
altered tuning (neurons implementing attention)
changes in preferred stimulus properties
how does attention affect neural activity?
- enhances firing rates of neurons responding to attended stimuli
- suppresses irrelevant stimuli
attentional blink
a gap in perception when detecting two targets in rapid succesion (the second target is often missed if it appears 200-500 ms after the first)
rapid serial visual presentation (RSVP) paradigm
stimuli appear in quick succesion at the same location (used to study attentional blink and perception limits)
feature integration theory (FIT)
features (color, shape) are processed separately and must be bound together by attention
the binding problem
how does the brain combine features into a unified perception?
what is guidance in visual search?
attention is guided by salient features and prior knowledge
feature search
- fast, parallel
- target “pops out”
conjunction search
- slower, serial
- target shares features with distractors
spatial configuration search
- even slower
- requires recognizing relationships
visual search paradigm
a task where participants find a target among distractors. The search efficiency is measured by reaction time vs. number of distractors
the “spotlight” metaphor of attention
attention is like a beam of light, enhancing what it illuminates. However, attention can split, shift, or diffuse, challenging the metaphor
differences between endogenous and exogenous attention cues
- endogenous –> voluntary, based on goals (e.g. looking for a friend in a crowd)
- exogenous –> involuntary, driven by sudden stimuli (e.g. flashing light)
what is the Posner cueing task
a test where a cue directs attention to a location before a target appears
- valid cues (correct location) –> faster responses
- invalid cues (incorrect location) –> slower responses
are different kinds of scenes processed differently
Yes, the brain categorizes scenes (e.g. urban vs. natural). Some areas, like the PPA (parahippocampal place area) specialize in scene recognition
how much do we actually notice or remember from what we see
- very little detail (most information is filtered out)
- we retain the gist, rather than specific details
- change blindness and inattentional blindness reveal our limited awareness
if we can only attend to a few things at once, why does the world seem rich and detailed?
- “gist perception” allows us to quickly understand scenes
- the brain fills in missing details based on experience
- rapid shifts of attention create an illusion of full perception
what changes in the brain when we “pay attention”?
- increased activity in visual and parietal cortex
- enhanced neural responses to attended stimuli
- suppression of irrelevant stimuli
how do we find what we are looking for?
- we use visual search and guidance from memory and expectations
- the brain prioritizes salient features (e.g. color, shape)
- top-down (goal-directed) and bottom-up (stimulus-driven) processed guide search
is attention really like spotlight?
Yes, in some ways (select specific locations, objects or features). But not exactly (attention can diffuse, tracking multiple things at once)
“zoom lens” model of attention
suggests attention can widen/narrow as needed
why can’t we process everything at once?
- the brain has limited cognitive resources and selects relevant information
- too much information would cause overload and slow down processing
- attention prioritizes what’s most important for survival or goals
visual problems that lead to abnormal development and stereoblindness
- strabismus (lazy eye)
- amblyopia
- congenital cataracts
strabismus
- lazy eye
- misalignment disrupts stereopsis
amblyopia
weaker eye suppressed in the brain
congenital cataracts
block visual input, preventing normal depth perception development
stages in normal development of stereopsis
- birth: no stereopsis
- 3-6 months: binocular coordination improves
- 4-6 months: stereopsis emerges
- childhood: refined depth perception
stereopsis
het vermogen van de hersenen om twee beelden van hetzelfde object, die elk door een oog worden waargenomen, te combineren tot één enkel 3D-beeld (diepte waarneming)
binocular rivalry
when two different images are shown to each eye, perception switches between them
suppression
the brain ignores conflicting information from one eye (e.g. in amblyopia)
how can misapplying depth cues lead to illusions
- Ames room (distorted size due to forced perspective)
- Ponzo illusion (parallel lines & how perspective tricks the brain)
- hollow face illusion (the brain expects convex faces)
how does the Bayesian approach apply to depth perception
the brain combines prior knowledge with sensory input to infer depth (helps resolve ambiguous or conflicting depth cues)
physiological basis of stereopsis and depth perception
- binocular neurons in V1 and beyond detect disparity
- MT (middle temporal area) processes motion and depth cues
- parietal cortex integrates depth information for spatial awareness
what is the correspondence problem in stereoscopic vision
the brain must match the correct points from each eye’s image
solved by:
- feature matching (edges, textures)
- global processing (analyzing whole scenes)
how do stereoscopes and stereograms create depth?
- present slightly different images to each eye to mimic binocular disparity
- the brain fuses them, creating 3D perception
retinal disparity
differences in images between right and left eyes
crossed disparity
objects closer than fixation point appear displaced outward
uncrossed disparity
objects farther than fixation point appear displaced inward
accommodation & convergence as triangulation cues
- accommodation = the lens adjusts shape to focus on objects at different distances
- convergence = the eyes rotate inward for near objects, outward for far objects
motion parallax as a triangultion cue
as you move, near objects move faster across your vision than far objects. The brain uses this difference to estimate depth and distance
difference between metrical and nonmetrical depth cues
- metrical cues provide exact distance (e.g. binocular disparity)
- nonmetrical cues provide relative depth but not exact distance (e.g. occlusion)
recognizing pictorial depth cues
- occlusion
- relative size
- linear perspective
- texture gradient
- shading & shadows
occlusion (pictorial depth cues)
objects blocking others appear closer
relative size (pictorial depth cues)
smaller objects appear further away
linear perspective (pictorial depth cues)
converging lines suggest depth
texture gradient (pictorial depth cues)
more texture means closer
shading and shadows (pictorial depth cues)
suggest 3D shape and position
how does 3D vision develop?
- infants rely more on monocular cues at birth
- binocular coordination develops around 3-6 months
- stereopsis emerges between 4-6 months
how does the brain compute binocular depth cues
- use binocular disparity
- disparity is processed in visual cortex to determine depth
how does the brain combine monocular depth cues
- use relative size, interposition, texture gradient, motion parallax, shading, and perspective
- integrates multiple cues for consistency in depth perception
why do we have two eyes
- provides binocular disparity which enhances depth perception (stereopsis)
- expands the field of view
- offers redundancy (one eye compensates if the other is damaged)
how does the brain reconstruct a 3D world from 2D retinal images?
- the brain combines monocular and binocular depth cues
- uses perspective, shading, texture gradients, motion, and disparities between two eyes’ images
- relies on experience and prior knowledge (Bayesian inference)
how does color influence perceived flavor?
- expectation effect (we expect a red drink to be cherry-flavored)
- cross-modal interactions (the brain combines visual and taste signals)
- marketing impact (people rate foods as tasting better when they have expected colors)
why is color vision useful?
- helps in finding food (ripe vs. unripe)
- aids in recognizing objects & emotions
- improves camouflage detection
color constancy and how it works
- the brain adjusts perceived color to remain consistent under different lighting
- uses context, memory, and lighting cues
predicting negative afterimage colors
after staring at a color, the opponent color appears in the after image (red –> green, and vice versa)
how can context influence color perception?
- color contrast (a color looks different depending on surrounding colors)
- color constancy (the brain adjusts for lighting to keep colors looking stable)
what is synesthesia?
a condition where one sense triggers another (e.g. seeing colors when hearing sound)
forms of anomalous color vision
- protanopia
- deuteranopia
- tritanopia
protanopia
missing L-cones (red-green deficiency)
deuteranopia
missing M-cones (red-green deficiency)
tritanopia
missing S-cones (blue-yellow deficiency)
does language influence color perception
some languages have fewer/more color words, which may affect perception (e.g. in russian they have seperate words for light and dark blue)
opponent color theory
it states that colors are perceived in opposing pairs (red-green, blue-yellow)
color cancellation experiments
adding a color cancels out its opponent (e.g. adding green cancels red)
how is 3D color space represented?
- RGB (red, green, blue) is used in screens and digital media
- HSB (hue, saturation, brightness) is used in color design
- CIE color space is more precise mapping of all visible colors
four ways cone outputs are pitted against each other in cone-opponent cells
- L-M (red-green opponent) compares long and medium-wavelength cones
- M-L (green-red opponent) is opposite of L-M
- S-(L+M) (blue-yellow opponent) compares short wavelength cones to the sum of L and M
- L+M (luminance channel) measures overall brightness
additive color mixing
- process: mixing light (e.g. red + green = yellow)
- example: computer screen
subtractive color mixing
- process: mixing pigments (e.g. cyan + yellow = green)
- examples: paint
Young-Helmholtz trichromatic theory of color vision
color vision results from three cone types corresponding to different wavelengths. The brain compares their responses to determine color
principle of univariance
a single photoreceptor cannot distinguish between different wavelengths, only overall light intensity
principle of metamers
two different light spectra can produce the same color perception because they stimulate cones the same way
spectral sensitivities of the three cone types
- S-cones: blue, ~420 nm peak
- M-cones: green, ~530 nm peak
- L-cones: red, ~560 nm peak
three types of cones that contribute to color vision
- S-cones: short wavelengths, blue
- M-cones: medium wavelengths, green
- L-cones: long wavelengths, red
three steps to color perception
- detection
- discrimination
- appearance
detection (color perception)
cones in the retina detect light
discrimination (color perception)
brain compares signals from different cones
appearance (cone perception)
brain interprets and assigns stable color
if an orange looks green, will it taste different?
color affects expectation of taste, but the actual taste depends on the chemical composition. But perception can be influenced by color, making us think it tastes different
why does an orange look orange in real life and in a photo, even though the physical basis is different?
in real life, we see the orange due to wavelengths of lights reflecting off it. In a photo, the camera captures those wavelengths and translates them into digital color signals that our screen displays. The brain interprets both using color constancy mechanisms
does everyone see the same color?
most people see similar colors, but color vision varies due to genetics, lighting, and perception. Color is affected by color blindness and cultural & linguistic differences
what is color for?
color vision helps us detect and recognize objects more effectively. It enhances contrast between objects and backgrounds
attention
a large set of selective mechanisms that enable us to focus on some stimuli at the expense of others
why are faces a special case of object recognition
they are processed holistically (not separate features). The FFA is specialized for face perception
objects labels at different levels of description
- superordinate = “animal”
- entry-level = “dog”
- subordinate = “golden retriever”
strengths and weaknesses of object recognition models
pandemonium model:
- strength = explains future-based recognition
- weakness = too simple for complex objects
template model
- strength = works well for specific images
- weakness = fails with variations
structural description
- strength = describes objects as 3D parts (geons)
- weakness = doesn’t explain textures, lighting effects
deep neural networks
- strength = excellent for real-world images
- weakness= requires large datasets
subtraction methods in the brain
compares brain activity with and without a stimulus
decoding methods in the brain
uses machine learning to interpret brain activity
receptive field properties of neurons that process objects and faces
- IT cortex (recognizes complex shapes and objects)
- FFA (specialized for faces)
Bayesian approach in perception
the brain combines prior knowledge and sensory input to interpret the world
methods the visual system uses to deal with occlusion
- edge interpolation (filling in missing parts)
- surface completion (assuming hidden parts continue)
- contour continuation (using Gestalt rules)
define figure-ground assignment
determines what is object (figure) and what is background (ground)
principles figure-ground assignment
- surroundedness (the enclosed region is usually the figure)
- size (smaller objects are typically figures)
- symmetry (symmetric shapes are often figures)
accidental view points in perception
occurs when an object aligns in a way that misleads perception (e.g. a person standing in front of the Eiffel tower may appear to be holding it)
Gestalt psychology
the principles state that perception is based on innate grouping rules
Gestalt principles
- proximity (close things group together)
- similarity (similar things group together)
- closure (we fill in the missing information)
- good continuation (we see smooth, continuous lines)
define midlevel (or middle) vision
the stage between low-level (edges, contrast) and high-level (objects, faces) processing. It organizes elements into coherent shapes
challenges in object recognition for the visual system
- viewpoint changes (objects look different from different angles)
- occlusion (objects can be partially hidden)
- lighting variations (shadows can distort appearance)
feed-forward processing
information flows one way (retina –> V1 –> IT)
reverse-hierarchy theory
higher areas can send feedback signals to refine early processing
visual agnosia
a condition where a person cannot recognize objects despite normal vision (caused by damage to the ventral stream, inferior temporal cortex)
dorsal pathway
- “where/how”
- location and action
- parietal lobe
- speed is fast
- damage causes optic ataxia (impaired grasping)
ventral pathway
- “what”
- object recognition
- temporal lobe
- speed is slower
- damage causes visual agnosia
concept of border ownership
cells in V2 assign edges to specific objects rather than seeing them as just standalone lines
differences between extrastriate cortex and striate cortex
- striate cortex = V1, processing of basic features (edges, orientation)
- extrastriate cortex = V2-V5, processes higher-level aspects (shapes, motion, object identity)
how could a computer recognize objects?
- feature detection
- template matching
- deep learning