Theory of methods Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

Psychophysics

A

Psychophysics is the study of the relationship between physical stimuli and the sensations or perception that they evoke. It is significant in psychology because it aims to quantify the psychological experience of stimuli and understanding of the underlying principles of sensory perception

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

In psychophysics, one method that we can use to get a threshold measurement is the method of constant stimuli. Describe this method and explain what additional information can be obtained when using this method.

A

In method of constant stimuli the researcher choose several stimuli with varying intensities, these are fixed intensity levels. The stimuli are then presented to the participant in a pseudorandomized order and the participant answer seen/not seen for each stimuli level. An example of method of constant stimuli is letting participants listen to beforehand chosen volume levels, lets say 1-5 and playing them at random maybe 3, 1, 4, 2, 5. The participant answer each time if they can hear it or not. The information that can be obtained from this is a psychometric function, which is a mathematical model of the relationship between stimuli intensity and perceptive experience.

Additional information obtained from the method of constant stimuli:

Sensitivity Analysis: By using a range of stimuli, the method allows researchers to assess the sensitivity of the participant’s sensory system across different levels of stimulation. This can provide insights into the participant’s ability to discriminate between different intensities or qualities of stimuli.

Detection Thresholds: In addition to measuring discrimination thresholds (the ability to distinguish between different stimuli), the method of constant stimuli can also be used to measure detection thresholds (the ability to detect the presence of a stimulus at all). This is particularly useful for understanding the absolute limits of sensory perception.

Individual Differences: By analyzing the variability in participants’ responses, researchers can gain insights into individual differences in sensory perception. This information can be useful for understanding factors such as age-related changes in perception or differences between clinical populations and healthy controls.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

On a signal detection/Yes-No task, the proportion of Hits and False Alarms depends on two parameters, sensitivity and criterion. Describe what these two parameters represent in signal detection theory.

A

SDT is a framework for analysing how people make decisions during uncertainty, particularly in perceptual tasks. The goals is to discriminate a signal (meaningful information that we are intrested in) from noise (not meaningful). This can result in four different answers, two correct and two wrong. Hits, which means detecting signal that is present. Miss, which means failing to detect a present signal, Correct rejection which means saying a signal isn’t present when it isnt and False alarm which means wrongly identifying a signal as present.
Criterion is the decision threshold for answering that the signal is present. The criterion can be influenced by bias and is subjectively set by the participant. If the perceptive information surpasses/exceeds the criterion you answer signal. If the criterion is lowered, the number of hits go up but the risk of false alarms also goes up. Sensitivity is how easy it is to detect the signal, set by the researcher or the one administrating the test. Let’s use where’s Waldo as an example. The goal by the participant is to find Waldo (signal) in a picture with a lot of people in it (noise). The criterion for Waldo might be that he has glasses, a hat and and a striped shirt. Let us say that you can’t find him so you change the criterion to only be a striped shirt.This will increase you chances of finding him but also increase the risk of misidentifying another person with a striped shirt as him. Using the same example for sensitivity, the researcher could choose to increase the size of waldo (easier to detect) or make the picture black and white (harder to detect). If the sensitivity goes up so does the hits while the risk for false alarms go down.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Who was Eadweard Muybridge and how was his work relevant for motion measurement? (1p)

A

Eadweard Muybridge was the first person to “film” a moving object. He did this by placing multiple cameras by a horse racing track and taking pictures of a horse. Then he put all the pictures together, showing the motion of a horse running, and proving that horses didn’t have any feet at the ground at one point while running. This revolutionized motion measurement and was the first ever motion picture

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Explain why “dead reckoning” can cause problems for researchers who use accelerometers to study behaviour and describe two principal strategies one can use to minimize ‘dead reckoning problems’ in accelerometer-based assessments

A

“Dead reckoning” in accelerometer-based behavior studies refers to the potential inaccuracies caused by estimating position solely from accelerometer data. To minimize these problems:

Integration with Other Sensors: Combine accelerometer data with gyroscopes, magnetometers, or GPS to improve accuracy.

External Calibration and Validation: Periodically validate accelerometer data against external reference points or measurements to correct for drift and errors.

These strategies help researchers mitigate dead reckoning issues and enhance the reliability of their behavior tracking.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Name and describe two different ways of assessing the reliability of observation coding.

Reliability in observational methods

A

One way to assess the reliability of observation coding is checking Interatter/observer reliability. You can do this by letting two or more observers observe the same thing. If the observers agree, and code behaviors the same way for example, the interobserver reliability is higher. A way to increase the chances of agreement is having clear definitions of behaviors and criteria for coding.
Another way to test reliability is Test-retest reliability. This can be assessed by letting the same observer observe the same thing more than once and see if they code the behaviors the same both times.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

When making observations the researcher often change the study participants’ behavior and make the observation less valid. Name at least four strategies to reduce participant reactivity to the observer.

A

One way to minimize participant reactivity is to make the observation double-blind, meaning that both the participant and the observer are blind to the study’s hypothesis, this reduces participant reactivity and expectancy effects from the observer. You can do this by taking in an observer that is not familiar with the ongoing research for example. Another strategy is to not observe for the first 10 minutes for example, it both minimizes observer drift and participant reactivity. You can also say you are going to actively observe for only 10 minutes during the hour, but not which ten minutes and thus minimize the participant reactivity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

You would like to know how the medial temporal lobe reacts when watching cat-gifs. You are going to use both fMRI and PET for this wonderful study. In the PET-study you are going to use a tracer for the serotonin transporter called DASB. Both the PET and fMRI study are using subtraction design. Using the simplest possible design for both PET and fMRI:

What is subtracted from what in the PET study? How many sessions will you run with each participant?

What is subtracted from what in the fMRI study? How many sessions will you run with each participant?

A

PET is a functional brain imaging technique where you can study neurochemical processes such as metabolism for example. In this method a radioactive tracer is injected and as it decays in the body it emits gamma rays that can give us information about where the neurochemical process takes place. In this study I will use a tracer for serotonin. In subtraction design for this method I will inject the tracer in the participant and measure the baseline activation before showing cat gifs, then I will repeat the process and do it while showing cat gifs. PET has poor temporal resolution and is not suited for measuring quick changes, thus I would run it two times. Since it is also exposure to radiation because of the tracers I wouldn’t do it in too many sessions. What is subtracted here is the normal baseline level of serotonin in the medial temporal lobe, from the serotinin level while watching cat-gifs. With this method you can see serotinin levels that are only cause by the task and not baseline.
fMRI is also a functional brain imaging technique, but in this method you can see activation in the brain based on the different magnetic properties of oxygenated and non-oxygennated blood. The activated parts of the brain uses more oxygen and thus this method is valuable for functional imaging. When using subtraction design with this method you first measure the “baseline” activation of the medial temporal lobe when you are not watching cat gifs and then measure again during the cat gifs-task. fMRI is an indirect measurement of activation and it’s spatial resolution is good, but since the temporal resolution is a bit worse (about 1s per frame) you could miss activation by only measuring the participant once, if I had resources I would make sure to run the session multiple times to not miss any activation. But the simplest
way would be to run it in in one session, first measuring the baseline and then showing the cat- gifs. What is subtracted is the baseline activation from the activation from the cat gifs-task. Which in fMRI would mean, the level of oxygenated blood in the medial temporal lobe while watching gifs minus the level normally. This would give me information about activation caused only by the cat gifs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Provide a psychological construct (e.g., extraversion) and create an implicit test to measure it.

A

Implicit Test: Implicit Association Test (IAT) for Self-Esteem

Task: Participants categorize self and other-related words/images using key presses.

Stimuli: Self-related and other-related words/images are presented randomly.

Category Pairings: Participants categorize stimuli, pairing self with positive/negative attributes.

Response Time: Faster response times for self-positive pairings indicate higher implicit self-esteem.

Analysis: The test calculates a D-score, indicating the strength of association between self-esteem and positive/negative attributes.

This test provides insights into individuals’ automatic evaluations of themselves.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Describe how to examine the construct validity and the reliability of your implicit measure

A

Convergent Validity: Assess the degree to which the IAT correlates with other measures of self-esteem, such as explicit self-report questionnaires like the Rosenberg Self-Esteem Scale. A strong positive correlation would support the construct validity of the IAT.

Reliability:

Internal Consistency: Calculate the internal consistency of the IAT by examining the correlation between responses to different pairs of self/other and positive/negative attributes. Higher correlations indicate greater reliability.

Test-Retest Reliability: Administer the IAT to the same participants on two separate occasions and examine the correlation between their scores. A high correlation suggests that the measure is stable over time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Magnitude matching
/adjustment task

A

ex. participant has a weight in their hand, and should say when the weight in other hand matches the first weight

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

ICC

A

Item characteristic curve

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Reflexive Thematic Analysis

A

FUCK GNUER
FCGNR

Not a theory, but an approach to how to process data in Qualitative research

Familiarize
Code ( categorically, descriptive or analytically )
Generate themes (related to research question
Name themes
Report

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Signal Detection Theory

A

Framework used to analyze decision making in the presence of uncertainlty, particularly in the situations where a participant must distinguish between a signal and noise
Signal: meaningful info, what we’re trying to detect
Noise: non meaningful info

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Occam’s razor

A

If two models have the same goodness of fit,
choose the more simple one
How well the model explains the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Ordered response categories

A

A higher person location (sum score) on the latent variable should entail an increased probability of a higher response (category) for all items and vice versa. Sometimes referred to as ‘monotonicity’

We can check this by looking at the item characteristic curves (ICC)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Which of the following items are included under “methods” when planning & reporting qualitative research?
Limitations
Interpretation
Context
Purpose/research question

A

Context

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Which research approach is based on the epistemological viewpoint of pragmatisms?
Mixed methods research
Quantitative research
Qualitative research
All of the above

A

Mixed methods research

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

2 ways of assessing reliability of observation coding

A

Check interrater/observer reliability Do this by letting 2 or more observers observe the same thing. If the observers agree about how they code behaviors, the reliability is higher. To increase the chance of this, have clear definitions of behaviors and criteria for coding

Test/retest reliability
Let the same observer observe the same thing more than once and see if they code the same all of the times.

20
Q

Criterion in SDT

A

The decision threshold for answering that a signal is present
Can be influenced by bias and is subjectively set
If the perceptive info surpasses/exceeds the criterion you answer “Signal”
If criterion is lowered (where’s Waldo looking for striped shirt instead of glasses or whole of him)
the number of hits go up, but there’s also a higher risk of false alarms

21
Q

Observations:

Unsystematic Observations

A

explore a phenomenon with minimal preconcieved notions

naturalistic setting, allowing for genuine insights & its flexibility to capture unexpected occurences

limits: potential bias, lack of structure, challenges with replicability

Good when aiming for in-depth understanding of a complex dynamic environment, esp in exploratory studies

22
Q

Observations:

Naturalistic

A

Observation without interference

The context makes actions meaningful, therefore it is important to study natural environments

High ecological validity
No manipulation
but there is a research question and hypothesis

23
Q

Observations:

Experimental

A

controlled, in a lab setting
Characteristic: test a hypothesis
Many factors are kept constant
Manipulate one factor… What happens?
Laboratories typically use this
tries to establish causal relationships

24
Q

Observations:

Field study

A

Mix between naturalistic and experimental observations.

Experimental method in a real environment

Introduces a systematic change in a natural environment

25
Q

Kinetic
&
Kinematics

A

Muscle force (kinetics)
creates movement (kinematics)

26
Q

Observational research scales

A

Interval: the same distance between each point on scale Like temperature.
There’s no absolute zero. like numbers! there is zero but no absolute… minus

Ordinal: not the same distance between the points. a specific order… gold silver bronze. Likert scale

Categorical (på svenska nominalskala): no inherent order. ex classifying behaviors eg. aggressive or friendly, man or woman

Ratio (skalnivå): There is an absolute zero. Like weight. Equal steps

27
Q

Comparison between IRT and CTT

A

Item response theory and Classical Test theory

In IRT they test one latent trait, but in CTTthey test different skills

IRT includes analysis of item location difficulty, not CTT

CTT assumes that reliability is constant across the continuum, & equal reliability for all respodents

In classical test theory the observed score (X) on a test is assumed to contain the true score (T) and an measurement error (E). In CTT the overall test score is of focus and the overall relibility of the test. It doesn’t consider probability or analyze any items separately. Any respondents are considered to have constant reliability. In Item response theory the aim is to measure an underlying construct, or a latent trait. All items are analyzed separately with consideration of three parameters: difficulty, guessing and discrimination. IRT also consider monotonicity which means that as the ability of the respondent goes up, the probability of choosing a correct/higher response should go up. CTT doesn’t explicty try to measure the latent ability and is more directed towards measuring overall score on a test.

28
Q

Which of the following is a phase in reflexive thematic analysis?

Counting codes
Familiarization
Coding reliability checking
All of the above

A

Familiarization

29
Q

ROC curve

A

Part of SDT

A graphical representation of an observer’s ability to discriminate between signal and noise at different decision criteria.
The curve shows the trade-off between hits and false alarms.

30
Q

Classical test theory

A

An observed score consists of
a true score and a random error

CTT maps overall test score (true score + error) and it doesn’t account for the probability of answering specific items correctly. It assumes constant reliability for all people and doesn’t check if more ability = higher scores.

31
Q

FACS

A

Observational: Facial action coding system

Standardized way of analyzing and describing facial expressions

standardized so no observer is physically needed to be present

32
Q

Longitudinal
cross sectional
and experimental
surveys

A

L: 1 person over time

C: Multiple people to get data frpm target demographic

Cross-sectional: Attitudes to bananas among the population or the
relationship between emotional instability and attitudes to bananas
Longitudinal: The evolution of attitudes to bananas in the population
Experimental: The effect of stress on attitudes to bananas

33
Q

When making observations the researcher often changes the study participants’ behavior and makes the observation less valid. Name at least 4 strategies to reduce participant reactivity to the observer

A
  1. Naturalistic observation: reducing the likelihood of altering their behavior in response to the observer’s presence
  2. Blind observation. concealing the identity or puspose of the observer
  3. Particpant familiarizarion. introducing the pps to reduce novelty effects
  4. Behavioral recording devices
    recording without observer
34
Q

Describe briefly how a psychophysical magnitude matching/adjustment task is performed

A
  1. Presentation of two stimuli: Present participant with one standard stimulus with fixed intensity and one comparison stimulus with adjustable intensity
  2. Matching/adjustment process: PP can adjust the intensity of the comparison stimuli until it matches the perceived intensity of the standard stimuli.
  3. Repeated trials: Multiple trials with different standard stimuli and comparison stimuli to assess perceptual judgments across a range of intensities
  4. Data collection: Record the settings or adjustments made by PP for each trial. –> Analyze perceptual sensitivity, discrimination thresholds or other psychophysical parameters
35
Q

In modelling, we make a distinction between descriptive and process models. Explain what is meant by a descriptive model and what is meant by a process model (in your answer, provide an example of a descriptive model and a process model)

A

Descriptive models give us information about data but don’t include any psychological content. Process models contain psychological content to describe the data. One example of a process model is the Generalized context model (GCM). In this model the process of facial recognition for example can be described as a calculation of similarity in regards to stored exemplars in the brain, with both recognition and categorization as parameters.

Descriptive models summarize observed data or patterns without capturing underlying mechanisms, often used for prediction or visualization (e.g., linear regression). Process models represent dynamic interactions and causal relationships among variables to simulate system behavior or understand complex phenomena (e.g., population dynamics simulations). Descriptive models aim to describe relationships, while process models aim to understand underlying processes. Both are valuable in modeling but serve different purposes: descriptive models provide summaries or predictions based on observed data, while process models offer insights into system dynamics and behaviors over time or under different conditions

36
Q

Modelling can be said to a cognitive aid for the scientist. Explain why this is the case

A

Modeling serves as a cognitive aid for scientists by facilitating understanding, prediction, and exploration of complex phenomena. It provides a structured framework for organizing knowledge, clarifying hypotheses, and testing theories. Through the process of creating and refining models, scientists can gain insights into the underlying mechanisms of systems, identify key variables, and make predictions about their behavior. Models also allow scientists to explore hypothetical scenarios and simulate experiments that may be difficult or impossible to conduct in the real world. Overall, modeling enhances cognitive processes by providing a systematic approach to conceptualizing, analyzing, and interpreting scientific phenomena.

37
Q

Describe the main differences between classical test theory and item response theory/Rasch measurement theory. Keep your explanations brief

A

In classical test theory the observed score (X) on a test is assumed to contain the true score (T) and an measurement error (E). In CTT the overall test score is of focus and the overall relibility of the test. It doesn’t consider probability or analyze any items separately. Any respondents are considered to have constant reliability. In Item response theory the aim is to measure an underlying construct, or a latent trait. All items are analyzed separately with consideration of three parameters: difficulty, guessing and discrimination. IRT also consider monotonicity which means that as the ability of the respondent goes up, the probability of choosing a correct/higher response should go up. CTT doesn’t explicty try to measure the latent ability and is more directed towards measuring overall score on a test.

Classical Test Theory (CTT) focuses on observed test scores and assumes that observed scores consist of true scores and measurement error. It does not consider item properties. In contrast, Item Response Theory (IRT) and Rasch Measurement Theory consider both item and person properties. IRT models how individuals respond to test items based on item difficulty, discrimination, and person ability, providing insights into item and person characteristics. Rasch Measurement Theory is a specific form of IRT that focuses on item difficulty and person ability, aiming for measurement scales with interval properties, allowing for more precise measurement across the entire continuum of ability.

38
Q

It is important to choose the right tools to do reliable observations. Compare the pro and cons with
using electronic registration versus paper and pencil. (2p)

A

Electronic registration is good when you want to observe a big amount of data, since you can
use different types of systems and programs on the computer to aid you in the observations.
Electronic registration also makes the observation notes easier to interpret, there’s no risk of
someone not being able to read what someone has written after the fact. With electronics you
can record videos and sound, which can later be played back to refresh the memory of the
researcher. The cons with electronic registration is that it can all be lost, if something would
happen to the computer (it breaks, or doesn’t work etc). Another con it that it takes up a lot of
space, and could be hard to bring with you as a researcher. If the equipment can be
transported it could also lead to novelty effect and other effects on the participant, if they feel
observed and therefore act differently.
The pros with paper and pencil is that people are used to it, and it might not be as intimidating
as electronics being used. A con is that if the researcher is stressed or want to note a lot of
things at the same time, this might lead to the notes being scribbles and therefore not readable
after the fact. The paper and pencil could also break, or be a finite resource, during the
observation. This could lead to the researcher having to abort the observation, or not being
able to write anything down for the remainder of the observation. The participant could also get
annoyed or affected by the sound of the researcher noting things. They could for example
change their behaviour if they notice that the researcher always write something down when
the participant does a specific action.

39
Q

You would like to know how the medial temporal lobe reacts when watching cat-gifs. You are going
to use both fMRI and PET for this wonderful study. In the PET-study you are going to use a tracer
for the serotonin transporter called DASB. Both the PET and fMRI study are using subtraction
design. Using the simplest possible design for both PET and fMRI:
a) What is subtracted from what in the PET study? How many sessions will you run with each
participant? (2p)
b) What is subtracted from what in the fMRI study? How many sessions will you run with each
participant? (2p)

A

a) In the PET study I will run 2 sessions with each participant. I will do this because because I
would like to get a scan of the participants neutral brain that is not being presented with cat-gif
stimulus, or any other stimulus for that matter. On the second session with the participant I
would present them with the stimulus of cat-gifs. I would then use the subtraction design, to
subtract the neutral brain scan of the participant from the scan where they were presented with
a stimulus. This will make the scan show what parts of the brain produce more serotonin when
presented with cat-gifs. Since the serotonin tracer is used in both sessions, you can see the
change in serotonin production (and regions in the brain) when presented with the stimulus.

b) In the fMRI study I would also use 2 sessions with each participant, to be able to use the
subtraction design most accurately. fMRI uses BOLD (blood oxygenated level dependent)
which means it measure where in the brain there’s a lot of oxygenated blood. If there’s a lot of
oxygenated blood going through a specific part of the brain, it means this part is activated more
than others.

Since both PET and fMRI are abrasive on the body I would not want to run too many
sessions with each participant, but I would regard 2 sessions as the minimum to get an
accurate result as well as the maximum to not make the participant be subject to too much
radiation.

I would also do the 2 sessions because using another persons brain in the subtraction design
would not be as accurate. Each person’s brain is different and individual to them, so
subtracting another person’s brain scan would not create the same results.
Both PET and fMRI are functional measures, and have good spatial resulotion.

40
Q

You have an (corneal reflection) eye tracking experiment where you are interested in whether
people look to object A or object B. The objects are the only things shown on the screen, and they
are spatially well separated (one to the far left and one to the far right).
For one of your participants, Jonas, you have problems calibrating the eye tracker, but otherwise
the experiment seems to be working and recording Jonas’ gaze. You consider using the previous
participant, Calle’s, calibration data and apply it to Jonas’ recording.
Could it be a good idea using Calle’s calibration data when running the experiment on Jonas in this
case? (1p)

A

Integrity of the Data
Researchers have an ethical obligation to ensure that the data collected is valid and scientifically sound. Using another participant’s calibration data compromises the integrity of the experiment. This introduces a systematic error, which can distort the findings, and results in misleading conclusions about Jonas’ behavior (e.g., where he is looking).

This issue is not just a methodological flaw (Individual Differences in Calibration, systematic errors, precision and accuracy
but also a breach of the ethical standard of accuracy in research.
From an ethical standpoint, using Calle’s calibration data for Jonas is not advisable. The ethical principles of accuracy, honesty, integrity, informed consent, and fair treatment of participants all dictate that each participant should undergo their own calibration process to ensure that the data you collect is valid and represents their true gaze behavior. Failing to do so risks violating these principles and can undermine the credibility and ethical standing of your research.

41
Q

Present one research question (of relevance for psychological science) that motivates the use of
a motion capture system and one research question which rather motivates the use of an
accelerometer. Motivate your answers. (3p)

A

Motion capture measures position. It has to be perfomed in a lab setting with the use of three
cameras, the participant will have some markers on his body and they will take the position.
This method is useful to measure actual performance, for example if we want to compare
analyze object manipulation at 6 months versus 12 months. We present the child with a
container and a set of blocks and we want to see how the infant approaches the objects
without any prior instruction. The research question: how does object manipulation changes
between the ages of 6 and 12 months?

Accelerometers are used to measure force or acceleration. It would be a good tool to measure
hyperactivity in toddlers. The research question: can hyperactivity be measured in toddlerhood
as a potential risk factor for ADHD? The benefits are that it is non instrusive, it is a small
device that the child can wear all day and anywhere.

42
Q

Name and describe at least 3 common metrics when doing quantitative behavior observations.

A

Frequency, duration, and latency are three core metrics in quantitative behavior observation that provide essential insights into how often, how long, and how quickly a behavior occurs. They allow researchers to quantify and analyze behavioral data objectively and are commonly used in psychology.

Definition: Frequency refers to the number of occurrences of a specific behavior within a given time period.

Definition: Latency measures the reaction time or response time to a stimulus or event. It is especially important in studies that assess reaction speed or decision-making in response to stimuli.

Definition: Duration measures the amount of time a specific behavior lasts or the time spent engaged in a particular activity.

43
Q

Name the five basic psychometric criteria presented at the lecture. Use one sentence for each
criterion to describe very briefly what it is about. 2p

A

Reliability refers to the consistency of a measurement, ensuring that the results are stable and repeatable across time, raters, or items.

Validity assesses the accuracy of a test in measuring what it is intended to measure, ensuring it truly evaluates the intended construct.

Sensitivity measures the ability of a test to correctly identify individuals who possess the condition or trait being assessed, minimizing false negatives.

Specificity evaluates the ability of a test to accurately identify individuals who do not have the condition or trait, minimizing false positives.

Fairness ensures that a test is free from bias and is equally appropriate for all individuals, regardless of background or group membership.

44
Q

Provide a psychological construct (e.g., extraversion) and create an implicit test to measure it.
(2p)

A

Step 1: Categorization Tasks

Participants will be asked to quickly categorize words into one of two categories (e.g., “Extraversion” vs. “Introversion”) using a computer. The categories and associated words will reflect extraverted and introverted behaviors or qualities. For example:

Extraversion words: Social, Energetic, Outgoing, Talkative, Enthusiastic, Active
Introversion words: Solitary, Quiet, Reserved, Shy, Private, Calm
Participants will also categorize the following self-relevant words to measure their implicit association with these categories:

Self-relevant words: Me, Myself, I, Us, We (indicating the participant’s self-concept)
Other-relevant words: Them, Their, They, Others (representing others in the participant’s mind)
Interpretation:
Stronger association with “Self” and “Extraversion”: If a participant responds faster when words like “Energetic” or “Talkative” are associated with “Self,” it suggests they have an implicit self-concept aligned with extraverted traits. This could indicate a higher degree of extraversion.

Stronger association with “Other” and “Introversion”: If the participant reacts more quickly when introverted words are paired with “Self,” it would indicate a stronger implicit association with introverted traits, suggesting a lower level of extraversion or a tendency toward introversion.

Benefits of the Implicit Extraversion IAT:
Automatic Associations: The IAT uncovers automatic, unconscious associations, which can often reveal more about an individual’s true psychological tendencies compared to self-report measures that may be subject to social desirability bias.

Objective Measurement: The IAT provides an objective measurement of how strongly a person associates certain traits with themselves (e.g., associating “Social” with “Me”) as opposed to relying solely on verbal or self-reported answers.

45
Q

Describe how to examine the construct validity and the reliability of your implicit measure. (2p).

A

To evaluate the construct validity of the Implicit Extraversion IAT, you would focus on convergent validity (correlation with other extraversion measures), discriminant validity (low correlation with unrelated traits), predictive validity (association with real-world extraverted behaviors), and theoretical consistency (alignment with expectations based on extraversion theory).

For reliability, you would assess test-retest reliability (stability over time), internal consistency (agreement within the test itself), inter-rater reliability (consistency of scoring or coding), and item reliability (consistency of words used in the test). A valid and reliable IAT would produce consistent and meaningful measurements of extraversion, helping to ensure that it accurately captures implicit associations related to this personality trait.