Part 3 (Final) Flashcards

1
Q

Name 4 areas where it is important to follow ethical guidelines in psychological assessment

A

test/questionnaire selection
scoring & interpretation
participant recruitment
treatment of participants

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does treatment of participant ethical guideline entail

A

Responsible treatment of participants
Even decisions made prior to seeing any participants are part of their ethical treatment
e.g. understand your measures, follow appropriate procedures, ensure participant wellbeing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is included in informed consent

A

It includes things like estimating the time commitment and informing them about their right to withdraw, and your goals but not hypotheses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Name some elements of ethical guideline “basics of wellbeing”

A
  1. informed consent
  2. assured confidentiality, as much as possible
  3. deliver feedback appropriately
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is included in confidentiality

A

this includes you practicing responsible handling of data

open-access data must be anonymous, people must be aware of the broad distribution of data, must plan in advance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is included in deliver feedback appropriately

A

ideally, there should be some benefit to the participants

usually, we cannot give them any scores if we do not understand the measure already

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are some regulations around children in research

A

Children 14-18 participating in research must be towards a project that directly benefits them or children like them

Children under 14 must have parental approval, parents and children can withdraw at any time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Define Responsible Data Handling

A

No unauthorized access to your data, this includes the government

Safe retention of data
anonymity goes a long way here - don’t collect personally identifying information if it’s not needed
traditionally, keep for 5 years then destroy (not necessarily anymore)

In testing, ensure the results won’t be interpreted inappropriately

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Define Responsible Data Removal

A

Data removal is allowed, but needs to be done responsibly as well

Balancing need to respect participants’ contribution with the need to reach accurate conclusions

Participants have the right to try to derail our studies, but not the right to succeed in doing so

Remember, our statistics have assumptions we should meet many statistics require complete data sets

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the assumptions of the factor analysis

A

Based on general factor model, factor analysis shares some important assumptions with more classical measurement models

  1. Errors are random and not correlated with the latent variable
  2. Correlations among items exist because they share a common latent variable(s)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a factor

A

Factor is another way we could refer to a latent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a component

A

Component is another way we could refer to a latent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Why would we use the words “factor” or “component” instead of “latent variable”?

A

The main reason to use these terms instead of latent variable is to better acknowledge that the unobserved influence was derived empirically

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How can you identify factors

A

Factor analysis is a process of trying to identify the latent variable(s) that influenced our measurements

Items that have stronger associations (correlations) with each other but weaker associations with other items will form identifiable clusters

we’re capitalizing on similarities and differences in correlations across items

if all items correlate strongly, there will be only one factor identified

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the main element to have a good factor analysis

A

We need quite large data sets for factor analysis to give accurate results: larger than what we need for good internal consistency reliability or validity analyses
we should have 50% more data for factor analysis than for reliability/validity analyses (around 300-400) to have a meaningful factor analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What should you keep in mind as you run factor analyses (4)

A
  1. What you put into the analysis dictates what you get out of it
  2. Every item has the potential to create a factor, and influence the creation of other factors
  3. Adding or dropping even one item will change the outcome
  4. Factor analysis should be viewed as a process requiring many iterations, thus it is time consuming
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

When should a cluster be called a factor

A

Technically, we always find more than one ‘cluster’ or pattern of responses in a questionnaire - only the important ones get called factors

Researchers typically select as factors any components with an eigenvalue > 1.0

The eigenvalue is a measure of the amount of information captured by an item

An eigenvalue of 1 is generally seen as indicating the factor capture as much information as one typical (good) item

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is a parallel analysis

A

Parallel analysis created random data with the same number of variables and observations as your data. A correlation matrix is created with the random data and then eigenvalues are calculated

When the eigenvalues from the random data are larger than the eigenvalues for your real data, you know the variables in the factors are not correlated better than random noise

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

How do you interpret a scree plot based on eigenvalues against a parallel analysis

A

All the eigenvalues are plotted, and so are the stimulated eigenvalues
Here the simulated eigenvalues comes from the randomly generated Parallel Analysis data set
What jamovi will call a factor is any blue dot that is before the first time a yellow dot rises above the blue dots
Factors are just numbered sequentially, after sorting the eigenvalues largest-to-smallest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

How do you interpret a traditional scree plot

A

The same eigenvalues are plotted, it’s just the comparison line that is different
Looking at the scree plot, where do you think we change mountain to ‘scree’
The number of factors identified to the left of where we think the screen starts is how many we should keep

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What are some changes you can do to an exploratory analysis that will make a difference

A

changes worth considering: dropping one or more item(s), adding a new item, using rotation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What are some changes you can do to an exploratory analysis that will not make a difference

A

changes that won’t impact anything: removing participants with missing data, reverse-scoring

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Briefly describe the exploratory factor analysis

A

Exploratory factor analysis is iterative, you repeat until a good solution arises
Identifying the number of factors is based on eigenvalue, using whatever method makes sense

24
Q

Explain the different rotations for EFA

A

Rotation allows us to spread the variability among our factors more evenly
There are two main forms of rotation worth considering

Orthogonal: this maximizes the squared variance in the factor loadings - clusters are as different as possible, unrelated

Oblique: this maintains some relationship between the factors-clusters are different but related things

25
Q

What is the varimax rotation for EFA

A

tried to maximize the differences between your clusters by rotating the factor axes
This is called varimax or quartimax rotation in jamovi
Varimax rotation is very commonly used. It makes it easier to identify the differences between clusters of items (or which items best represent which cluster)

26
Q

What is the oblique rotation for EFA

A

Tries to maintain some of the association between your clusters (and measure it) while rotating the factor axes
This is called oblimin or promax rotation in jamovi

Oblimin rotation is used to minimize the squared loading covariance of the factors, while allowing them to be correlated
as different as possible while still being related

This oblimin rotation is, obviously, not an orthogonal rotation

This means the horizontal and vertical axis (for 2 factors) aren’t at 90 degrees
The strength of the correlation dictates the extent of the departure from a 90-degree angle

The results obtained in jamovi would look the same as with a varimax rotation, or even an unrotated solution, but it will have different factor loadings

27
Q

Where do we expect the variance to come from for EFA

A

Multiple sources of variance

In Exploratory Factor Analysis we expect the latent variable(s) to be only part of the forces acting on our measurements. The other influences are probably just random error, though.

That means the eigenvalues we calculate are affected by our conception of how to attribute the sources of variance

28
Q

What is the principle component analysis

A

In Principle Component Analysis we expect all variance to come from common sources among the items, through not necessarily to an equal extent in each

This means there is no choice of extraction method, as these options under EFA are based on different ideas of the sources of variance

It’s quite possible for some of the common variance to still be error, though; accordingly, we still want to discard any ‘factors’ that seem too trivial - on the basis of their eigenvalues

29
Q

When should you use PCA or EFA

A

Technically, PCA is not considered factor analysis. You should use it if you want to maximize the amount of explained variance

If someone asks you to do a factor analysis, you should do EFA

30
Q

Describe the EFA and the CFA

A

Exploratory Factor Analysis (EFA) allows very little control over the latent variables, theories are introduced through rotation or forcing a set number of factors

Confirmatory Factor Analysis (CFA) requires you to exercise control over which items theoretically go with which latent variable; with the right software, not jamovi (yet), you could even specify expectations for how strong the relations should be

31
Q

What should be the process for CFA

A

If you go back and look at the process for EFA, it has a very similar structure
The main difference here is that we need an a priori theory and we need to check that theory
The examination of the model fit dictates whether we are done or need to revise our model or our theory

32
Q

What statistics should you look at in a CFA

A

use standardized estimates, preferable, recognizing that each of the factor loadings should be significant (p < .05)

examine the chi squared as well as the CFL, TLI & RMSEA statistics

you want the chi squared to be non-significant, though that’s unlikely

you want the CFI & TLI to be above .95 (below .90 is very bad)

you want the RMSEA to be below .05 (not more than .08 upper CI)

33
Q

Give the definitions for standardize (3)

A
  1. turning into z-scores
  2. norming
  3. an expected set of procedures, or that the items on a questionnaire have been previously established and we didn’t change them, that a measure has established normative information available, or some combination of all these. Unstandardized means one or more of these is missing
34
Q

What is a proprietary measure, which are standardized vs unstandardized

A

Another big difference between many standardized and unstandardized ones are proprietary. That is, you have to pay to use them.

Standardized: Controlled administration conditions, normative information available, interpretation guidelines available, instruction manual available, carefully vetted (?), and probably costly

Unstandardized: May not require controlled conditions, typically lacks normative information, less interpretable for giving feedback, mainly useful for research purposes

35
Q

Are proprietary measures of personality better?

A

for measuring personality, free tests are better: more reliable and efficient

36
Q

What is a hierarchical factor analysis

A

You may recall I mentioned that it’s possible to theorize and (with different software) test multiple levels of latent variables. There’s a name for this.

Hierarchical Factor Analysis is one where you have proposed one or more latent variable(s) causing changes in one or more other latent variable(s)

The idea here is you have a primary latent variable that influences the other latent variables

37
Q

Why use a hierarchical factor analysis

A

We’re often focused on measuring just one factor, but this isn’t always what we want.

A test for a course may not capture just one construct
An IQ test captures G (overall intelligence), but it also has multiple components within it

A personality “test” assesses many aspects of personality; the Big 5: Openness, Conscientiousness, Extraversion, Agreeableness, Neuroticism
they aren’t really tests, they are questionnaire

38
Q

What inter-item correlations should we expect between latent variables in a hierarchical factor analysis

A

above .3, no more than .8

39
Q

Why use experiments

A

In experimental research we are usually trying to manipulate something, and to observe the effect of that manipulation on an outcome.

Many studies use performance on some task as their outcome

Often, they just compare two different groups of people as the “manipulation”
Very rarely are psychometrics evaluated in these contexts

40
Q

Why should we evaluate the psychometric properties of experiments

A

If we don’t, we may:

Making less-than-optimal design decisions (e.g., power, task selected)
Failure to find predicted results
Incorrectly interpreting results

41
Q

Define discrimination, reliability and validity in a cognitive task

A

Discrimination: Scores aren’t entirely random (e.g., high/low performance by some individuals)
similar to good distribution and range from questionnaire, can be thought of as criterion validity

Reliability: Scores are a precise estimate of one’s ability

Validity: Scores reflect the right cognitive faculty

42
Q

What is the risk of assuming the same reliability for control and treatment groups

A

The control group may not have the same reliability, because of response biases

Assuming reliability would be equal could lead to incorrect conclusions

43
Q

What is the difference between tests and questionnaires

A

For tests, items have right and wrong answers, whereas for questionnaires they do not (except with respect to the truth of the response for that individual).
validity, reliability but also suitability of purpose is important for tests

We are still interested in psychometric qualities for tests, but not all concepts apply equally well:
Internal consistency reliability may not make much sense

This depends on how many factors we expect to capture using the test, and how strongly correlated the factors are

44
Q

Define discrimination and difficulty for tests

A

For proper tests, there are two important statistics to consider – in addition to reliability and validity.

Discrimination: The ability to accurately separate individuals into high/low performers
Discrimination is based on a relatively simple formula, especially compared with everything else we’ve seen.
Discrimination = Ph/Nh - Pl/Nl

Difficulty: The likelihood of getting a question correct

Ph = People responding correctly (H) 
Pl = People responding correctly (L) 
Nh = Number of people in group (H) 
Nl = Number of people in group (L)
45
Q

What is the goal of discrimination in tests

A

Remember, the goal of discrimination is good categorization of individual cases into high and low performers.

That means we ideally have approximately half of people getting items wrong, and we want to be able to predict in which half an individual falls with relatively good accuracy – for each item

So, there is typically a connection between discrimination ability and item difficulty

46
Q

How do you calculate difficulty in tests

A

The item’s discrimination score should be interpreted in light of the item’s difficulty score.
Difficulty is 1 minus the proportion of correct responses from all individuals
Difficulty = 1 - Correct/N

47
Q

How would you optimize a test for research

A

If you are designing a test for research purposes, you want to maximize the variability in the scores (think, full range of possible responses).

That means you should aim for most items to have a difficulty score near the optimal .50
This gives you equal room above and below, across participants

With optimal difficulty, you have optimized the probability of getting a high discrimination score – though it is by no means certain

This means your item can separate good performers from poor performers

48
Q

What are some psychological considerations when creating tests

A

People are naturally inclined to want to do “well” on their tests. For research, this may or may not be a key concern – it depends on the test.

Getting just over 50% on a test doesn’t fit most peoples’ definition of “doing well”, even if it is above average
Pretty much only psychometricians would think this way

So, sometimes you need to sacrifice some psychometric quality to accommodate psychological considerations
Considering a test as a whole, there may be utility in having some very difficult items even if that means they have poor discrimination (this would depend on you choice of split, of course)

49
Q

How can you get around psychological considerations of tests

A

If it’s an option, standardization can help resolve the contrast between good psychometrics and psychological considerations.

In this case, standardization means taking the raw scores, which are likely very low, and converting them into a new score that looks more acceptable

Percentiles would be one way of doing it
IQ scores typically use standardization via Z scores, where a Z of 0 becomes 100 (a psychologically pleasing number) and each Z difference of 1 adds or subtracts 15 points

50
Q

If you have multiple factors in your factor analyses, how should you assess reliability?

A

Run one reliability analysis per factor (e.g. 2 factors = 2 reliability analysis)

51
Q

Contrast differences between PCA and EFA

A

For PCA, covariance matrix, not correlation matrix and there is no real definition of random error

PCA tend to have higher correlations because what EFA would attribute to error (unexplained variance), PCA will attribute to covariance (explained variance)

52
Q

What is the advantage of doing factor analysis rather than just averaging out correlations

A

Factor analyses can account for error and use “true score” better, allowing us to have more precise models

53
Q

What do we need for good measurements

A

We need good theories
We need to capture all our dimensions well
We need reliability
We need convergent validity (and ideally also discriminant validity)

54
Q

Describe the issue of content validity with the depression scales

A

There are 280 different measures of depression.
The most cited scale for depression is the CES-D
The CES-D has 20 items and 1/3 of them do not appear in any other commonly used measure of depression
Therefore, all 279 other research teams did not agree that a third of the questions should be measuring the construct of depression
this could led us to believe this scale measures depression and another component
Could lead to p-hacking, where one chooses the scale of depression that fits their hypothesis and reject their ones that don’t to demonstrate significance

55
Q

Describe the issue of reliability with the depression scales

A

The situation isn’t any better when looking at reliability for depression.
Inter-rater reliability of major depression diagnoses was .28 (presumably this is an r)
Cronbach’s alpha has many limitations, but in most cases is the only statistic ever used to demonstrate reliability
At least 20% of studies don’t even report alpha
Some report other people’s alpha, not in relation to their own sample

Disconnect between construct validity and reliability, while they should go together

56
Q

How is removing items from a scale bad for validity

A

Adding or removing items is a common practice, without providing a clear reason for doing so
Changing items isn’t necessarily a bad thing; the lack of justification is
Many studies don’t try to demonstrate validity, and many that do simply provide a citation to some other study