Item Response Theory Flashcards
How does IRT modeling differ from Rasch modeling?
- IRT modeling seeks to fit a model to the data (model fitness) whereas Rasch modeling seeks to fit data to a model (model parsimony).
- Data that does not fit a Rasch model is discarded.
- Rasch modeling does not permit abilities to be estimated for extreme items and persons
What are the ____ advantages of IRT over CTT?
fdsfd
What is the purpose of IRT?
To relate item characteristics (item parameters) to individual characteristics (latent traits) to create a model that is used to predict the probability of a particular response to an item.
Please list some of the limitations of CTT.
- It is impossible to separate items and examinees
- *insert here*
- You are not able to predict the likelihood of a correct response to an individual item
- You are unable to solve many important practical testing problems
- Test fairness analysis
- Computerized adaptive testing
- Equating test scores
- Designing tests
- Dealing with missing data
What does theta (Θ) represent in IRT?
It represents the amount of a latent trait that an individual possesses.
What are two ways of describing what P(Θ) represents?
- It is the probability of answering item *j *correctly for a randomly sampled examinee with Θ.
- It is also the frequency of correct answers for a given examinee with Θ after administering item j repeatedly.
What are three other ways to depict P(Θ)?
- Expected score conditional on Θ (only for dichotomous items)
- Regression of score on ability Θ (only for dichotomous items)
- P(Θ) is a probability and also a conditional expectation.
What are the three steps that must be taken when creating an adaptive test?
- Where should examinees begin (e.g., at what difficulty level should the first items be set at)?
- What should we them give next (e.g., after getting the first set of items correct/incorrect, how difficult should the next set of items be)?
- When should we have them stop (e.g., how many item sets do we have them complete before we are satsified)?
What is the probability of a “correct response” to a dichotomous item?
P(Θ)
What is the probability of an “incorrect response” to a dichotomous item?
1 - P(Θ) = Q(Θ)
What is a polytomous item?
An item that has a different score for each possible response option.
E.g., Likert scale items or political affiliation
What is a dichotomous item?
An item that has only two possible score values (regardless of how many response options are available)
E.g., Correct vs. Incorrect
What is the purpose of an item response function (IRF)/item category response function (ICRF)?
To graphically represent (a) the probability of an examinee answering an item correctly, or (b) the probability of an examinee endorsing a particular response category.
Why can the IRT assumption of unidimensionality never be truly met?
It can never be met because there are always other latent traits that are responsible for a given response (e.g., motivation, reading ability, etc.) for any testing instrument.
What is another name for the assumption of local independence?
Assumption of conditional independence.
If the assumption of unidimensionality is true, is the assumption of local independence also true?
Yes!
If the assumption of local independence is true, is the assumption of unidimensionality true?
Not necessarily. For example, a model that include multiple latent traits could account for the complete latent space (satisfying the assumption of local independence) but would not satisfy the assumption of unidimensionality.
What does the item parameter *b *represent?
*b *= an item’s location/difficulty
Location = the amount of the latent trait that is needed to have a 0.5 probability of endorsing the item.
What does the item parameter a represent?
a = an item’s discrimination/slope
Discrimination = how strongly related the item is to the latent trait like loadings in a factor analysis
It also is proportional to the slope of the ICC at the point bi (location) on the ability scale (which is the maximum slope of the IRF)
What does a c item parameter represent?
c = guessing
Guessing = is included when respondents very low on a trait may still choose the correct answer
This is mostly used with multiple choice testing and should not vary excessively from the reciprocal of the # of response options (e.g., 0.25 would be expected if there are 4 response options).
What does a *d *item parameter represent?
d = upper asymptote
Upper asymptote = included when respondents very high on the latent trait are not guaranteed (i.e., have less than 1 probability) to endorse the item.
What is the typical range of *b *values?
-3 to +3
If a b parameter is adjusted (made bigger or smaller), while holding all other parameters constant, which direction does it move ICC?
If adjusted to be smaller then it moves the ICC left.
If adjusted to be larger then it moves the ICC right.
Are the terms item response function (IRF) and item characteristic curve (ICC) interchangeable?
Yes!