Creating a Test Flashcards
Give examples of some of the issues to be considered when conceptualizing a test
Consider what’s been done before and if the proposed test offers anything new; who’s going to use it; context in which it will be used; how long or effortful the measure can be; any practicalities to be considered; any ethical issues to address
Give an example of the sort of practicalities that might need to be considered when conceptualizing a test
Large enough font size if using older adults; Size of budget and amount of time; expense if administered to thousands of people, and skill of examiners
What sort of ethical issues might you need to consider when conceptualizing a test?
Anonymity and whether this will create problems; whether there’s any potentially sensitive or offensive content
What sort of issues should you consider when creating the materials for a new test?
Format of test items; scoring of test; whether items yield sufficient response variability (e.g. if norm-referenced); whether the items sample the domain of interest sufficiently; if items test people on the criterion; number of items (more items = more reliable but takes longer); whether you want/have face validity
What key considerations might you need to take into account when drafting written content for a test?
Avoid ambiguity and lack of clarity; include reversed questions
If someone asks you advice on designing a good multiple-choice question, what would you tell them?
Make sure distractors are plausible; reduce the chance of a respondent guessing the correct answer without actually knowing it
What are selected-response questions?
They test recognition; e.g. multiple-choice questions; matching options; true/false
What are constructed-response questions?
Where respondents generate the answer (recall); e.g. fill in the blank; write an essay/account
In a multiple-choice question, what is the “stem”?
The question part
Describe five strategies you could use for pilot research into your new test
- Examine previous literature;
- Interview people who might know something about the area;
- Give people a more open-ended form of the test to generate ideas for items;
- Get people to complete a preliminary or “faked-up” version of the test and comment on it;
- Get people to give a running commentary about their thoughts as they complete a draft of the test
List five strategies you could use for evaluating the quality of items in an achievement test
Item difficulty index; Reliability item discrimination index; Validity item discrimination index; If multiple choice, examine pattern of responses across options; Further in-depth examination
What’s the item difficulty index?
The proportion of respondents who get the item correct
What the validity item discrimination index?
The item’s contribution to the scale’s relationship to an external criterion measure
When examining the quality of items, what further in-depth questions need to be asked?
Are there items where respondents expected to get it correct get it wrong?; Are there items where respondents expected to get it wrong are scoring above chance?
Describe the four strategies you could use to evaluate item quality in a personality-type test
Examine histogram/frequency table of responses to each item to determine the spread; reliability item discrimination index; validity item discrimination index; further in-depth examination (expectations of high vs. low scorers)