h&m chap 5 Flashcards
Predictor
Any variable used to forecast a criterion
Psychometric
The measurement (“metric”) of properties of the mind (from the Greek word “psyche”). The standards used to measure the quality of psychological assessments
Reliability
A standard for evaluating tests that refers to the consistency, stability, or equivalence of test scores. Often contrasted with validity
Test-Retest Reliability
A type of reliability that reveals the stability of test scores upon repeated applications of the test
coefficient of stability
reflects the stability of the test over time
On a reliable test, do those with high and low scores stay consistent or bounce around between two trials? What happens on an unreliable test
If the test is reliable, those who scored high the first time will also score high the second time, and those who scored low on the first will also score low on the second. If the test is unreliable, the scores will “bounce around” in such a way that there is no similarity in individuals’ scores between the two trials
At what number is a reliability coefficient considered professionally acceptable?
0.70 (although 0.80 and above are better)
Equivalent-form reliability
A type of reliability that reveals the equivalence of test scores between two versions or forms of the test
coefficient of equivalence
reflects the extent to which the two forms are sufficiently comparable measures of the same concept
What is the least popular major type of reliability and why?
Of the three major types of reliability, test-retest reliability is the least popular because it is usually challenging to come up with one good test, let alone two
If the resulting coefficient of equivalence between two tests is high, what does that mean for their reliability?
If the resulting coefficient of equivalence is high between two tests, the tests are sufficiently comparable and are viewed as reliable measures of the same concept
Internal Consistency Reliability
A type of reliability that reveals the homogeneity of the items comprising a test
Split-half reliability
a test is given to a group of people, and when it is time to score the test, the researcher divides the items in half
If a test has internal-consistency reliability, what will there be between responses to the two half-tests?
If the test has internal-consistency reliability, there will be a high degree of similarity between the responses to the items from the two halves
Inter-rater reliability (aka inter-judge, inter-observer, or concept reliability)
A type of reliability that reveals the degree of agreement among the assessments provided by two or more raters
Validity
A standard for evaluating tests that refers to the accuracy or appropriateness of drawing inferences from test scores. Often contrasted with reliability
Operationalization
The process of determining how a construct will be assessed
Construct Validity
The degree to which a test is an accurate and faithful measure of the construct it purports to measure
convergent validity coefficients
reflect the degree to which these scores converge (or come together) in assessing a common concept
divergent validity coefficients
reflect the degree to which these scores diverge (or are separate) from each other in assessing unrelated concepts
Criterion-related validity
The degree to which a test forecasts or is statistically related to a criterion
Concurrent validity
used to diagnose the existing status of some criterion, whereas predictive validity is used to forecast future status
What kind of validity focuses on how well a predictor measures the criterion at the same point in time?
In measuring concurrent criterion-related validity, we are concerned with how well a predictor can predict a criterion at the same time, or concurrently.
In what kind of validity do we collect predictor information to forecast future criterion performance?
In measuring predictive criterion-related validity, we collect predictor information and use it to forecast future criterion performance
Validity Coefficient
A statistical index (expressed as a correlation coefficient) that reveals the degree of association between two variables
What does a greater correlation between the predictor and the criterion tell us?
Tell us more about the criterion based on the predictor
Content Validity
The degree to which subject matter experts agree that the items in a test are a representative sample of the domain of knowledge the test purports to measure
Which form of validity was most relevant in achievement testing?
Content