Psychological Testing & Assessment; Assumptions & Norms Flashcards
The gathering and integration of Psychology-related data for the purpose of making a psychological evaluation that is accomplished through the use of tools.
Psychological Assessment
The process of measuring Psychology related variables by means of devices or procedures designed to obtain a sample of behavior.
Psychological Testing
Its objective is typically to answer a referral question, solve a problem or arrive at a decision through the use of tools of evaluation.
Psychological Assessment
In psychological assessment, the _ is the key to the process of selecting tests and other tools of evaluation as well as in drawing the conclusions.
Assessor
What is the typical outcome of Psychological testing?
Test scores
What are the 2 different approaches to assessment?
Collaborative Psychological Assessment
Dynamic Assessment
In this approach, the assessor and assessee may work as partners from the initial contact through final feedback.
Collaborative Psychological Assessment
An interactive approach to Psychological Assessment that usually follows a model of: evaluation> intervention of some sort> evaluation. This provides a means for evaluating how the assessee processes or benefits from some type of intervention.
Dynamic Assessment
A measuring device or procedure.
Test
Psychological tests almost always involve analysis of _.
Sample of behavior
The subject matter of the test.
Content
Form, plan, structure, arrangement and layout of test items as well as related considerations. It also refers to the form in which a test is administered.
Format
Demonstration of various kinds of tasks demanded of the assessee, as well as trained observation of an assessee’s performance.
Administration procedures
For tests that are designed for administration on _ may require an active and knowledgeable test administrator.
One-to-one basis
The process of assigning such evaluative codes or statements to performance on tests, tasks, interviews or some other behavior samples.
Scoring
Most tests of intelligence come with __, that are explicit about scoring criteria and the nature of interpretations.
Test manuals
Refers to how consistently, how accurately a psychological test measures what it purports to measure, the usefulness or practical value that a test or other tool of assessment has for a particular purpose.
Psychometric soundness
The method of gathering information through direct communication involving reciprocal exchange.
Interview
Samples of one’s ability and accomplishment.
Portfolio
Refers to records, transcripts and other accounts in written, pictorial or other form that preserve archival information, official and informal accounts and other data and items relevant to an assessee.
Case History Data
A report or illustrative account concerning a person or an event that was compiled on the basis of case history data.
Case study
Monitoring the actions of others or oneself by visual or electronic means while recording quantitative and/or qualitative information regarding those actions.
Behavioral observation
Observe behavior of humans in natural settings in which the behavior would typically be expected to occur.
Naturalistic Observation
A tool of assessment wherein assessees are directed to act as if they were in a particular situation. Assessees may then be evaluated with regard to their expressed thoughts, behaviors, abilities and other variables.
Role-Play Tests
It can serve as test administrators and as highly efficient test scorers.
Computers
What are the different types of scoring reports?
Simple scoring report
Extended scoring report
Interpretive report
Consultative report
Integrative report
A scoring report that includes statistical analysis of the testtaker’s performance.
Extended scoring report
A scoring report that includes numerical or narrative interpretive statements in the report. Some of it contain relatively little interpretation and simply call attention to certain high, low or unusual scores that need to be focused on.
Interpretive report
A scoring report that is usually written in language appropriate for communication between assessment professionals and may provide expert opinion concerning analysis of the data.
Consultative report
A scoring report that integrates data from the sources other than the test itself into the interpretive report.
Integrative report
Create tests or other methods of assessment.
Test developer
Psychological tests and assessment methodologies are used by a wide range of professionals. These are called _.
Test user
Anyone who is the subject of an assessment or an evaluation.
Testtaker
Reconstruction of a deceased individual’s psychological profile on the basis of archival records, artifacts and interviews previously conducted with the deceased or people who knew him or her.
Psychological autopsy
Test that evaluates accomplishment or the degree of learning that has taken place.
Achievement test
Tool of assessment used to help narrow down and identify areas of deficit to be targeted for intervention.
Diagnostic test
Nonsystematic assessment that leads to the formation of an opinion or attitude.
Informal evaluation
In this setting, tests are mandated early in school life to help identify children who may have special needs.
Educational settings
In this setting, tests and many other tools of assessment are used to help screen for or diagnose behavior problems.
Clinical settings
Group testing in clinical settings is primarily used for _ - identifying those individuals who require further diagnostic evaluation.
Screening
In this setting, the ultimate objective of many such assessments is the improvement of the assessee in terms of adjustment, productivity or some related variables.
Counseling setting
In these settings, a wide range of achievement, aptitude, interest, motivational and other tests may be employed in the decision to hire as well as in related decisions regarding promotion, transfer, job satisfaction and eligibility for further training.
Business and Military Setting
What is the well known application of measurement in governmental settings?
Governmental licensing certificate
An observable action or the product of an observable action including test-or assessment-related responses.
Overt Behavior
The more the testtaker respond in a particular direction as keyed by the test manual as correct or consistent with a particular trait, the higher that testtaker are presumed to be on the targeted ability or trait.
Cumulative scoring
Understanding of behavior that has already taken place. It is typically the use of psychological tests in forensic matters.
Postdict
Refers to factors other than what a test attempts to measure will influence performance on the test.
Error
Component of a test score attributable to sources other than the trait or ability measured.
Error variance
What are the potential sources of error variance?
Assessee
Assessor
Measuring instruments
The test performance data of a particular group of testtakers that are designed for use as a reference when evaluating or interpreting individual test scores.
Norms
The group of people whose performance on a particular test is analyzed for reference in evaluating the performance of individual testtakers.
Normative sample
The process of deriving norms.
Norming
A method of evaluation and a way of deriving meaning from test scores by evaluating an individual teststaker’s score and comparing it to scores or a group of testtakers.
Norm-referenced testing and assessment
The process of administering a test to a representative sample of testtakers for the purpose of establishing norms.
Standardization
In the process of developing a test, a test developer has targeted some defined group as the population for which the test is designed.
Sampling
The complete universe or set of individuals with atleast one common, observable characteristic.
Population
A portion of the universe of people deemed to be representative of the whole population.
Sample
The process of selecting the portion of that universe deemed to be representative of the whole population.
Sampling
A sampling method where differences with respect to some characteristics of subgroups within a defined population are proportionately represented in the sample. It helps prevent sampling bias and ultimately aid in the interpretation of findings.
Stratified sampling
Arbitrarily selecting some people because it is believed to be representative of the population.
Purposive sampling
People who are most available to participate in the study.
Incidental sample or convenience sample
What are the 6 types of norms?
Percentile
Developmental Norms
National Norms
National Anchor Norms
Subgroup Norms
Local Norms
An expression of the percentage of people whose score on a test falls below a particular raw score.
Percentile
Applied broadly to norms developed on the basis of any trait, ability, skill or other characteristic that is presumed to develop, deteriorate or otherwise be affected by chronological age, school grade or stage of life.
Developmental Norms
A developmental norm that indicates the average performance of different samples of testtakers who were at various ages at the time the test was administered.
Age norms
A developmental norm designed to indicate the average test performance of testtakers in a given school grade.
Grade norms
Derived from a normative sample that was nationally representative of the population at the time the norming study was conducted.
National Norms
A type of norms that provides some stability to test scores by anchoring them to other test scores.
National anchor norms
A normative sample that is segmented by any of the criteria initially used in selecting subjects for sample.
Subgroup norms
A type of norm that provides normative information with respect to the local population’s performance on some test.
Local Norms
The distribution of scores obtained on the test from one group of testtakers is used as the basis for the calculation of test scores for future administration of the test.
Fixed-reference group scoring system
Evaluating the test score in relation to other scores on the same test. The usual area of focus is how an individual performed relative to other people who took the test.
Norm-referenced
A method of evaluation and a way of deriving meaning from test scores by evaluating an individual’s score with reference to a set standard.
Criterion referenced testing and assessment
It is an index of reliability, a proportion that indicates the ratio between the true score variance on a test and the total variance.
Reliability coefficient
The degree of the relationship between various forms of a test that can be evaluated by means of an alternate-forms or parallel-forms coefficient of reliability.
Coefficient of Equivalence
The means and the variances of the observed test scores are equal.
Parallel forms
The estimate of the extent to which item sampling and other errors have affected test scores on versions of the same test when, for each form of the test, the means and variances of observed test scores are equal.
Parallel Forms reliability
Typically designed to be equivalent with respect to variables such as content and level of difficulty.
Alternate forms
Refers to an estimate of the extent to which these different forms of the same test have been affected item sampling error of other error
Alternate forms reliability
It refers to the degree of correlation among all the items on a scale. Calculated from a single administration of a single form of a test.
Internal consistency of reliability
An index of _ consistency is useful in assessing the homogeneity of the test.
Inter-item consistency
The statistic of choice for determining the inter-item consistency or dichotomous items, primarily those items that can be scored right or wrong.
Kuder-Richardson Formula 20 or KR-20
It is appropriate for use on tests containing no dichotomous items.
Coefficient alpha
Because negative values of alpha are theoretically impossible, it is recommended under such circumstances that the alpha coefficient be reported as _.
Zero
It is a measure used to evaluate the internal consistency of a test that focuses on the degree of difference that exist between item scores.
Average proportional distance
The higher the reliability of a test, the _ the SEM/SEE.
Lower
Yield insights regarding a particular population of test takers as compared to the norming sample described in a test manual.
Local validation studies
It describes a judgement of how adequately a test samples behavior representative of the universe of behavior that the test was designed to sample.
Content validity
He developed a formula termed content validity ratio.
C.H. Lawshe
A method for gauging agreement among raters or judges regarding how essential a particular item is.
Quantification of content validity
When fewer than half the panelists indicate “essential” in content validity ratio.
Negative CVR
When exactly half the panelists indicate “essential” in content validity ratio.
Zero CVR
When more than half the panelists indicate “essential” in content validity ratio.
Positive CVR
The content validity ratio ranges between _.
- and .99
The standard against which a test or a test score is evaluated.
Criterion
A judgment of how adequately a test score can be used to infer an individual’s most probable standing of some measure of interest.
Criterion- related validity
What are the 3 characteristics of a criterion
Relevant
Valid
Uncontaminated
The term applied to a criterion measure that has been based, atleast in part, on predictor measures.
Criterion contamination
It is a type of criterion-related validity that indicates the extent to which test scores may be used to estimate an individual’s present standing on a criterion.
Concurrent validity
It is a type of criterion-related validity. It is the measure of the relationship between the test scores and a criterion measure obtained at a future time.
Predictive validity
Judgments of criterion-related validity are based on 2 types of statistical evidence:
Validity coefficient
Expectancy data
A correlation coefficient that provides a measure of the relationship between test scores and scores on the criterion measure.
Validity coefficient
The degree to which an additional predictor explains something about the criterion measure that is not explained by predictors in use.
Incremental validity
Table that illustrate the likelihood that the test takers will score within some interval of scores on a criterion measure-an interval that may be seen as “passing” or “acceptable”.
Expectancy table
A graphic representation of an expectancy table.
Expectancy chart
A judgment about the appropriateness of inferences drawn from test scores regarding the individual standings on a variable called a construct.
Construct Validity
An informed, scientific idea developed or hypothesized to describe or explain behavior. Unobservable, presupposed traits that a test developer may invoke to describe test behavior or criterion performance.
Construct
Viewed as the unifying concept for all validity evidence.
Construct validity
It refers to how uniform a test is in measuring a single concept.
Homogeneity
It may be used in estimating the homogeneity of a test composed of multiple-choice items.
Coefficient alpha
What are the evidence of construct validity?
Homogeneity
Changes with age
Pretest-posttest changes
Distinct groups
Convergent evidence
Discriminant evidence
Factor Analysis
A validity coefficient showing little relationship between test scores and/or other variables with which scores on the test being construct validated should not theoretically be correlated. This provides what kind of evidence?
Discriminant evidence
An experimental technique useful for examining both convergent and discriminant validity. It is the matrix or table that results from correlating variables (traits) within and between methods.
Multitrait-multimethod matrix
It is designed to identify factors or specific variables that are typically attributes, characteristics or dimensions on which people may differ. It is employed as a data reduction method in which several sets of scores and the correlations between them are analyzed.
Factor analysis
Factor analysis is conducted on three basis:
Exploratory factor analysis
Confirmatory factor analysis
Factor loading
A factor analysis that typically entails estimating or extracting factors, deciding how many factors to retain, and rotating factors to an interpretable orientation.
Exploratory factor analysis
Factor analysis wherein the researchers test the degree to which a hypothetical model includes factors that fit the actual data.
Confirmatory factor analysis
It conveys information about the extent to which the favor determines the test scores.
Factor loading
High factor loadings would provide _ evidence of construct validity.
Convergent
Moderate to low factor loadings would provide _ evidence of construct validity.
Discriminant
A factor inherent in a test that systematically prevents accurate, impartial measurement.
Bias
A numerical or verbal judgment that places a person or an attribute along a continuum identified by a scale of numerical word descriptions known as rating scale.
Rating
A judgment resulting from the intentional or unintentional misuse of a rating scale.
Rating error
3 types of rating errors:
Leiniency or Generosity error
Severity error
Central Tendency Error
One way to overcome restriction of range errors. It is a procedure that requires the rater to measure individuals against one another instead of against an absolute scale.
Ranking
The extent to which a test is used in an impartial, just and equitable way.
Fairness