Assessment and Statistics Flashcards
All key terms from the Assessment in Counselling textbook.
Accessibility
The notion that all examinees should have an unobstructed opportunity to demonstrate their capabilities on the construct(s) being assessed.
Accommodations
An action taken in response to an individual with a disability in which there is a departure from the standard testing protocol in an attempt to adjust for the disability.
Acculturation
A process of change and adaption that individuals undergo as a result of contact with another culture.
Achievement test
An assessment in which the person has “achieved” knowledge, information, or skills through instruction, training, or experience. Achievement tests measure acquired knowledge and do not make any predictions about the future
Adaptation
The change in original instrument in terms of design or administration to increase the accessibility to individuals (e.g., visually impaired, limited English proficiency).
Affective instrument
An instrument that assesses interest, attitudes, values, motives, temperaments, and the noncognitive aspects of personality.
Age or grade equivalent scores
Scores used to compare individuals with other individuals at the same age that are calculated by item response theory or by using a norm-referenced approach
Alternate or parallel forms
Two forms of an instrument that can be correlated, resulting in an estimate of reliability.
Analogue observation
In this type of observation, the counselor creates a simulated environment that is reflective of the client’s natural environment
Appraisal
Another term for assessment.
Aptitude test
A test that provides a prediction about the individual’s future performance or ability to learn based on his or her performance on the test. Aptitude tests often predict either future academic or vocational/ career performance.
Assessment
A procedure for gathering client information that is used to facilitate clinical decisions, provide clients with information, or for evaluative purposes.
Authentic Assessment
Performance assessments that involve the performance of “real” or authentic applications rather than proxies or estimators of actual learning.
Behavioral Assessment
An assessment method in which the focus is typically on observing and recording of the precise behaviors of the client
Bias Testing
A term that refers to the degree to which construct-irrelevant factors systematically affect a specific group’s performance.
Cluster Sampling
A technique that involves using existing units or cluster rather than selecting individuals.
Coaching
It involves longer training or practice on questions that are the same or similar to the items on the test.
Coefficient of Determination
This statistic estimates the percent of shared variance between two sets of variables that have been correlated. The coefficient of determination (r2) is calculated by squaring the correlation coefficient.
Computer Adaptive Testing (CAT)
A testing in which the computer adapts the next question for the student based on his or her response
Concurrent Validity
A type of validation evidence in which there is no delay between the time the instrument is administered and the time the criterion information is gathered.
Conditional Standard Errors of Measurement
Type of standard error of measurement that takes into account the different score levels.
Construct Underrepresentation
The degree to which the instrument is unable to capture significant aspects of the construct.
Construct Irrelevance
The degree to which scores or results are affected by that are extraneous to the instrument’s intended purpose.
Construct Validity
One of the three traditional forms of validity that is broader than either content or criterion-related validity. Many experts in assessment now argue that evidence of construct validity, which includes the other traditional forms of validity, applies in all types of psychological and educational assessment. This type of validation involves the gradual accumulation of evidence. Evidence of construct validity is concerned with the extent to which the instrument measures some psychological trait or construct and how the results can be interpreted.
Content-Related Validity
One of the three traditional forms of validity in which the focus was on whether the instrument’s content adequately represented the domain being assessed. Evidence of content-related validity is often reflected in the steps the authors used in developing the instrument
Convergent Evidence
Validation evidence that indicates the measure is positively related with other measures of construct
Correlation Coefficient
A statistic that provides an indication of the degree to which two sets of scores are related. A correlation coefficient (r) can range from —1.00 to +1.00 and, thus, provides an indicator of both the strength and direction of the relationship. A correlation of +1.00 represents a perfect positive relationship; a correlation of —1.00 represents a perfect negative or inverse relationship. A correlation coefficient of .00 indicates the absence of a relationship.
Correlation Method
A statistical tool often used in providing validation evidence related to an instrument’s relationship with other variables.
Criterion-Referenced Instrument
Instruments designed to compare an individual’s performance to a stated criterion or standard. Often criterion-referenced instruments provide information on specific knowledge or skills and on whether the individual has “mastered” that knowledge or skill. The focus is on what the person knows rather than how he or she compares to other people
Criterion-Related Validity
One of the three traditional types of validity in which the focus is the extent to which the instrument confirms (concurrent validity) or predicts to (predictive validity) a criterion measure.
Cronbach’s Alpha or Coefficient Alpha
It is one of the methods of estimating reliability through the examination of the internal consistency of the instrument. This method is appropriate when the instrument is not dichotomously scored, such
as an instrument that uses a Likert scale.
Decision Theory
A method that examines the relationship between an instrument and a criterion or predictor variable, which usually involves an expectancy table. Expectancy tables frequently are used to determine cutoff scores or to provide clients with information regarding the probability of a certain performance on the criterion that is based on scores on the assessment.
Differential Item Functioning (DIF)
A set of statistical methods for investigating item bias that examines differences in performance among individuals who are equal in ability but are from different groups (e.g., different ethnic groups).
Discriminant Evidence
Validation evidence that indicates the measure is not related to measures of different psychological constructs.
Domain Sampling Theory
Another term for generalizability theory.
Duty to Warn
The requirement or permission for mental health practitioners to disclose information about a client when that client is going to harm someone else.
Event Recording
One of the methods used in behavioral assessment, where the counselor records the number of times a target behavior or behaviors occur during a specified time period.
Expectancy Table
A method of providing validity evidence that involves charting performance on the criterion based on the instrument’s score. It is often used to predict who would be expected to fall in a certain criterion category (e.g., who is likely to succeed in graduate school) and to determine cutoff scores.
Factor Analysis
A term that covers various statistical techniques that are used to study the patterns of relationship among variables with the goal of explaining the common underlying dimensions (factors). In assessment, factor analysis is often used to examine if the intended internal structure of an instrument is reflected mathematically. For example, a researcher would analyze whether all items on each subscale “load” with the other items on the appropriate subscale (factor) and not with another factor.
False Negative
In decision theory, a term used to describe when the assessment procedure is incorrect in predicting a negative outcome on the criterion.
False Positive
In decision theory, a term used to describe when the assessment procedure is incorrect in predicting a positive outcome on the criterion.
Formative Evaluation
A continuous or intermediate evaluation typically performed to examine the counseling services process
Frequency Distribution
A chart that summarizes the scores on an instrument and the frequency or number of people receiving that score. Scores are often grouped into intervals to provide an easy-to-understand chart that summarizes overall performance.
Frequency Polygon
A graphic representation of the frequency of scores. The number or frequency of individuals receiving a score or falling within an interval of scores is plotted with points that are connected by straight lines
General Ability Test
Another term for intelligence test.
Generalization Theory
An alternative model to the true score model of reliability. The focus of this theory is on estimating the extent to which specific sources of variation under defined conditions influence scores on an instrument.
Grade Equivalent Norms
Norms that are typically used in achievement tests and provide scores in terms of grade equivalents. In some instruments, grade equiv- alent scores are not validated on each specific grade but are extrapolated scores based on group performance at each grade level
High-Stales Testing
A type of testing where the outcome of such tests have significant consequences (e.g., high school graduation examination),
Histogram
A graphic representation of the frequency of scores in which columns are utilized
Hypomanic Episode
Characterized by similar symptoms as a manic episode except it lasts for only four days
Individualized Education Plan (IEP)
An educational plan that is developed for each student who is receiving special education and related services. The plan is developed by a team of educators and the child’s parents or guardians.
Instrument
An assessment tool that typically is not related to grading. In this book, instruments include tests, scales, checklists, and inventories.
Intelligence Tests
Instruments that are designed to measure the mental capabilities of an individual. These assessments are also referred to as general ability tests.
Intercepts
In regression, it is the constant in the regression equation or it is where the line crosses the y-axis.