Quiz 2 Flashcards
Five principles of assessment
- It is thorough 
- It uses a variety of assessment methods
- It is valid
- It is reliable
- It is tailored to the individual client
Possible essay question about five principles of assessment
A good assessment must have foundational integrity. There are five principles we can use to test if an assessment is a good measuring tool. First, a good assessment is thorough. It should have as much relevant information as possible so we can come to a correct diagnosis and meaningful recommendations for therapy can be made. Second, a good assessment uses a variety of assessment tools. It should consist of interviews, case history information, client observations along with formal and informal testing. Third, a good assessment is valid. This means it evaluates the intended skills. Forth, a good assessment is reliable. Repeated administration of the same test to the same child should yield the same results. Finally, a good assessment is tailored to the individual client. It should be appropriate for the clients age, gender, skill level, and cultural background.
Norm-reference tests
They are always standardized. They compare an individual’s performance to the performance of a larger group. This type of test is preferred for school districts and insurance companies for third-party payment and qualification purposes.
Mean
Determines the peak of the bell scale and it represents the average performance.
(In a perfect distribution, the peak also depicts the median, which is the middle of the distribution And the mode, which is the most frequently occurring score.)
Empirical Rule
For a normal bell curve the empirical rule states that:
68% of all outcomes fall within one standard deviation of the mean (34% on each side)
95% of all outcomes will fall within to standard deviation of the mean (47.5% on each side)
99.7% of all outcomes will fall within three standard deviation of the mean (49.85% on each side)
Mode
Most frequently occurring score
Median
The middle of distribution.
In a perfect distribution, the peak also the pics the median.
Standard deviation
A measurement used to quantify the variability of data dispersion, In a set of given values around the mean.
Criterion-referenced tests
May or may not be standardized.
Don’t attempt to compare an individuals performance to anyone else’s.
Identifies what client can and can’t do compared to a predefined criterion.
How does my client compared to an expected level of performance?
Assumes that there is a level of performanceThat must be met for a behavior to be acceptable
**Use most often when assessing a client for neurologic disorders, fluency invoice disorders
Authentic assessment
These are non-traditional assessments.
Identifies with a client can and can’t deal
Emphasis on contextualized test stimuli
it’s ongoing. Information is maintained in a client portfolio. Provides an opportunity for the client to self-monitor, self evaluate.
Systemic observations social skills observed during lunch, recess.
Real-life situation, Language sampling, structured symbolic play
Self-monitoring self-assessment. Use of anecdotal notes in checklists video and audio taping and involvement of caregiver and other professionals.
Dynamic assessment
This is a form of authentic assessment.
Purpose is to evaluate a learning potential Based on his or her ability to modify responses after the clinician provided teaching or other assistance. (Test-teach-retest)
Great for kids with cognitive communicative disorders And kids from culturally and linguistically with that verse background
Mediated learning experience (MLE)
The clinician teaches a strategies specific to the skill being evaluated, observing the clients response to instruction and adjusting teaching accordingly.
Test-teach-retest
Validity
Test truly measures what it claims to measure.
Validity types
Face validity
Content validity
Construct validity
Criterion validity
Concurrent validity
Predictive validity
Face validity
Test looks like it assesses the skills it claims to assess. Face validity alone is not a valuable measure Of validity because it is based merely on appearance, not on content or outcome.
Content validity
Test contest are representative of the constant domain of a skill being assessed.
A valid articulation test should elicit all phonemes, thereby assessing the spectrum of articulation. Content validity is related to face validity; content validity, though, Judge is the actual content of the text rather than superficial appearance and is judged by individuals with expert knowledge.
Construct validity
Measures a predetermined theoretical content, which is an explanation of a behavior or Attribute based on empirical observation. For example the theoretical content that preschool children’s language skills improve with aged is based on language development studies. Therefore, a valid test of early language development well show improved language skills when administered to normally developed preschool children of progressively increasing age.
Criterion validity
Refers to a validity that is established by use of an external criterion there are two types of criterion validity.
Concurrent validity refers to a test levity in comparison to a widely excepted standard.
For example, the Stanford-Binet Intelligence scale is already excepted as a valid assessment for intelligence. You are intelligent tests are compared to the Stanford-Binet which serves as the criterion measure.
Predictive validity refers to the tests ability to predict performance (the criterion measure) in another situation or at a later time. It implies that there is a known relationship between the behaviors the test measures and the behaviors or skills exhibited at some future time. College entrance exams, such as the graduate were record examination (GRE), are used because of their predictive validity. The GRE scores are expected to predict future academic performance.
Reliability
Results are replicable. When administered properly, the test gives consistent results on repeated administrations or with different interpreters judging the same administration. There are several types of reliability.
Types of reliability
Test-retest reliability Split-half reliability Rater reliability Intra-rather reliability Inter-rather reliability Alternate form reliability/parallel form reliability
Test-retest reliability
Refers to a test stability overtime. It is determined by Administering the same test multiple times to the same group and then comparing the scores. If the scores from the different administrations are the same or very similar, the test is considered stable and reliable.
Split-half reliability
Refers to a tests Internal consistency. Scores from 1/2 of the test correlate with results from the other half of the test. The halves must be comparable in style and scope and all items should assess the same skill. This is often achieved by dividing the test into odd numbered questions and even numbered questions.
Rater reliability
Refers to the level of agreement among individuals rating a test. It is determined by administering a single test and audio or video taping it so it can be scored multiple times there are two types of greater reliability.
Intra-rater reliability is established if results are consistent when the same person reads the test on more than one occasion.
Inter-rater reliability Is established if results are consistent when more than one person rate the test

Alternate form reliability Also called parallel form reliability
Refers to a test correlation coefficient with a similar test. It is determined by administering a test (test A) To a group of people and then administering a parallel form of the test (test B) To the same group of people. The two sets of test results are compared to determine the tests alternate form reliability.