Study Group - Evaluation & Research Flashcards
What are nonintervention costs of programs?
Resource costs that are part of intervention
What are the parameters for addressing utilization of findings?
- Study design
- Preparation
- Possible feedback
- Follow through
- Information distribution
- Any possible further uses of information
Attributes for Evaluation Recommendations
- Defensible
- Timely
- Realistic
- Targeted
- Simple
- Specific
What are the cost expenditure categories in CEA?
- Developmental
- Production
- Implementation
- Evaluation
Cost-Minimization Analysis (CMA)
type of CEA which program A & program B have identical outcomes
How does cost-effective analysis help decision makers?
allocation of limited resources & still achieve desired health benefits
Where are evaluation indicators created from?
Logic Model
What is Accuracy (evaluation standard)?
Provide accurate information for determining merits of program
Construct Validity
Whether specific measure of concept is associated with 1+ measures that is consistent with theoretically derived hypotheses
*- How accurately inferences about specific features of program reflect constructs
- Underlying theory is correct
Meta-Analysis
Quantitative technique for combing results from multiple, different evaluations on same topic
- Could provide information as to whether findings are strong over variations of populations,, settings, programs & outcomes
What are intervention costs of programs?
All resources used in delivery of intervention
Translational Research
Studying & understanding progression of “bench-to-bedside-to-population”
- How scientific discoveries lead to effectiveness & efficacy in studies which lead to dissemination into practice
_____________ & _____________ aid in program cost effectiveness (in addition to types of analyses).
- Cost effectiveness ratio
- Value threshold
Types of Evidence (most to least rigorous)
- Systematic reviews & meta-analysis
- Scientific literature
- Public health surveillance data
- Program evaluations
- Reports from community members & other stakeholders (e.g. needs assessment)
What does Implementation Science identify?
Identifies factors, processes, & methods that increase likelihood of evidence-based interventions to be adopted & used to sustain improvement in population health
What is external use?
Benefits decision makers & administrators not connected with program by considering program in different setting & how to change similar program that is not performing well
How data is managed is dependent on what?
- Type of data
- How data is collected
- How data is used throughout project lifestyle
Ordinal measurement & give an example
Provides information based on order, sequence, or rank
- scale from strongly disagree to strongly agree
Impact Evaluation
- Focuses on ultimate goal, product, or policy
- Often measured in terms of HEALTH STATUS, MORBIDITY, & MORTALITY
How long is data stored according to security rule?
5-10 years
What is the difference between qualitative & quantitative data?
Qualitative - describes what is occurring or why is occurring (non-numerically)
Quantitative - Numerical data that describes what is happening
What does data screening (found in data analysis plan) allow/tell evaluator/researcher?
- Assesses accuracy of data entry
- How outliers & missing values will be handled
- If statistical assumptions are met
What processes can quantitative & qualitative data be useful?
- Program planning
- Implementation
- Evaluation
Guidelines for Developing Recommendations
- Invest time
- Start early
- Consider all issues as fair game
- Case wide net
- Work closely with decision makers & program staff
- Decide whether recommendations should be as general or specific as possible
- Consider program context
- Consider program closure
- Describe expected benefits/costs
- Decide whether change should be incremental vs fundamental
- Avoid recommending another evaluation
What is feasibility (evaluation standard)?
Conduct evaluations that are VIABLE & REASONABLE
What are delimitations?
Parameters or boundaries placed on study by researchers that help manage scope of study
What is a purpose statement?
- Tool to identify what is to be learned from evaluation and/or research
- Serves to focus & steer collection & analysis of data
Longitudinal Design
data about program collected at 2+ POINTS IN TIME
Hawthorne Effect
If people in intervention group become sensitive to repeated introduction/removal of intervention
- Threat to Internal Validity
Implementation Evaluation
Comprehensive, retrospective determination of extent to which program was delivered as designed & whether variations may have held significant implications for effects of program
- AKA Process evaluation *
What is Instrumental Use?
Decision makers change program (expand to other sites, terminate, change how it is implemented) based on answers to evaluation questions
Steps of Effective Evaluation
- Defining research population
- Identifying stakeholders & collaborators
- Defining evaluation objective
- Selecting research design that meets evaluation objective
- Selecting variables for measurement
- Selecting sampling procedure
- Implementing research plan
- Analyzing data
- Communicating findings
Procedural Equity
Maximizing fairness in distribution of services across groups
Descriptive Design vs Explanatory Design
Descriptive: DESCRIBE events, activities, or behavior that occurred (what went on in program)
Explanatory: EXPLAINS events, activities, or behavior that occurred (improve understanding)
What is maturation?
before-after changes due to changes occurring inside people rather than program
Substantive Equity
Minimizing disparities in distribution of health across groups or different populations
What is a type V error?
- Reporting intervention has statistically effect but effect is too small
- No significance to decision makers
What are problematic outliers?
Outliers not representative of population
Bivariable Analysis
Determines whether variables in database are correlated with each other
- Compares 2+ groups to see whether a characteristic is similar/different
- Find out whether program outcomes are significantly different between 2 groups OR 1 group overtime (impact evaluation)
Ratio measurement & give an example
Common measurement between each score & have true zero
- height, weight, age, etc.
What are the different types of evaluation designs?
- one group posttest only
- one group pre- & posttest
- Comparison group posttest only
- two group pre- & posttest
- one group time series
- Multi-group time series
- two group retrospective (case control)
- two group prospective (cohort)
- two group pre- & posttest with random assignment (RCT)
Test-Retest
Same measurement administered at 2 points in time
What should be tested/assessed for when considering using existing data collection instruments? Why?
Literary reading level (using or adapting) to ensure validity of responses
What are intangible benefits?
Non-monetary, subjective, or difficult to measure gains attributable to program intervention
Cost-Effective Analysis (CEA)
Determines differences between 2 programs based on what it costs for delivery of the programs
- Relationship b/w program cost (input) & impact (output)
Internal Reliability/Consistency
Consistency measuring multiple/all items it is meant to measure
What analyses provides information on program cost effectiveness?
- Cost-benefit Analysis (CBA)
- Cost-minimization Analysis (CMA)
- Cost-utility Analysis (CUA)
- Sensitivity Analysis
Evaluations should be __________________
Useful, feasible, ethical, accurate, & accountable
History (threat to internal validity)
before-after changes is due to other factors in the environment rather than the program
IRB
Group of individuals that review potential research proposals that involve human subjects/participants
- Approval must be granted prior to beginning data collection*
Institutional Review Board
Formative Evaluation
Conducted before program begins
- designed to produce data & information used to improve program during developmental phase
- Documents appropriateness & feasibility of program implementation
- ensure fidelity of program
Causal Inference
Intellectual discipline that considers assumptions, study designs, & estimation strategies
- Allows researchers to draw conclusions based on data
What data should be demonstrated/included when using it for policy analysis?
- Burden of health of public
- Priority over other issues
- Pertinence at local level
- Interventional benefits
- Personalization of issue by using stories about how lives as impacted
- Estimated intervention costs
What are the characteristics of indicators to ensure credibility?
- Clearly linked to intervention outcome
- Presented in specific, measurable terms
- Appropriate for population being served
- Feasible given data collection, resources, & skills
- Valid & reliable to stakeholders
Data collection must be _____________ by decision makers & stakeholders
Relevant
What are the research/evaluation errors HES should be able to identify?
- Sampling errors
- Lack of precision
- Variability of measurement
- Selection bias
- Instrumental bias
- Internal threats to validity
Embedded Design
Either qualitative or quantitative has priority or is more vital for answering main question of evaluation
What is a type III error?
Rejecting program as ineffective when program was never implemented as intended or technology flaws undermined program effectiveness
Summative Evaluation
Evaluation occurs after program has ended
- designed to produce data on program’s efficacy or effectiveness during implementation
- Provides data on extent of achievement of goals regarding learning experience
What should evaluator consider when choosing evaluation design?
- Causality
- Bias
- Retrospective vs Prospective
- Time span
- Finances
- Current political climate
- # of participants
- Type of data being collected
- Data analysis & skills
- Access to group to use for comparative purposes
- Possibility to distinguish b/w exposed & unexposed to program intervention
- Type of outcome being evaluated (unbound vs bound)
Considerations for data collection implementation
- Find reliable, trustworthy, & skilled people to collect, enter, analyze, & manage data
- Define roles, responsibilities, & skills needed
- Monitor data collection
- Maintain integrity of data collected
- Ensure protocols address quality control measures
Dissemination
spreading information widely
- new publications take 17 years to be widely implemented
Interval measurement & give an example
Common unity of measurement with no true zero
- Temperature
What are the advantages to using existing data collection instruments?
- Previously tested for reliability & validity
- Direct comparison measures
- Reduced cost
- User familiarity
Interrater Reliability
Correlation between different observers at same point in time
SWOT Analysis
Assesses internal & external environment
Convergent Design
basic steps in evaluation process implemented independently at same time
What is missing data?
Observations that were intended to be made but were not
What does HIPPA protect?
All information in health records, billing, & conversations among individuals & healthcare providers
Continuous Quality Improvement (CQI)
Tool to reduce costs while improving quality of services
- enhances organizational effectiveness
What are direct costs of programs?
All goods, services, & other resources used to deliver intervention
Internal Validity
Degree program caused change that was measured
- Were changes in participants due to program or by chance?
Population Effectiveness
Improving health of populations & communities through medical and/or non-medical services
What are baseline indicators?
Value of indicator prior to implementation
Sequential Design
Basic steps in evaluation process implemented sequentially (either qualitative or quantitative first)
Why would HES make modifications to content, format, or presentation of question, questionnaire, or instrument?
- Adapting to data needs
- To have results that are more versatile & useful
What does most appropriate data collection instrument depend on?
- Intent of program
- Intent of evaluation
- Information being acquired
What can cause lesser approach (compared to most rigorous available) to be used in either research or evaluation?
- Ethics
- Cost
- Politics
- Availability of resources
What is a type IV error?
- Evaluation is conducted for sake of evaluation
- Questions are asked about program that no one cares about
- Answers are of no interest to decision makers
Steps for Conducting CEA
- Define problem & objectives
- Identify alternatives
- Describe production relationships
- Define perspective/viewpoint of CEA
- Identify, measure, & value cost
- Identify & measure effectiveness
- Discount future costs & effectiveness
- Conduct sensitivity analysis
- Address equity issues
- Use CEA results in decision making
What are target indicators?
Expected value of indicator at a specific point in time
Random Selection
Random identification from intended population of those who will be in program and/or evaluation
Ethical Standards of Participant Data
- Respect for autonomy
- Social justice
- Promotion of good & avoidance of harm
- Have evaluation/research plan that protects privacy of participants
- Participant data must be stored, utilized, & disclosed ensuring protection of participant privacy
Types of Program costs
- Direct costs
- Intervention costs
- Indirect costs
- Nonintervention costs
- Cost savings vs future costs as result of program/implementation
Multi-Phase Design
Evaluations divided into multiple parts/phases that are implemented over time
What do performance measures require?
- Object
- Standard - accepted level of performance expected
- Indicator - determines whether performance standard is achieved
- Measure - quantitative representation of capacity, process, or outcome
What is a type II error?
Inferring program has no impact when it does (occurs when sample size is too small)
External Validity
Generalizability of results beyond participants
- Would results be the same with different target population?
Social Desirability Effect
Bias that occurs when people answer questions in a way they think will make them seem favorable to others
- Threat to internal validity
What is Conceptual Use?
- Evaluations produce new information about what goes on in the program through answers to questions raised about a program
- Reveals insights about program (what they think of the program, understanding the importance of program) in addressing underlying problem
Systems-Analysis Evaluation Model
Uses instruments that serve to quantify program’s effects
Standards/Steps of Evaluation
- Engage stakeholders
- Describe program
- Focus evaluation design
- Gather credible evidence
- Justify conclusions
- Ensure use & share lessons
CBPR
Research in which evaluators collaborate with community members
- Improves likelihood of success & stronger impact with target population
Data Management Plan
- Procedures for transferring data from instruments to data analysis software
- Scoring guide to tell researcher/evaluation team how to code variables
Security Rule
- Establishes rules for safeguarding information
- Requires IRB
- Guidance provided by Code of Federal Regulations
What measurements are used in descriptive data?
Frequency, mean, median, mode
Response Bias
Intentional or unconscious way individuals select responses
Advantages of Meta-Analysis
- Ability to tell if results are more varied that expected
- Derived statistical testing of overall factors/effect size in related studies
- Potential generalization to population of studies
- Ability to control & use moderators to explain variations between studies
Measurement vs Classification
Measurement - process of sorting & assigning #s to people in quantitative evaluations
Classification - assigning people into set of categories in qualitative evaluations
What aspects are included in an effective evaluation report?
- Timely provision
- Effective & detailed summary of how stakeholders were involved
- List of strengths & limitations/weaknesses of findings
Evidence-based approach in findings & scientific evidence need to be incorporated into what areas of health programming?
- Decision making
- Policy development
- Program implementation
What is utility (evaluation standard)?
Ensure information needs of intended users are satisfied
Unbounded vs Bounded Outcomes
Unbounded - possibility of existing before/after program
Bounded - outcomes can only occur once at particular time by a specific event/time
What are limitations?
Boundaries placed on study by factors or people other than researcher
Allocative efficiency
Combining inputs to produce maximum health improvements given available resources
Measurement tools must be ____________
Valid & Reliable
Types of Designs (most to least rigorous)
- Systematic reviews
- RCT
- Cohort
- Case-control
- Case series, case reports
- Editorials, expert opinion
Reliability
Whether results can be measured consistently (can be reproduced under similar circumstances)
Mediation Analysis
Identification of pathway between health promotion program, its impact on hypothesized psychosocial mediators, & its effects on behavioral outcomes
Informal Interviewing
One–ended conversation with goal of understanding program from respondent’s perspective
- Continues until no new information is gathered & there is full understanding of the program
Types of Program Benefits
- Tangible benefits
- Intangible benefits
- Economic
- Personal health
- Social
Discriminant Validity
measure of different concepts correlated to each other
- type of construct validity
What is a sensitivity analysis?
systematic approach for determining whether CEA yields same results if different assumptions are made
What is construct confounding in regard to construct validity?
Failure to define all constructs may result in incomplete construct inferences or confusion among constructs
What are evaluation questions designed to do?
- Designated boundaries for evaluation
- Determine what areas of program are the focus
What is attrition?
differences between program and another group due to loss of people from either or both groups rather than the program
Limitations in Comparing Evaluation Results
- Examine & analyze data to look for patterns, recurring themes, similarities/differences
- Address patterns or lack of patterns that justify/don’t justify answers to evaluation questions
- Possible reasons for deviations in established patterns
- Study how patterns are supported/negated by previous studies or evaluations
Process Evaluation
- Any combination of measures that occurs as program is implemented
- Ensures or improves quality of performance or delivery
- Assesses how much intervention was provided (dose), to whom, when, & by whom
Cost savings vs Future costs of programs?
Cost Savings - savings tha occur from prevention or alleviation of disease
Future costs - costs of disease unrelated to intervention
Placebo Effect
Individual’s health improves after taking fake treatment
- In control group and they think they are receiving intervention
- Threat to Internal Validity
What are different types of evaluation indicators?
- Baseline
- Target
Ways to Measure Reliability
- Test-Retest
- Internal Reliability/Consistency
- Split-Half Method
Decision-Making Evaluation Model
- Uses instruments that focus on elements that yield context, input, processes, & products to use when making decisions
- Evaluates criteria that are used for making administrative decisions in the program
Split-Half Method
- 2 parallel forms to administered at same point in time
- Correlation calculated b/w them
Nonresponse Bias
- Lack of responses
- Failure of providing data
- May be due to attrition
Efficiency
How well program and/or intervention can produce positive results
few inputs + higher outputs = MORE EFFECIENT
How is value threshold used?
Determining & allocating resources to intervention rather than another program
Cost-Utility Analysis (CUA)
type of CEA which outcomes of program A & program B are weighted by their value/quality
When are evaluation findings justified?
When they are linked to evidence gathered & judged against agreed upon values or standards set by stakeholders
Threats to Internal Validity
- Ambiguous Temporal Precedence
- History
- Maturation
- Testing
- Instrumentation
- Regression Artifacts
- Selection
- Attrition
- Expectancy threat
- Hawthorne effect
- Social desirability
- Placebo effect
What are performance measures?
Indicators of process, output, or outcomes that have been developed for use as standardized indicators by health programs, initiatives, practitioners or organizations
What is inadequate explanation of constructs in regard to construct validity?
Failure to adequately explicate construct may lead to incorrect inferences
What is the goal of longitudinal designs?
Track changes in factors over time
What is goal of CMA?
Determine which program has lower cost
Implementation Documentation
Collecting data specified in process objectives carried out to demonstrate extent of program implementation to FUNDERS
What is a disadvantage to using existing data collection tools?
Potential for unreliable measures with different population demographics & situations
What is a value threshold?
Benchmark for designating whether service is cost effective
Ambiguous Temporal Precedence
Lack of clarity about whether treatment occurred
What does implementation assessment provide?
- Managerial guidance & oversight
- Informs decision making to which aspects of organizational or service utilization plan are ineffective in accomplishing process objectives
Content Validity
Assesses whether test is representative of all aspects of construct
Triangulation
Examines changes or lessons learned from different points of view or in different ways
What is vote counting?
Defines findings as significantly positive/negative OR nonsignificant
What is process use?
Engagement of designing & conducting evaluation that may lead to better understanding & new ways of thinking about the program
Intrarater Reliability
Correlation between observations made by same observer at different points in time
Cost-Benefit Analysis (CBA)
method of economic evaluation which all benefits & costs of program are measured
Cross Sectional Design
Data about program, events, activities, behaviors, & other factors collected at ONE POINT IN TIME
Goal-Free Evaluation Model
Instruments provide all outcomes (including unintentional positive/negative outcomes)
What needs to be considered when proposing possible explanations of findings?
- Standards
Analysis & Synthesis - Interpretation
- Judgements
- Recommendations
Nonrandom Error
Measure is systematically higher or lower than true score
What should HES consider when using existing data collection instruments?
- If item is appropriate for intended purpose
- If language is appropriate or population
- Whether test has been performed using sample from intended audience
Outcome Evaluation
- Short term, immediate, & observable effects of program leading to desired outcomes
- What changed about public health problem?
Implementation Assessment
Ongoing, nearly real-time activity of collecting data for purpose of making timely corrections or modifications to implementation through changes to elements of process theory
*AKA Program or Process Monitoring **
What is instrumentation in regard to threat to internal validity?
before-after changes due ot changes in the instrument or those administering instrument rather than program
What is descriptive data used for?
To decrease large quantity of data into few elemental measurements that entirely describe data distribution
What are evaluation indicators of program?
Information or statistics that provide evidence of progress toward outcomes
What is mon-operation bias?
Inferences are complicated when definition of construct both underrepresent construct of interest & measures irrelevant constructs
Attainment Evaluation Model
Uses evaluation standards & instruments upon elements that yield objectives & goals of programD
What should HES/researcher consider when choosing method for data collection?
- Specifically target most important elements of study
- Clearly prove or disprove hypothesis
- Appropriate to scale of study
- Do not cost too much or require too much time
Clinical Effectiveness
Improving health of individual patients through medical care services
What is evaluation used in needs assessment?
- Evaluating primary, secondary data, observations, & interviews
- Evaluating literature
What are 2 approaches to meta-analysis?
Vote-Counting & Classic (or Glassian) meta-analysis
Quality Assurance
Using minimum acceptable requirements for processes & standards for outputs
What are tangible benefits?
Benefits that are quantifiable & measurable
Measurement Reliability
type of random error which same measure gives same results on repeated applications
what is selection in regard to internal validity?
Difference between program & another group due to differences in people in the groups rather than the program
Multiple Method vs Mixed Method Designs
Multiple Method: combining qualitative & quantitative DATA to answer evaluation questions
Mixed Method: combing qualitative & quantitative METHODS to answer evaluation questions
What is evaluation used in program implementation?
Evaluating progress of program based on health indicators
What is regression artifacts in regard to internal validity?
If subjects are selected on basis of their extreme score, before-after changes may be affected partly by extreme scores naturally shifting toward mean
Statistical Significance
Likelihood one would be to get result by chance
- 0.05 (usually used)
What does descriptive data describe?
Data that answers a questions
Random Assignment
Process of determining on random basis who does & does not receive health program/intervention
What is testing in regard to threat to internal validity?
Before-after changes due to giving pretest rather than program
CDC evaluation standards
Utility, Feasibility, Propriety, Accuracy
Why is process evaluation important?
- Understanding internal & external forces that can impact activities of program
- Maintain and/or improve quality & standards of program performance and delivery
- May serve as documentation of provisions & success of those provisions of program
What are things to monitor/evaluate to ensure efficiency & effectiveness?
- Simplicity
- Flexibility
- Acceptability
- Sensitivity (proportion of disease)
- Predictive value positive
- Representativeness
- Timeliness
- Stability
What can correlates (relationships b/w variables where on is affected/dependent on another) be derived from?
- Reach & effectiveness
- Size of effect
Concurrent Validity
Assesses degree measure correlates with an already validated measure
- Type of Criterion Validity
- constructs may be same or different
- Related constructs
what can Process Use provide?
Collaboration of different perspectives among interest groups
What is Classic or Glassian Meta-Analysis?
- Defines questions to be examined
- Collects studies
- Codes study features & outcomes
- Analyzes relations b/w study features & outcomes
What is propriety (evaluation standard)?
Behave legally, ethically, & with regard for welfare of participants of program and those affected by program
Production Efficiency
Combining inputs to produce services at lowest cost
What is Persuasive Use?
Evaluation results used to support or criticize program
What should HES do when only using part of data collection instrument to maintain validity?
- Aspects of questions should be retained
- Give credit for using item/collection tool
What does SMOG stand for?
Simple Measure of Gobbledegook
What are the field procedures for collecting data?
- Protocols for scheduling initial contacts with respondents
- Introducing instrument to respondent
- Keeping track of individuals contacted
- Follow up with non-respondents (when appropriate)
What should performance measures be aligned with?
Objectives
What are indirect costs of programs?
Lost or impaired ability to work or engage in leisure activities as a direct result of intervention
Divergent Validity
Measure of construct does not correlate with other measures that should not be related to
HIPPA
Federal regulations for protection of privacy of participant data
Multivariable Analysis
Estimates size & direction of program’s effect in randomized & non-randomized study designs with treatment and control group
What are beneficial outliers?
Outliers that are representative of population
Effectiveness
Degree of how successful program is in producing desired result
What are multivariate outliers?
Unusual combinations of scores on different variables
Clinical Significance
Likelihood intervention is to have noticeable benefit to participants
Validity
Accuracy of measurement (do results represent what should be measured)
Data Analysis Plan
- How data will be scored & coded
- How missing data will be managed
- How outliers will be handled
- Data screening
Convergent Validity
Measure of same concept correlated to each other
- type of construct validity
What is mono-method bias?
When construct is measured using same method and method is part of the construct itself
How does effective data management help evaluator/researcher?
- Organization of data
- Ability to access data
- Analysis of data
- Ensures quality of research
- Supports published results
How can evaluators have less bias in their data collection?
use evaluation questions that allow for more than 1 answer
Evaluation Plan Framework
- Organize evaluation process
- Procedures for managing & monitoring evaluation
- Identify what to evaluate
- Formulate questions to be answered
- Timeframe for evaluation
- Plan for evaluating implementation objectives (process)
- Plan for evaluating impact objectives
- Targeted outcomes (outcome objectives)
What should be used to guide data analysis?
- Research/evaluation questions
- Level of measurement of data
- is it for research? evaluation?
Rigor
Confidence findings/results of evaluation are true representation of what occurred as result of program
What types of systematic errors can occur with findings from evaluation/research?
- Sampling
- Design
- Implementation
- Analysis
What specific readability tools are there to help with this?
SMOG & FleschKincaid
Why must evidence be interpreted in regard to health programs?
Allows for determination of significance & drawing relevant inferences to plan future programs/interventions
What is included in evidence-based practice approach?
- Combination of best available scientific evidence & data
- Program planning frameworks
- Community is engaged
- Programmatic evaluation
- Disseminated results
Nominal/Dichotomous measurement & give an example
Cannot be ordered hierarchically but are mutually exclusive
- Male/Female
- Yes/No
Predictive Validity
Assesses degree measure predicts criterion measure assessed at later time
Threats to Construct Validity
- Inadequate explanation of constructs
- Construct cofounding
- Non-operation bias
- Mono-method bias
- Confounding constructs with levels of constructs
Equity
Maximum potential effect under ideal circumstances
Expectancy Effect
Occurs when researcher’s expectations influence results
Why is conducting meta-analysis important when synthesizing data?
Combination of results to answer research hypotheses
What is the goal of CUA?
Determine which program produces most at lower cost
What are confounding variables?
Extraneous variables or factors outside scope of intervention that can impact results
Face Validity
Common acceptance or belief that measure actually measures what it is supposed to measure
- Expert decides if scale “appears” to measure construct
Criterion Validity
Measure correlates with outcome