Practice and Process Improvement Flashcards

1
Q

Research process

A
Identify the problem
Review of literature
Identify conceptual framework
Define concepts and variables
Select design
Identify population and sample
Determine measurement methods
Plan data collection and analysis
Implement: collect, analyse, interpret
Disseminate findings (most important)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Qualitative inquiry

A

Subjective (words)
Inductive
Broad focus
Meaning/discovery

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Quantitative inquiry

A
Single, measureable reality
Objective (numbers and statistics)
Deductive
Narrow focus
Cause and Effect
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Research study critique

A

Introduction: is problem identified
Problem statement: is it clear
Purpose statement: clearly stated goal or objective?
Review of literature: is the review comprehensive
Theory and conceptual framework; Clearly identified and discussed
REsearch question: clearly stated? Further delineates the problem area of study?
Design/methodology: is the design appropriate to test the study
Ethics: appropriate IRB approval
Data: were collection methods clearly described? Statistics clearly discussed? Demographics?
Findings: are findings presented clearly and objectively? Is each hypothesis addressed separately?
Limitations: any that may have influenced findings?
Implications: suggestions for nursing practice, education or further nursing research

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Evidence-based Practice steps

A

Formulate a burning question using PICOT:
Population or patient; interest or intervention; comparison; outcomes; time (optional) it takes to achieve outcome
Search best evidence
Review evidence
Integrate evidence with clinical expertise and patient preferences/values in making practice change
Evaluate effectiveness is of change meeting desired outcomes
Disseminate findings

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Rapid critical appraisal

A

Validity: was research study conducted in a rigorous manner
reliability: can others realistically expect to achieve the same results
Applicability: does the study apply to my patients/learners?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Levels of evidence

A
Systematic review or meta-analysis
RCT
Controlled trial without randomization
Case control or cohort study
Systematic review of qualitative or descriptive studies
Qualitative or descriptive study
Expert opinion or consensus
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Evaluating the evidence: 3 important question

A

Is the evidence strong enough to implement the findings into practice?
Is the evidence so weak that a research study needs to be conducted?
Should we return to the literature and do a more complete search, widening the sources used?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Principles of quality management

A

Customer focus
Leadership
Involvement of people at all levels of the organization
Process approach
System approach to management of interrelated processes
Continual improvement
Factual approach to decision making
Mutually beneficial supplier relationships

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

6 measures of quality

A

Safe; timely; effective; efficient; equitable; patient-centered

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

QSEN versus IOM Measures of healthcare quality

A

QSEN : IOM
Pt centered care : delivering pt centered care
Teamwork and collaboration : working as part of an interdisciplinary team
EBP : Practicing EBP
QI : focusing on QI
Informatics : using information technology
Safety

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

FOCUSPDCA

A
Find a problem
Organize a team
Clarify the problem
Understand a problem
Select an intervention
Plan
Do 
Check the results
Act
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Six Sigma Improvement

A

Near-elimination of defects from every product, process or service; minimizing variances;
Approaches: DMADV, DMAIC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

six sigma approaches

A

DMADV (define, measure, analyze, design and verify) :developing a new process
DMAIC (define, measure, analyze, improve and control) improving an existing process

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Lean thinking

A

Provide perfect value to the customer through zero waste (time waste, supplies where they need to be, improves efficiency)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Rapid cycle improvement

A
Small, concrete changes that can be implemented and piloted quickly
Set the aim
Define the measures
Make the change
Test changes using PDSA
Repeat cycle if changes or did not work.
17
Q

FMEA. Failure mode and effect analysis

A

Looking in a future way where is there a potential for errors to occur and the likelihood for that to happen, goal is error prevention, may be used to evaluate new or existing processes

18
Q

Steps in FMEA

A

Select a process
Select a team
Describe the process
Identify what could go wrong at each step of process
Work on eliminating the problems chosen
Design and implement changes to reduce or prevent problems
Measure the success of the process changes

19
Q

Dashboards

A

Internal formative assessment of performance and quality care

20
Q

PICOT

A

P - population or patient - most important characteristics
I - interest or intervention - Issue or intervention to be studied, measure, improve, or affected
C - comparison - Describe the main alternatives
O - outcomes - Describe what is to be accomplished
T - Time it takes for intervention to achieve outcome - describe the timeframe

21
Q

3 components of EBP

A

Best evidence
Clinician (NPD) Expertise
Patient (learner) values/preferences

22
Q

Quality improvement skills

A

Using systematic processes and tools to define problems (flowchart, cause effect diagrams)
Recognizing potential causes of problems (RCA)
Planning and implementing change
Using tools and data to measure quality, performance and variation (control chart, run chart)
Selecting appropriate methods for testing possible solutions (PDCA)
Evaluating effects of change to improve care and practice

23
Q

Flowchart

A

Flow of work through a process including activities, decisions, and measurement points. Oval – beginning and end, rectangle Dash steps in the process, diamond – decision making point

24
Q

Line graph or run chart

A

Shows data trends over time Y axis shows the time delay, defects or cost and X axis shows the time

25
Q

Pareto chart

A

Bar graph showing frequency of measures in descending order, focuses on the improvement effort by identifying the 20% of the contributors that create 80% of the time delay, defects, or cost in any process

26
Q

Cause-effect diagram or Fishbone diagram

A

Begins with major causes and work backward to read causes, five whys; Problem statement and brainstorm to identify potential issues related to specific categories – people, process, materials, management, environment, policy, etc.

27
Q

Control chart

A

Helps analyze, sustain, and monitor the current levels of process stability and identify key issues for problem-solving or root cause analysis. Looks for variation over time.

28
Q

Histogram

A

Bar graph showing frequency and pattern of distribution of one variable - numeric data

29
Q

Bar graph

A

Compares multiple variables – categorical data

30
Q

Matrix

A

Compares two or more groups of ideas, determines relationships among the elements, and makes decisions. It helps prioritize tasks or issues to a decision making and shows linkages between large groups of characteristics, functions, and tasks.

31
Q

Tree diagram

A

Systematically links ideas, target, objectives, goals, or activities in greater and greater detail it shows key goals, some goals, measures, and tasks required to accomplish an objective.

32
Q

Scatter diagram

A

Plotting data on a graph to look for correlations between variables (positive, negative, or no correlation)

33
Q

FADE Methodology

A

Focus, analyze, develop, execute/evaluate

34
Q

CIPP Model

A

Model of program evaluation: Context, Input, Process, Product; evaluating to aid in decision making and improve process
Context - goals, needs assessment; Input - planning, Process - action/implementation, Product - evaluate outcomes
provides formative and summative evaluation