Measures of Reliablity and Validity Flashcards

1
Q

requires constant collection, evaluation, analysis, and use of quantitative and qualitative data.

A

Clinical Medicine

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Error

A
  • Mistakes in the diagnosis and treatment of patients
  • Mistakes due to clear negligence
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Goal

A

minimize error in data so as to guide, not mislead

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

PROMOTING

A

Accuracy and Precision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What errors do we need to reduce

A

Differential and Indifferential Erros

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

We need to reduce what variability?

A

intraobserver and interobserver

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

closer to the true value

A

Accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Also known as “reproducibility” or “reliability”

•Ability of a test to give the same result or a similar result with repeated measurement of the same factor

A

Precision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Differential error

A

information errors differ between groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

information is incorrect, but is the same across groups

A

Nondifferential error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

refers to any systematic error that may occur during the collection of baseline or follow-up data

A

Measurement bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Examples of Measurement Bias

A

• Blood pressure values
• measuring height with shoes on
• laboratories and the use of different methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

• Variability and unpredictability
• results in lack of precision
• some observations are too high and some are too low

A

Random error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

one observer examining the same results more than once

A

Intraobserver variability (within the observer)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Interobserver variability (between observers)

A

2 or more observers examining the same material

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

measure of the consistency of a metric or a method

A

Reliability

17
Q

• Overall percent agreement
• Paired observation
• Multiple variables
• Kappa test ratio

A

MEASURES OF RELIABILITY

18
Q

Common way to measure agreement

A

Overall Percent Agreement (OPA)

19
Q
  • does not include prevalence
  • does not show how disagreement occurred
  • Agreement might be due to chance alone
A

Drawbacks of OPA

20
Q

Percent Agreement Formula

A

PA = a+d/a+b+c+d(100)

21
Q

Measures the extent to which agreement exceeds that expected by chance

A

KAPPA TEST RATIO

22
Q

Kappa =
(Percent Agreement Observed)-(Percent agreement expected by chance alone)/100% -(Percent agreement expected by chance alone)

A

FORMULA FOR KAPPA TEST RATIO

23
Q

INTERPRETATION OF KAPPA

A

<0 = Less than Chance agreement
0.01-0.20 = Slight
0.21-0.40 = Fair
0.41-0.60 = Moderate
0.61-0.80 = Substantial
0.81-0.99 = Almost perfect agreement

24
Q

Ability of a test to distinguish between
WHO HAS a disease and WHO DOES NOT

A

Validity

25
Q

Screening tests

A

is performed as a preventative measure

26
Q
  • able to correctly identify who has the disease
  • Reliably finding a disease when it is present
  • Avoids false negative results
A

Sensitivity

27
Q

Specificity

A

– correctly identifies who does not have the disease
- Reliably excluding a disease when it is absent
- Avoids false positive results

28
Q

Type 1 error / false-positive error / alpha error

A

Finding a positive result in a patient in whom disease is absent

29
Q

Finding a negative result in a patient whom disease is present

A

Type II error / false-negative error / beta error

30
Q

Sensitivity Formula

A

TP/TP+FN

31
Q

Specificity Formula

A

TN/TN+FP

32
Q

False Positive error rate formula

A

b/(b+d)

33
Q

False Negative Error rate formula

A

c/(a+c)

34
Q

Predictive Values

A

Describes the probability of having actual disease given the results of a test

35
Q

Positive predictive value (PPV)

A

Indicates what proportion of the subjects with positive test results actually have the disease

36
Q

Negative predictive value (NPV)

A

Indicates what proportion of the subjects with negative test results actually do not have the disease

37
Q

Formula for PPV

A

PPV = number of people with gold-standard evidence of disease who test positive (a)/ number of people who test positive (a+b)

38
Q

NPV FORMULA

A

NPV = number of people with gold-standard absence of disease who test negative (d)/ number of people who test negative (c+d)

39
Q

Sensitivity & specificity versus
predictive value

A

Sensitivity and specificity are characteristics of a test.

Positive predictive value (PPV) and negative predictive value (NPV) are best thought of as the clinical relevance of a test.