Classification in Diagnostic Tests Flashcards
What is sensitivity?
Sensitivity is the true positive rate of a test or measure. This means the percentage of participants, of who you know that they have a certain disorder/pathology, are correctly classified as having a certain pathology/disorder/etc.
What is specifity?
Specifity is the true negative rate of a test or measure. This means the percentage of participants who belong to the healthy control group get also classified as having no disorder/pathology.
What is the positive predictive power of a test?
This is the confidence you can have (in terms of percentage) that an outcome of a test is actually predicting what the measure should be predicting.
What is negative predictive power?
Related to specifity; the confidence you can have in that an outcome of a test which says that someone has no pathology/disorder is in reality also predicting this.
Which of these TOC values are the most important for clinical practice?
PPP and NPP
Which factor influences PPP and NPP? And which implication does this have for clinical practice?
Prevalence rates of a given disorder; if this is higher, the PPP will increase. If this is a lower prevalence, the NPP will increase. This means that you really have to be careful to have the same base rate of the disorder in your sample as is the case in the population, otherwise you will have over- or underestimations of PPP and NPP.
What is the current problem with TOC values in NP-research?
That most researchers only include calculations of sensitivity and specifity, and don’t consider PPP and NPP, thereby excluding important information. Meaning, that clinicians have (with only S and S) no idea how confident they can be that a certain measure rules in/out a disorder correctly.
Clinicians need information on true positives (1), true negatives (2), but also false positives (3) and false negatives (4) when using a test or measure for clinical decision-making.
- Sensitivity
- Specifity
- PPP
- NPP
What are the main recommendations for TOC-values according to Lange & Lippa?
(1) Always calculate sensitivity, specificity, PPP, and NPP.
(2) Calculate PPP and NPP values based on a range of hypothetical base rates of the
condition in the target population, not the actual base rate of the condition in the
experimental sample.
(3) Always evaluate the clinical utility of a test/measure based on the interpretation of
sensitivity, specificity, PPP, and NPP together. Never interpret sensitivity and specificity
in isolation. Never interpret positive predictive power and negative predicting
power in isolation.
What is the main difference in studying test results in NP-research and medical tests?
In NP-research, we mainly focus on finding significant associations and causality (this is often sufficient), while in medical tests this is far from sufficient. To know if results have clinical implications, they also need information on the sensitivity, specifity, CI’s and likelihood rations (more generally speaking, descriptive statistics). P-values and odds ratio’s are, so to say, secondary statistics in medical tests.
Which two forms of variability are there in tests results?
- Intra-observer variability = the lack of reproducibility in results when the same observer/laboratory performs the test on the same specimen at different times.
- Inter-observer variability = the lack of reproducibility among two or more observers.
Studies of reproducability adress (1)….., not (2)…. or (3)…..
(1) precision
(2) accuracy
(3) validity
- –> All observers can agree with each other and still be wrong.
What is the basic design to assess reproducibility?
Comparing test results from more than one observer (inter-observer) or that were performed on more than one occasion (intra-observer)
How can the inter-observer agreement be measured in categorical variables?
- The percentage of observations on which the observers agree exactly. Cons: hard to interpret, and it counts partial agreement the same as complete agreement.
- Kappa measures the extent of agreement beyond what would be expected from observers’ knowledge of the prevalence of abnormality, and can give credit for partial agreement.
How can the inter-observer agreement be measured in continuous variables?
This depends on the design of the study; can be done with mean differences between the measurements, SD’s, etc. Another possibility is the Coefficient of Variation (CV) which is the standard deviation of all of the results obtained from a single specimen divided by the mean value.