Sensitivity and specificity are statistical measures of the performance of a binary classification test, also known in statistics as classification function. Sensitivity (also called recall rate in some fields) measures the proportion of actual positives which are correctly identified as such (e.g. the percentage of sick people who are correctly identified as having the condition). Specificity measures the proportion of negatives which are correctly identified (e.g. the percentage of healthy people who are correctly identified as not having the condition). These two measures are closely related to the concepts of type I and type II errors. A theoretical, optimal prediction aims to achieve 100% sensitivity (i.e. predict all people from the sick group as sick) and 100% specificity (i.e. not predict anyone from the healthy group as sick), however theoretically any predictor will possess a minimum error bound known as the Bayes error rate.
For any test, there is usually a trade-off between the measures. For example: in an airport security setting in which one is testing for potential threats to safety, scanners may be set to trigger on low-risk items like belt buckles and keys (low specificity), in order to reduce the risk of missing objects that do pose a threat to the aircraft and those aboard (high sensitivity). This trade-off can be represented graphically as an Receiver operating characteristic curve.
Contents |
Imagine a study evaluating a new test that screens people for a disease. Each person taking the test either has or does not have the disease. The test outcome can be positive (predicting that the person has the disease) or negative (predicting that the person does not have the disease). The test results for each subject may or may not match the subject's actual status. In that setting:
Sensitivity relates to the test's ability to identify positive results.
Again, consider the example of the medical test used to identify a disease. The sensitivity of a test is the proportion of people who have the disease who test positive for it. This can also be written as:
If a test has high sensitivity then a negative result would suggest the absence of disease.[1] For example, a sensitivity of 100% means that the test recognizes all actual positives – i.e. all sick people are recognized as being ill. Thus, in contrast to a high specificity test, negative results in a high sensitivity test are used to rule out the disease.[1]
From a theoretical point of view, a 'bogus' test kit which always indicates positive, regardless of the disease status of the patient, will achieve 100% sensitivity. Therefore the sensitivity alone cannot be used to determine whether a test is useful in practice.
Sensitivity is not the same as the precision or positive predictive value (ratio of true positives to combined true and false positives), which is as much a statement about the proportion of actual positives in the population being tested as it is about the test.
The calculation of sensitivity does not take into account indeterminate test results. If a test cannot be repeated, the options are to exclude indeterminate samples from analysis (but the number of exclusions should be stated when quoting sensitivity), or, alternatively, indeterminate samples can be treated as false negatives (which gives the worst-case value for sensitivity and may therefore underestimate it).
A test with a high sensitivity has a low type II error rate.
Specificity relates to the ability of the test to identify negative results.
Consider the example of the medical test used to identify a disease. The specificity of a test is defined as the proportion of patients who do not have the disease who will test negative for it. This can also be written as:
If a test has high specificity, a positive result from the test means a high probability of the presence of disease.[1]
From a theoretical point of view, a 'bogus' test kit which always indicates negative, regardless of the disease status of the patient, will achieve 100% specificity. Therefore the specificity alone cannot be used to determine whether a test is useful in practice.
A test with a high specificity has a low type I error rate.
In medical diagnostics, test sensitivity is the ability of a test to correctly identify those with the disease (true +ve rate), whereas test specificity is the ability of the test to correctly identify those without the disease (true -ve rate).[2]
If 100 patients known to have a disease were tested, and 43 test positive, then the test has 43% sensitivity. If 100 with no disease are tested and 96 return a negative result, then the test has 96% specificity.
A highly specific test is unlikely to give a false positive result: a positive result should thus be regarded as a true positive. A sign or symptom with very high specificity is often termed pathognomonic. An example of such a test is the inspection for erythema chronicum migrans to diagnose lyme disease.[3] In contrast, a sensitive test rarely misses a condition, so a negative result should be reassuring (the disease tested for is absent). A sign or symptom with very high sensitivity is often termed sine qua non. An example of such test is a genetic test to find an underlying mutation in certain types of hereditary colon cancer.[4][5]
SPIN and SNOUT are commonly used mnemonics which says: A highly SPecific test, when Positive, rules IN disease (SP-P-IN), and a highly 'SeNsitive' test, when Negative rules OUT disease (SN-N-OUT).
Condition (as determined by "Gold standard") |
||||
Condition Positive | Condition Negative | |||
Test Outcome |
Test Outcome Positive |
True Positive | False Positive (Type I error) |
Positive predictive value = Σ True Positive Σ Test Outcome Positive
|
Test Outcome Negative |
False Negative (Type II error) |
True Negative | Negative predictive value = Σ True Negative Σ Test Outcome Negative
|
|
Sensitivity = Σ True Positive Σ Condition Positive
|
Specificity = Σ True Negative Σ Condition Negative
|
Patients with bowel cancer (as confirmed on endoscopy) |
||||
Condition Positive | Condition Negative | |||
Fecal Occult Blood Screen Test Outcome |
Test Outcome Positive |
True Positive (TP) = 20 |
False Positive (FP) = 180 |
Positive predictive value
= TP / (TP + FP)
= 20 / (20 + 180) = 10% |
Test Outcome Negative |
False Negative (FN) = 10 |
True Negative (TN) = 1820 |
Negative predictive value
= TN / (FN + TN)
= 1820 / (10 + 1820) ≈ 99.5% |
|
Sensitivity
= TP / (TP + FN)
= 20 / (20 + 10) ≈ 67% |
Specificity
= TN / (FP + TN)
= 1820 / (180 + 1820) = 91% |
Related calculations
Hence with large numbers of false positives and few false negatives, a positive FOB screen test is in itself poor at confirming cancer (PPV = 10%) and further investigations must be undertaken; it did, however, correctly identify 66.7% of all cancers (the sensitivity). However as a screening test, a negative result is very good at reassuring that a patient does not have cancer (NPV = 99.5%) and at this initial screen correctly identifies 91% of those who do not have cancer (the specificity).
In information retrieval positive predictive value is called precision, and sensitivity is called recall.
The F-measure can be used as a single measure of performance of the test. The F-measure is the harmonic mean of precision and recall:
In the traditional language of statistical hypothesis testing, the sensitivity of a test is called the statistical power of the test, although the word power in that context has a more general usage that is not applicable in the present context. A sensitive test will have fewer Type II errors.
Source: Fawcett (2004). |
|