Skip to main content
Elsevier - PMC COVID-19 Collection logoLink to Elsevier - PMC COVID-19 Collection
. 2020 Jun 11;185:88–90. doi: 10.1016/j.puhe.2020.06.006

Test, test, test for COVID-19 antibodies: the importance of sensitivity, specificity and predictive powers

N Kumleben a, R Bhopal b, T Czypionka c,d, L Gruer b,e,, R Kock f, J Stebbing g, FL Stigler h
PMCID: PMC7287442  PMID: 32590234

Abstract

Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) antibody tests of varying specificity and sensitivity are now available. For informing individuals whether they have had coronavirus disease 2019 (COVID-19), they need to be very accurate. For measuring population prevalence of past infection, the numbers of false positives and negatives need to be roughly equal.

With a series of worked examples for a notional population of 100,000 people, we show that even test systems with a high specificity can yield a large number of false positive results, especially where the population prevalence is low. For example, at a true population prevalence of 5%, using a test with 99% sensitivity and specificity, 16% of positive results will be false and thus 950 people will be incorrectly informed they have had the infection. Further confirmatory testing may be needed.

Giving false reassurance on which personal or societal decisions might be based could be harmful for individuals, undermine public confidence and foster further outbreaks.

Keywords: COVID-19, Antibody tests, Specificity, Sensitivity

Highlights

  • COVID-19 antibody tests need to be very reliable to inform individual decisions.

  • Even with high specificity, antibody tests can produce many false positives.

  • Giving false reassurance could undermine public confidence and foster new outbreaks.


To help reverse the current lockdowns while suppressing COVID-19 rates, we need to identify who currently has the infection and who has had it and recovered. As reverse transcriptase polymerase chain reaction (RT-PCR) testing to detect current infection has been recently discussed in detail,1 we focus in this article on antibody tests. The presence or absence of antibodies can inform individuals if they have had the infection or not and guide personal and societal decisions about if and when they can return to normal activities. Antibody testing thus needs to be particularly accurate. It can also be used to provide an estimate of the population prevalence of previous infection. We demonstrate that for this purpose high accuracy is not required, but the numbers of false positives and false negatives need to be approximately equal.

Antibody tests are increasingly available but with variable accuracy. It is hoped they can be used to identify people who are at least partially immune. Immunity certificates, a more appropriate phrase than immunity passports that promises too much, for individuals thought to have recovered from COVID-19, are being discussed internationally.2, 3, 4 Whether tests are carried out for clinical diagnosis, screening or immunity certificates, we need to have sufficient confidence they are accurate.

A sensitive test will detect the presence of antibodies to SARS-CoV-2 (the virus that causes COVID-19), and a specific test will not react to other antibodies e.g. to other coronaviruses. No diagnostic or screening test is perfect and incorrect results are inevitable, not least because the timing of the test is critical. Seroconversion takes time, with IgM, IgG and IgA antibodies usually developing in that order, and can be variable and dependent upon the severity of the illness and the individual's immune system. Antibody levels subsequently decline with time. Antibody test systems may perform less well than the manufacturers' results suggest. For example, both Roche and Abbott reported their antibody test had 100% sensitivity for samples taken 14 days or more after the onset of symptoms, yet Public Health England found sensitivity at 14 or more days of only 87% and 93.4%, respectively.5 , 6

We show here how to measure the test's accuracy and how this changes along with the prevalence of disease (12 tables showing the results with varying sensitivity, specificity and population prevalence of 1%, 5%, 10% and 20% are available in the Supplementary File). The two key measures of its accuracy are sensitivity and specificity, set out in Table 1 , with the cells identified as A (true positives), B (false positives), C (false negatives) and D (true negatives). Sensitivity (A/A + C) is the proportion of people with a disease who, when tested, receive a positive test result. It is also known as the true positive rate. Specificity (D/D + B) is the proportion of individuals without a disease who, when tested, receive a negative test result. It is also known as the true negative rate.

Table 1.

Predictive powers of a test with 90% sensitivity and specificity (5% prevalence).

Test result (90% sensitivity and 90% specificity) People truly with disease People truly without disease Totals
Positive 4500 (A) 9500 (B) 14,000
Negative 500 (C) 85,500 (D) 86,000
Total 5000 95,000 100,000

Predictive value of a positive test: A/A + B = 32.1%.

Predictive value of a negative test: D/D + C = 99.4%.

To establish sensitivity and specificity, we could test a sample of patients with proven disease (in this case laboratory detection of SARS-CoV-2), and a sample of people known to be free of disease (for example, using stored blood samples taken before COVID-19 existed in humans). In practice, a test's performance will usually be poorer than the values established due, for example, to problems in storing or transporting specimens or the variable time lag from the onset of infection until antibodies appear in the blood (seroconversion) and then decline. The proportion of test results that are false partly depends on the prevalence of the disease in the population. With a low prevalence, even a test with high sensitivity and specificity will produce a high proportion of false positives. In this article, we focus on the outcomes of tests of variable accuracy with 5% population prevalence in a hypothetical group of 100,000 people, of whom 5000 have had the infection and 95,000 have not. This is a plausible current prevalence of past COVID-19 in many countries7, 8, 9 although it could be a lot higher in some areas.

Table 1 shows that if the sensitivity is 90%, 4500 people will correctly test positive, but 500 will incorrectly test negative and be wrongly told they have no antibody evidence of the disease. If the specificity is 90%, 85,500 people will correctly test negative, but 9500 will incorrectly test positive and be wrongly told they have antibody evidence of previous infection. Thus, of the 14,000 people who received positive test results, only 32% (4500/14,000; A/A + B) had the disease. This is referred to as the predictive value (or power) of a positive test. The other 68% would be given wrong information. Of the 86,000 people who received negative tests, 99% (85,500/86,000; D/C + D) would receive a correct result. This is called the predictive value (or power) of a negative test.

Sensitivity and specificity vary with different tests but, for any particular antibody test, these can be adjusted by altering the level of antibody required to determine a positive result. Requiring a higher level of antibody for a positive result would increase the specificity but lower the sensitivity. This would reduce the false positives (C) but increase the false negatives (B). Choosing a test that has 80% sensitivity and 99% specificity, as shown in Table 2 , 81% of people who test positive have had the disease, an increase from 32%. Now, about one in five people who test positive will not have had the disease. This shows that when the prevalence of the disease is low, antibody testing, even with a specificity as high as 99%, still produces many false positives so the predictive power of a positive test is far from 100%.

Table 2.

Predictive powers of a test with 80% sensitivity and 99% specificity (5% prevalence).

Test result (80% sensitivity and 99% specificity) People truly with disease People truly without disease Totals
Positive 4000 (A) 950 (B) 4950
Negative 1000 (C) 94,050 (D) 95,050
Total 5000 95,000 100,000

Predictive value of a positive test: A/A + B = 80.8%.

Predictive value of a negative test: D/D + C = 98.9%.

If a test is extremely accurate, as is claimed for the Roche and Abbott systems, say 99% sensitivity and specificity, the results are shown in Table 3 . Even now, the predictive power of a positive test has only risen from 81% with a sensitivity of 90%, to 83.8%. If the prevalence rises to 20% then the predictive power of a positive test is 96.1% and of a negative test 99.7% (Supplementary File Table A12).

Table 3.

Predictive powers of a test with 99% sensitivity and 99% specificity (5% prevalence).

Test result (99% sensitivity and 99% specificity) People truly with disease People truly without disease Totals
Positive 4950 (A) 950 (B) 5900
Negative 50 (C) 94,050 (D) 94,100
Total 5000 95,000 100,000

Predictive value of a positive test: A/A + B = 83.8%.

Predictive value of a negative test: D/D + C = 99.9%.

If immunity certificates, or personal or societal decisions about returning to normality, are based on these results, a significant proportion will be incorrect. Where the disease has become highly prevalent, for example, among health care and care home workers, the power of a positive test would be higher, therefore more reliance could be placed on it. Even with a prevalence of 20% and 99% sensitivity and specificity, the test itself does not give a guarantee at the individual level, and personal and clinical judgements are required in applying the findings. A major hope of antibody testing is that those who test positive can resume work and social activities more fully and confidently than those who test negative. The presence of antibodies should signify the same illness will not recur, the person is not contagious and there is at least partial immunity to future COVID-19 infections. We need to establish whether this is true.10

If the purpose of antibody testing is to assess the prevalence of COVID-19 in a representative sample of the population, these clinical issues do not apply. The veracity of the prevalence derived by such measurements will depend upon achieving equal false positives and false negatives. For example, although the true prevalence is 5%, Table 1, Table 2, Table 3 give a prevalence in the hypothetical population of 100,000 people of 14% (14,000 positives), 4.95% (4950 positives), and 5.9% (5900 positives), respectively. Perhaps surprisingly, the test with 80% sensitivity and 99% specificity (Table 2) gives the most accurate estimate at this level of population prevalence.

In conclusion, at currently reasonable estimates of the general population prevalence, even high sensitivity and specificity will produce an important number of false positives. People testing positive, especially those without indicative case histories, may need further testing to confirm the result. Given the current uncertainty about the level of immunity signalled by antibodies, all those testing positive for antibodies would be well advised to maintain protective measures. More information is also urgently needed to ascertain the strength and duration of immunity in people who have recovered from COVID-19, and whether some can still be infectious or become reinfected. Giving false security and reassurance could be harmful for individuals, undermine public confidence and foster further outbreaks.

Author statements

Ethical approval

Not required.

Funding

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Competing interests

None declared.

Author contributions

The article was conceived by J.S. during a discussion initiated by R.B. in the COVID-19 researchers Google Group. The manuscript was drafted by N.K. All authors commented on the drafts and agreed the final version. L.G. is the corresponding author and guarantor of the manuscript.

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.puhe.2020.06.006.

Appendix A. Supplementary data

The following is the Supplementary data to this article:

Multimedia component 1
mmc1.docx (19.6KB, docx)

References

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.docx (19.6KB, docx)

Articles from Public Health are provided here courtesy of Elsevier

RESOURCES