Do not be fooled by statistics!

Scientific notes

Recently, I came across a diagnosis ELISA Kit which was claimed to provide a sensitivity of 100% and a specificity of 99%. These statistics were computed from a skewed sample of sera, with only 5 positive cases and more than 300 negative cases.  The potential client should be aware that the confidence on these statistics is very poor for such a low number of positive cases considered.  Let me support my discussion with a similar simulated example. Imagine 300 negative readings of a diagnosis kit, representing non infected sera, are simulated following a normal distribution with 0.5 mean and 0.25 standard deviation. Also, 5 positive readings are simulated with a normal distribution with 1.5 mean and 0.5 s.d. This may be seen as a very realistic situation, according to our experience. With these simulated data, the optimum cut-off following a given criterion (the Youden index) is 1.01, and the sensibility and specificity of the (simulated) diagnosis kit is 100% and 97%, respectively. These are quite close statistics to those reported by the test. But the question is: are they reliable values for the client? The answer is no. Mainly because, as already commented, five positives are not enough.  Another relevant consideration when using a reduced number of samples is that the sensibility and specificity should always be computed from independent data to that used to set the cut-off. Otherwise, we get very optimistic results. As a matter of fact, the less the size of the sample used, the more optimistic the statistics are. To illustrate this, imagine that a client buy the assay, with cut-off 1.01 and sensibility and specificity of 100% and 97%, respectively. I simulated new samples of positive and negative readings, representing the new sera analyzed by the client with the kit, and obtained a sensibility and specificity of 84% and 98%. Why the sensibility was so low for these new readings? Because the number of positive sera used to compute the cut-off (5) was too low to obtain a reliable statistic. To avoid this problem, resampling techniques such as cross-validation provides more realistic estimates. Cross-validation estimates of the sensibility and specificity were of 80% and 97%, much closer to those observed by the client than the original ones. As a conclusion, statistical validation of diagnosis kits should be always supported by sounded statistical procedures and enough data. Otherwise, performance statistics reported are simply not reliable. Written by José Camacho.

Related news

Notas científicas

Vacunación vs infección natural. Un desafío en el diagnóstico de pacientes asintomáticos con COVID-19, cuando se desconoce el perfil de inmunidad.

Se desarrollan intensas pruebas contrarreloj para obtener la mejor vacuna anti-COVID-19. Dos organismos, uno del Reino Unido (Universidad de Oxford) y otro de China (CanSino Biologics) han logrado los resultados más prometedores hasta el momento. Un vector de adenovirus produce ambas vacunas (ChAdO..

Notas científicas

Vuelta a la normalidad a pesar del COVID-19

En este momento de pandemia, un punto muy crítico es saber si cada uno de nosotros estamos inmunizados y al mismo tiempo si contagiamos o no. Saber si las personas están inmunizadas contra el COVID-19 es una información imprescindible para poder deshacer este estado de cuarentena en el que estamo..

Notas científicas

Detección de la nueva enfermedad por coronavirus 2019 (COVID-19) mediante ensayos de serodiagnóstico rápido

El 31 de diciembre de 2019, la OMS fue informada de un grupo de casos de neumonía de causa desconocida detectados en la ciudad de Wuhan, provincia china de Hubei. El 7 de enero las autoridades chinas identificaron la enfermedad por coronavirus (COVID-2019) como el virus causante. Como parte de la r..