Hiding part of the data leads to wrong conclusions
There have been numerous papers published showing how well the vaccines protect people after the second dose. Some of this effect is an illusion. The effect happens as a result of inaccurate measuring and a phenomenon called survivorship bias.
Survivorship bias happens when a group is compared at two time points, but the members of the group change between the time points. It would be like assessing the quality of a swimming school which favours the technique of throwing people into the middle of the ocean, leaving them for a couple of hours and claiming credit for how well the remaining students can swim. After two hours, the only people left would be the ones who could already swim and possibly a few who learnt to swim the hard way! The poor souls who drowned in the interim don’t even make the count. Attributing the remaining people’s swimming ability to the coach who turned up 2 hours later would obviously give a very misleading picture. Pointing out that no-one drowned in later lessons would be equally misleading in determining the success of the ‘teaching technique’.
With covid vaccination there is a two week period after vaccination that is not included in the data. The rationale given for this is that vaccines take a while to induce antibodies and therefore the first two weeks’ data are not relevant. Obviously this is flawed. What if the vaccines have deleterious effects that are visible straightaway, that have nothing to do with antibody production? An example is the high rate of shingles seen after covid vaccination, suggesting there is a problem with viral reactivation. This may explain why Sars-CoV-2 infection rates are actually higher in the vaccinated than in the unvaccinated in the first two weeks after vaccination.
The effect of eliminating the first two weeks is a misleading data bias. If people become infected and are dying during that period, this needs to be included. The possibility that the vaccine itself may exert an effect on infection rate cannot be overlooked and the entire dataset needs to be included in order to accurately assess effectiveness. By only measuring the period after the higher risk of infection (0-14 days) it is possible to be deceived. Any signal would be missed.
Aside from it being nonsensical in terms of individual risk to remove this period of time, there will also be an impact on the wider community. If the vaccine in fact causes a spike in infections during the first two weeks, this will inevitably increase spread and will lead to an increased number of infections in that community during that time. Therefore, the assessment of the impact of the vaccination programme must include not only the effect on the individual, but the impact on the wider community.
This point is of particular relevance for close-knit communities where many are being vaccinated at the same time, such as schools and in particular communities with a high number of vulnerable people such as care homes and hospitals. What we are effectively doing is ‘speeding up’ the wave of infections (and deaths). Ultimately at the end of the viral season, the same number of people died. Because of excluding the earlier deaths (1-14 days), we are misled into thinking that the vaccines were more effective than they actually were. By only looking at the later period and seeing fewer deaths during that time, the illusion was created that lives were saved. This is evident in data from many countries following vaccine roll-out. The graph below showing the UK versus Europe illustrates this point, as the UK was the fastest to roll out the vaccine. The total number of deaths, represented by the area under the curve, was similar to other countries, but is just compressed into a shorter time period.
Figure 1: Covid Deaths in winter in UK and the European Union