At this point we have huge amounts of data from all around the world that are relevant to determining how effective the vaccines are against infection, hospitalization, death and transmission.
We do not, however, have any really good data.
Having really good data would mean knowing how many Covid infections, hospitalizations, deaths and transmissions occur in a group of vaccinated people and, in similar circumstances, in a similar group of unvaccinated people. The groups have to be similar in terms of age, sex, comorbidities, health-related behavior and so on. The circumstances have to be similar in terms of virus incidence, living situations, government restrictions, healthcare facilities and so on.
If we had such numbers we could straightforwardly calculate vaccine effectiveness, expressed in percentages, using this formula:
(attack rate in unvaccinated group minus attack rate in vaccinated group)
divided by
(attack rate in unvaccinated group)
times
100
But we don’t have these numbers.
The data that we do have are flawed to varying degrees. This is because establishing what the causal effects of pharmaceutical interventions are is fraught with difficulties. In the processes of producing, collecting, handling, interpreting, and presenting the data necessary to determine how effective the vaccines are, there are countless steps that introduce potential sources of bias, inaccuracy, or even downright deception.
The best way to avoid most of these problems is through well designed, properly conducted, double blind randomized controlled trials (RCTs). There are many good reasons to be cautious if those data do not directly come from such trials. This is especially the case in an atmosphere in which there is tremendous pressure and incentive to make the vaccines look as good as possible
And sadly, we simply don’t have RCTs that provide high quality data to tell us how effective these vaccines actually are against:
The absence of medium and long-term RCTs and of high quality active and passive adverse events surveillance systems also mean that we do not have good evidence about how safe these vaccines are, especially in the long-term.
The only RCT data we do have — from the RCTs conducted by the pharmaceutical companies that developed the vaccines — provide us with relatively short-term data on how effective the vaccines are at preventing symptomatic infection. The trials were not designed to demonstrate effectiveness against severe disease or death. The number of trial participants— especially elderly people and people with severe comorbidities who are most vulnerable to severe disease and death — was simply much too small.
Moreover, because the control groups were unblinded a few months into the trials when the participants were given the opportunity to take what now appeared to be a highly effective vaccine, there could also not be any meaningful medium and long-term safety and effectiveness data. Lastly, there have been credible accusations that for example the Pfizer trial was not actually properly conducted, nor reliably double-blinded even before the official unblinding took place.
Most of the other evidence we have about vaccine effectiveness comes from observational studies, cases/hospitalizations/deaths numbers released by public health institutions, experimental data showing various types of immune responses, and from the governments and media reporting on such data. All of these can be highly problematic, although of course some non-RCT data are higher quality than other data, and some institutions and media are more capable or honest in how they interpret and represent those data than others.
When comparing the number of cases, hospitalizations and deaths that occur in vaccinated and unvaccinated populations, scientists, public health institutions and media will ideally:
This is by no means an exhaustive list of the criteria that studies that compare results in vaccinated and unvaccinated people should meet. But it is enough to give an impression of just how difficult it is in the absence of well designed and properly conducted double blind RCTs to make such comparisons in meaningful and accurate ways.
When it comes to vaccine effectiveness data biases creep in not just in how data are handled but in how data are generated and collected as well. For example, case rates may be influenced by differences in testing behaviors:
And when it comes to hospitalization and deaths data, hospitals and doctors may have different Covid testing and coding rules and practices for vaccinated and unvaccinated people:
These are just some of the ways in which slight biases and differential treatments can distort the raw data that will be used by researchers to compare differences in case, hospitalization and death rates between vaccinated and unvaccinated people.
By no means are all these factors present everywhere and all of the time. Some hospitals and doctors will have stricter rules and practices in place that reduce the effect such biases could have than others do.
These decisions are also typically made on the margin: If a fully vaccinated person is hospitalized with obvious signs of severe Covid disease then doctors will be as unlikely to disregard the patient’s positive Covid test as they would in the case of a similar Covid patient who is unvaccinated. But in less clear-cut, more ambiguous situations doctors might make different reporting and treatment decisions for unvaccinated and vaccinated people.
Such different decisions do not imply that the doctors and hospitals intend to bias the results to make vaccination seem more effective. Many of the examples mentioned involve reasonable decisions, based on reasonable assumptions, that healthcare professionals have to make in situations in which time, resources and information are scarce. But the net effect may be as distortive as when there was an intention to deceive.
It is however also not completely out of the realm of possibility that some doctors may subconsciously or consciously provide their vaccinated patients with better treatment than their unvaccinated patients, as a result of their personal antipathy or their sense that scarce resources are better spent on people who ‘did their part’. An example is what this influential doctor wrote (and apparently felt comfortable enough to write) in a discussion of how Ontario hospitals should use the scarce resource of lifesaving monoclonal antibodies:
Such examples, one would hope, are rare.
The individual net effect of each or most of the biases just mentioned will be small but when several such biases are at play they can add up. Moreover, there may be a self-fulfilling prophecy effect. For example, the assumption that vaccines are highly effective at preventing severe disease can result in several biases that in turn make the vaccines seem more effective at preventing severe disease. Over time the realization that the vaccines are not as effective as previously thought may also result in changes in the protocols and practices that then in turn result in data that further confirm this. It is possible that part of the appearance of waning vaccine effectiveness in the data is in fact the result of protocols and practices being changed in response to indications that the vaccines are less effective than previously thought.
What the total effect of all these potential biases is — how it will all shake out — is impossible to determine. The best thing to do is simply avoid them. And the best way to do that is through well designed, properly conducted and double blind RCTs. Blinding the participants, for example, removes many of the potential biases that could result from vaccination causing changes in behavioral patterns such as risk-seeking or willingness to get tested. Making the distribution of the real vaccines random means that different personality types won’t be overrepresented or underrepresented among either the vaccinated or unvaccinated, which would have resulted in different tendencies to get tested or seek medical attention. And double blinding removes the biases that might lead doctors and hospitals to make different reporting, coding and treatment decisions for vaccinated and unvaccinated people.
Moreover, if a trial is well designed and properly conducted it will for example require each participant to get tested regularly and on a fixed schedule, not just if a patient experiences symptoms.
Alas, as mentioned before, we simply don’t have vaccine efficacy data from well designed, properly conducted and double blind RCTs other than data from the (possibly not quite so well conducted and possibly not quite reliably double blinded) RCTs conducted by the pharmaceutical companies, or rather, the companies their hired to do that for them. And these only give us data about short-term efficacy against symptomatic infection.
So we don’t have pristine data devoid of distortions that are caused by biases in the data collection. But we can in principle still avoid many of the above-mentioned pitfalls when it comes to the handling, representation and interpretation of the data. In practice, however, very few of the reports published in the media and/or by public health institutions take care to avoid these pitfalls.
This CBC report, for example, reaches spectacular conclusions about vaccine effectiveness, but while its methodology avoids some of the pitfalls mentioned above it fails to avoid several others, which I will leave as an exercise for the readers to identify in the article:
Even academic papers will typically fail in one or more respects. And even the ones that are careful to avoid as many as possible, will typically still have to rely on imprecise estimates for some of the numbers they need to do the calculations used for their comparisons.
For example, in populations with very high vaccination rates it is very difficult to accurately estimate the population of unvaccinated people. Typically that population is estimated by taking the total number of people in that population and subtracting the number of vaccinated people from it. Because public health institutions keep track of the number of vaccinations in that area the number of vaccinated people can typically be reliably estimated, although people moving in and out of the area complicates things somewhat. But to estimate the number of unvaccinated people you need reliable estimates of the total number of people in an area. And if vaccination rates are very high, then relatively small differences in these estimates of the total population will cause relatively large differences in estimates of the unvaccinated population, and hence in the rates of infections, hospitalizations and deaths in the unvaccinated population.
To give an extreme example, in Ontario the vaccination rate of people 80 years and older is listed as 99.99%. The total population of that age group is 655,835. That implies there are only 66 unvaccinated people 80+ in the province. But suppose the total population in that age group is underestimated by a mere 1,000, or 0.15%. That means the unvaccinated population would be 1,066 instead of 66, or 16 times as high. So a small difference in the estimate of the total population means a huge difference in the unvaccinated population, and hence in the rate of infections, hospitalizations and deaths in that age group.
In conclusion, with all the data purporting to tell us something about how effective the vaccines are, we should keep in mind that those data are very much imperfect and should be treated with considerable caution. Even more so when the people involved in the generation, collection, processing and presentation of the data have a strong belief in the effectiveness of the vaccines, such that these assumptions can unintentionally introduce biases that distort the data. Let alone when they may have incentives to intentionally distort and misrepresent.
Even absent deliberate misrepresentation, in the current atmosphere reports that are relatively careless in adjusting for biases that result in a more favorable view of the vaccines will tend to face considerably less scrutiny than reports with a less favorable view.
The solution for all this lies in well designed, properly conducted and double blind randomized controlled trials. These are expensive. But governments have spent enormous amounts of money on their pandemic response in general and the vaccination program in particular. There is no possible justification for not spending a tiny fraction of that money on the only method that generates reliable and accurate data to inform this response.