tldr; The recent study tying lack of vaccinations to traffic accidents illustrates everything that is wrong with the state of our peer-reviewed scientific process. The study uses data that do not capture any of the important behavioral factors driving one's vaccination status and propensity to traffic accidents, and yet draws conclusions as if the biases in the observational data have been sufficiently cured. The researchers ignores some findings that contradict their headline and which reveal shortfalls of their methodology. These inconvenient results suggest that their regression model did not adequately correct the biases. The model structure does not align with the hypothesized causal structure, which in any case is far too complex to be handled by a basic regression. Worse of all, the study foregrounds an entirely inappropriate analysis, on to which reporters latched. Peer reviewers somehow looked the other way.
Late last year, several readers prodded me to blog about a sensational study urging people to take Covid-19 vaccines in order to prevent traffic accidents (link). Indeed, I had already noticed the breathless headlines, and flagged the study as suspected “science for emergency use” (i.e., not the usual science).
Now that I’ve read the paper, I confirm the initial suspicion. The study is not frivolous, as these researchers made efforts to strengthen the evidence on a finding that is predictably controversial. Nevertheless, they fell too much in love with a headline grabber, and missed some red flags.
As with most Covid-19 observational studies, these researchers gained access to a set of government databases – vaccinations, Covid-19 cases, demographics, traffic accidents, etc., linked them together, selected a “test group” and a “control group”, and performed several statistical analyses, ranging from simple tests of proportions to regressions with simple adjustments. Reflecting the preference during pandemic times, this type of research is conducted purely remotely, in an arms-length manner, without any engagement with the persons or objects under study. It’s quintessential ivory-tower data mining, no use for shoe leather.
For this study, the researchers obtained “N=All” data from Ontario, Canada. They selected about 11 million people to include in their analyses, divided into those who have taken zero vaccine shots (“unvaccinated”) and those who have taken one or more vaccine shots (“vaccinated”). The unvaccinated group comprised only 16% of the analysis population at the time of the analysis.
The researchers chose to highlight their least convincing result, perhaps because it carries the biggest “effect”. This is the claim that “unvaccinated individuals [experienced] a 72% increased relative risk [of traffic crashes] compared with those vaccinated”. This claim orginates from a simple, unadjusted comparison of the rate of traffic accidents between the two groups.
This simple, unadjusted comparison is completely useless! It is valid only under the obviously false assumption that the only difference between the “test group” and the “control group” is their vaccination status. Nevertheless, reporters latched onto it like sticky glue. I like to use the term “XYopia” (i.e. X-Y myopia) to describe the fallacy of pretending that the observed outcome Y (traffic accidents) is affected only by the offered explainer X (vaccinations) and nothing else.
In fact, the observed difference in traffic accident rates could be partially or wholly due to any number of other factors. After all, this type of research study is an exercise in data mining, not a carefully designed clinical trial. In every province and country, the unvaccinated group is expected to vary from the vaccinated group, not just in their vaccination status but due to self-selection and local, discriminatory vaccination policies. Thus, any unadjusted analysis ignoring other factors is pure garbage.
Why did the researchers let this happen? It’s not like they can’t smell the stench from a mile away. This is elementary stuff taught in the first lecture on observational studies. The paper actually included several adjusted analyses. Maybe they will argue that the simple analysis was “pre-registered”, and thus according to the norms of scientific publishing, they are “obligated” to headline it. The protocol is not published so we don't know. If true, this situation exposes the Achille’s heel of using “pre-registration” to preempt bad science – since patently absurd statistical methods could be pre-approved. (Besides, in many other Covid-19 studies, including trials, the protocols were continuously "revised" up to the time of analysis, rendering pre-registration toothless.)
After adjustment, the relative risk increase dropped to about 50% (the adjustment is insufficient, as indicated below).
***
The researchers are well aware that their finding would be controversial, obviously a prime example of correlation, not necessarily causation. This is also where "causation creep" shows up. Just like the majority of such studies, the authors clearly state that "correlation is not causation", but their conclusion - "Together, the findings suggest that unvaccinated adults need to be careful indoors with other people and outside with surrounding traffic" - makes sense only under a causal interpretation of the observed correlation.
Think about the situation faced by all research on observational data. If the finding is merely an interesting correlation, there is no reason to publish it, as the correlation could be spurious. The only studies that could get published are the ones in which the researchers somehow turn that observed correlation into causation. What's strange and wrong in my opinion is the declaration that they have found only a correlation when they have convinced themselves (and peer reviewers) that they have offered sufficient evidence of causality. If that's what they believe, they should own it and say it clearly.
To strengthen their own conviction of causality, the researchers sliced and diced the data some more, and found a similar signal almost everywhere. Instead of ringing alarm bells, such unlikely consistency is interpreted as confirming the aggregate result. But the signal isn’t “everywhere”, only everywhere they looked. If an overlooked factor was driving the aggregate result, it may well drive all the subgroup analyses, with the neglected factor cutting across all the known subgroups.
For example, in Table 2, the analysts replicated the simple unadjusted comparison of rates for a variety of subgroups, such as males, younger or older age, people with cancer, etc. We learn that among those older than 65 years old, those unvaccinated have a 30% lower risk of traffic accidents than those who are vaccinated. (This finding, just like the other findings, assumes all else is equal, which is obviously false.) Applying the same reasoning these authors used to interpret the 50% overall percentage increase in risk, we’d have to advise Ontario residents above 65 years old against getting the Covid-19 vaccine as doing so may make them more vulnerable to traffic accidents! This effect remains even after adjustment.
Next, in Table 3, the researchers performed the simple comparison of vaccinated vs. unvaccinated on subsidiary outcomes. One such outcome was involvement as pedestrian (rather than driver or passenger). Amusingly, being unvaccinated increased the risk of getting hit by vehicles by 40% (p = 0.00). Moreover, not getting the vaccine more than doubled the risk of traffic accidents *at night*.
Any attempt to explain away these inconvenient findings will involve invoking some other factors e.g. maybe unvaccinated people are more likely to have a night life – such theorizing does not rescue the analytical model but merely proves that it failed to account for all relevant factors.
The wayward results were disclosed in the paper – the research team chose to look the other way, focusing on the outcomes that support their headline-grabbing conclusion. But these results demonstrate that the "adjustments" applied have failed to cure the selection biases in the observational data.
***
It comes as little surprise that the simple adjustments do not cure the selection biases. (And when such biases are not cured, they could infect any analysis.) This problem is endemic in the Covid-19 literature. Most studies - including this one - rely on basic regression models that include just main effects of various demographic variables (age, gender, location, etc.). Very few of these studies publish the actual models - almost never do we see any tables of coefficient estimates or goodness of fit statistics. Almost no one discusses the causal structure associated with the structure of their regressions. Nevertheless, they all proceed as if their adjusted models have completely cured any selection biases.
The regression model in the traffic accidents study accounted for "baseline demographic and diagnostic predictors", which are not explicitly listed. Presumably they contain the items in Table 2: age (<40, 40-64, >65), gender, urban/rural, socioeconomic status (high, medium, low), prior alcohol misuse, prior sleep apnea, prior diabetes, prior depression, prior dementia, prior hypertension, prior cancer, prior Covid-19 infection.
Missing from this list is the most obvious one: prior traffic accidents. Another obvious one is prior vaccinations. I'd also include prior traffic violations. Yet another relevant factor is prior hospitalizations or hospital contacts. I'm not sure how peer reviewers missed all these.
This last factor is clearly important - but most of the Covid-19 researchers make a ridiculous decision - that is, they excluded people with recent hospital contacts, which improperly injects a healthy person bias into all such studies. This factor should not be used as an exclusion criterion but should be an adjustment variable.
The study's appendix includes a very telling causal graph:
I'm not going to adjudicate as to whether this diagram properly captures the complex relationships between these factors, nor whether it includes all relevant factors. Just focus on the white nodes - these are the only factors/outcomes that have real-world measurements, and are represented in the regression model. Everything else is not directly observable.
The regression model ignores all the unmeasured quantities so it is inconsistent with the above diagram. And yet, they assert that all biases have been cured. Alright - they didn't really claim all biases have been cured. In fact, they compiled a long list of factors that they knowingly ignored. In describing the "limitations" of the study, they listed the following factors: distrust of government, belief in freedom, misconception of everyday risks, faith in natural protection, antipathy toward regulation, chronic poverty, exposure to misinformation, insufficient resources, other personal beliefs, political identity, negative past experiences, limited health literacy, social networks, amount of driving, vehicle factors (speed, spacing, configuration, location, weather, distance driven), access to care, risk compensation, non-reported crashes, etc.
The researchers' stance is that since they don't have data for these factors, they are left as homework exercises for future research. That regrettably is normal research practice: even though many important factors are ignored, the study is still publishable since the researchers acknowledge the "limitations". It's like an investment banker disclosing all the reasons why a tech startup would tank, while still selling the IPO stock fervently. Here is a sample from Twitter's 2013 IPO that is good for giggles:
- "Our products and services may contain undetected software errors, which could harm our business and operating results."
- "We rely on assumptions and estimates to calculate certain of our key metrics, and real or perceived inaccuracies in such metrics may harm our reputation and negatively affect our business."
- "We have incurred significant operating losses in the past, and we may not be able to achieve or subsequently maintain profitability."
- "If we fail to grow our user base, or if user engagement or ad engagement on our platform decline, our revenue, business and operating results may be harmed."
- "If we fail to effectively manage our growth, our business and operating results could be harmed."
Maybe it's just me - if I believed those risk factors (plus many others), I wouldn't buy this stock. But of course, I wouldn't buy any tech stocks since they all use similar language.
If had read the limitations of the traffic accidents study first, I'd have saved a lot of time by not reading the study.
P.S. This post has run too long, and I'll post the second part next week.
Recent Comments