The Centers for Disease Control and Prevention’s plan for who gets vaccines and in what order saved nearly as many lives and prevented nearly as many infections as a theoretically perfect rollout, according to a new mathematical model we developed to assess the rollout of covid-19 inoculations in the US.
In December 2020, with a limited number of vaccines available, the CDC had to make a hard decision: Who gets the covid-19 vaccines first? It decided to divide the US population into four groups for vaccine prioritization based on age, occupation, living condition, and known covid-19 risk factors.
Using a new model and an Iowa State University supercomputer, we compared the real-world CDC recommendations with 17.5 million possible strategies that also staggered the rollout in up to four phases. To calculate how well a vaccine allocation strategy performed, our model measured total deaths, cases, infections, and years of life lost.
We found that the CDC allocation strategy performed exceptionally well—within 4% of perfect—in all four measures.
According to our model, the CDC’s decisions to not vaccinate children initially and prioritize healthcare and other essential workers over nonessential workers were both correct. But our model also showed that giving individuals with known risk factors earlier access to vaccines would have led to slightly better outcomes.
No single rollout was able to simultaneously minimize deaths, cases, infections and years of life lost. For example, the strategy that minimized deaths led to a higher number of cases. Given these limitations, the CDC plan did a good job of balancing the four goals of vaccination and was particularly good at reducing deaths.
Why it matters
Many other studies have looked at a small number of alternative covid-19 vaccine rollouts. Our project incorporated more characteristics of the current pandemic and considered 17.5 million possible strategies. We believe this gives our results more authority.
Our model includes differences in disease severity and susceptibility to the coronavirus due to age. It also incorporates social distancing levels that change over time, as well as variable infectivity rates to account for more contagious virus strains such as the delta variant.
All this gave us the ability to accurately assess the CDC’s past decisions. But the greater value of our modeling approach lies in how it could help guide future policy.
By changing model inputs, we were able to show how optimal rollout strategies should change given different vaccine hesitancy rates and for different vaccines that can protect in different ways against infection or death. For countries that are currently planning covid-19 vaccination strategies, our model could help decision-makers develop the most effective strategies given their local resources and specifics. And even in the US, our modeling technique can inform allocation strategies for booster shots and future vaccine rollouts so healthcare administrators can make the best use of limited resources.
What still isn’t known
Any model is a simplification of reality. Our model did not account for re-infections or varying levels of vaccine hesitancy based on socioeconomic status, political ideology, or race. We also assumed that the level of hesitancy was constant over time.
Additionally, some important factors for how the coronavirus spreads—like contact rates between individuals of different age and demographic groups and the contagiousness of asymptomatic and vaccinated individuals—are still unknown. Better data on these parameters would improve the accuracy of our results.
Now that we have the model built, we can extend it. For example, we can study how waning immunity and booster shots might affect the spread of the disease. Our computer code is available to the public, and we hope it will guide healthcare policymakers in the US and around the world.
This article is republished from The Conversation under a Creative Commons license. Read the original article.