Mathematical modeling of the West Africa Ebola epidemic
Abstract
As of November 2015, the Ebola virus disease (EVD) epidemic that began in West Africa in late 2013 is waning. The human toll includes more than 28,000 EVD cases and 11,000 deaths in Guinea, Liberia, and Sierra Leone, the most heavily-affected countries. We reviewed 66 mathematical modeling studies of the EVD epidemic published in the peer-reviewed literature to assess the key uncertainties models addressed, data used for modeling, public sharing of data and results, and model performance. Based on the review, we suggest steps to improve the use of modeling in future public health emergencies.
https://doi.org/10.7554/eLife.09186.001eLife digest
The outbreak of Ebola that started in West Africa in late 2013 has caused at least 28,000 illnesses and 11,000 deaths. As the outbreak progressed, global and local public health authorities scrambled to contain the spread of the disease by isolating those who were ill, putting in place infection control processes in health care settings, and encouraging the public to take steps to prevent the spread of the illness in the community. It took a massive investment of resources and personnel from many countries to eventually bring the outbreak under control.
To determine where to allocate people and resources during the outbreak, public health authorities often turned to mathematical models created by scientists to predict the course of the outbreak and identify interventions that could be effective. Many groups of scientists created models of the epidemic using publically available data or data they obtained from government officials or field studies. In some instances, the models yielded valuable insights. But with various groups using different methods and data, the models didn’t always agree on what would happen next or how best to contain the epidemic.
Now, Chretien et al. provide an overview of Ebola mathematical modeling during the epidemic and suggest how future efforts may be improved. The overview included 66 published studies about Ebola outbreak models. Although most forecasts predicted many more cases than actually occurred, some modeling approaches produced more accurate predictions, and several models yielded valuable insights. For example, one study found that focusing efforts on isolating patients with the most severe cases of Ebola would help end the epidemic by substantially reducing the number of new infections. Another study used real-time airline data to predict which traveler screening strategies would be most efficient at preventing international spread of Ebola. Furthermore, studies that obtained genomic data showed how specific virus strains were transmitted across geographic areas.
Chretien et al. argue that mathematical modeling efforts could be more useful in future pubic health emergencies if modelers cooperated more, and suggest the collaborative approach of weather forecasters as a good example to follow. Greater data sharing and the creation of standards for epidemic modeling would aid better collaboration.
https://doi.org/10.7554/eLife.09186.002Introduction
On March 23, 2014, the Ministry of Health Guinea notified the World Health Organization (WHO) of a rapidly evolving outbreak of Ebola virus disease (EVD), now believed to have begun in December 2013. The epidemic spread through West Africa and reached Europe and the United States. As of November 4, 2015, WHO reported more than 28,000 cumulative cases and 11,000 deaths in Guinea, Liberia, and Sierra Leone, where transmission had been most intense (World Health Organization, 2016).
As the emergency progressed, researchers developed mathematical models of the epidemiological dynamics. Modelers have assessed ongoing epidemics previously, but the prominence of recent EVD work, enabled by existing research programs for infectious disease modeling (National Institutes of Health, 2016a; National Institutes of Health, 2016b) and online availability of EVD data via WHO (World Health Organization, 2016), Ministries of Health of affected countries, or modelers who transcribed and organized public WHO or Ministry of Health data (Rivers C) may be unprecedented. The efforts for this outbreak have been numerous and diverse, with major media incorporating modeling results in many pieces throughout the outbreak. U.S. Government decision making has benefited from modeling results at key moments during the response (Robinson R).
We draw on this vigorous response of the epidemiological modeling community to the EVD epidemic to review (Moher et al., 2009) the application of modeling to public health emergencies, and identify lessons to guide the modeling response to future emergencies.
Results
Overview of modeling applications
We identified 66 publications meeting inclusion criteria (Figure 1).

Literature search flow.
https://doi.org/10.7554/eLife.09186.003Models addressed 6 key uncertainties about the EVD epidemic: transmissibility, typically represented by the reproduction number (R, the average number of people each infected person infects; assessed in 41 publications); effectiveness of various interventions that had been or might be implemented (in 29 publications); epidemic forecast (in 29 publications); regional or international spreading patterns or risk (in 15 publications); phylogenetics of EVD viruses (in 9 publications); and feasibility of conducting vaccine trials in West Africa (in 2 publications) (Table 1, Supplementary file 1).
Overview of modeling publications on the 2013-present EVD epidemic.
Ref. | Date of latest EVD data | Date published | EVD data was pre-existing and public | Uncertainties addressed | |||||
---|---|---|---|---|---|---|---|---|---|
R | Interventions | Forecast | Spread | Phylogenetics | Clinical trials | ||||
Baize et al., 2014 | 3/20/14 | 4/16/14 | No | * | |||||
Dudas and Rambaut, 2014 | 3/20/14 | 5/2/14 | Yes | * | |||||
Alizon et al., 2014 | 6/18/14 | 12/13/14 | Yes | * | * | ||||
Gire et al., 2014 | 6/18/14 | 8/28/14 | No | * | * | ||||
Stadler et al., 2014 | 6/18/14 | 10/6/14 | Yes | * | |||||
Volz and Pond, 2014 | 6/18/14 | 10/24/14 | Yes | * | |||||
Pandey et al., 2014 | 8/7/14 | 10/30/14 | Yes | * | * | * | |||
Gomes et al., 2014 | 8/9/14 | 9/2/14 | Yes | * | * | * | |||
Valdez et al., 2015 | 8/15/14 | 7/20/15 | Yes | * | * | * | * | ||
Merler et al., 2015 | 8/16/14 | 1/7/15 | Yes | * | * | * | * | ||
Rainisch et al., 2015 | 8/16/14 | 2/18/15 | Yes | * | |||||
Althaus, 2014 | 8/20/14 | 9/2/14 | Yes | * | |||||
Fisman et al., 2014 | 8/22/14 | 9/8/14 | Yes | * | * | ||||
Nishiura and Chowell, 2014 | 8/26/14 | 9/11/14 | Yes | * | * | ||||
Poletto et al., 2014 | 8/27/14 | 10/23/14 | Yes | * | * | ||||
Meltzer et al., 2014 | 8/28/14 | 9/26/14 | Yes | * | * | * | |||
Agusto et al., 2015 | 8/29/14 | 4/23/15 | Yes | * | * | ||||
Althaus, 2015 | 8/31/14 | 4/19/15 | Yes | * | |||||
Scarpino et al., 2014 | 8/31/14 | 12/15/14 | Yes | * | |||||
Weitz and Dushoff, 2015 | 8/31/14 | 3/4/15 | Yes | * | * | ||||
Drake et al., 2015 | 9/2/14 | 10/30/14 | Yes | * | * | * | |||
Towers et al., 2014 | 9/8/14 | 9/18/14 | Yes | * | * | ||||
Bellan et al., 2014 | 9/14/14 | 10/14/14 | Yes | * | |||||
Chowell et al., 2015 | 9/14/14 | 1/19/15 | Yes | * | |||||
Cooper et al., 2015 | 9/14/14 | 4/14/15 | Yes | * | |||||
Read et al., 2015 | 9/14/14 | 11/12/14 | Yes | * | * | ||||
WHO Ebola Response Team, 2014 | 9/14/14 | 9/23/14 | No | * | * | * | |||
Faye et al., 2015 | 9/16/14 | 1/23/15 | No | * | * | ||||
Bogoch et al., 2015 | 9/21/14 | 10/21/14 | Yes | * | * | ||||
Yamin et al., 2014 | 9/22/14 | 10/28/14 | No | * | * | ||||
Lewnard et al., 2014 | 9/23/14 | 10/24/14 | Yes | * | * | * | |||
Webb et al., 2015 | 9/23/14 | 1/30/15 | Yes | * | * | * | |||
Shaman et al., 2014 | 9/28/14 | 10/27/14 | Yes | * | * | ||||
Chowell et al., 2014 | 10/1/14 | 11/20/14 | Yes | * | * | ||||
Fasina et al., 2014 | 10/1/14 | 10/9/14 | Yes | * | |||||
Khan et al., 2015 | 10/1/14 | 2/24/15 | Yes | * | |||||
Rivers et al., 2014 | 10/5/14 | 10/16/14 | Yes | * | * | * | |||
Xia et al., 2015 | 10/7/14 | 9/8/15 | Yes | * | * | ||||
Majumder et al., 2014 | 10/11/14 | 4/28/15 | Yes | * | * | ||||
Kiskowski, 2014 | 10/15/14 | 11/13/14 | Yes | * | * | ||||
Fisman and Tuite, 2014 | 10/18/14 | 11/21/14 | Yes | * | * | * | |||
Althaus et al., 2015 | 10/20/14 | 1/15/15 | Yes | * | * | * | |||
Simon-Loriere et al., 2015 | 10/25/14 | 6/24/15 | No | * | |||||
Rainisch et al., 2015 | 10/31/14 | 6/16/15 | Yes | * | |||||
Fast et al., 2015 | 11/1/14 | 5/15/15 | Yes | * | |||||
Kucharski et al., 2015 | 11/1/14 | 2/18/15 | Yes | * | * | * | |||
Tong et al., 2015 | 11/11/14 | 5/13/15 | No | * | * | ||||
Hoenen et al., 2015 | 11/21/14 | 3/26/15 | No | * | |||||
Cope et al., 2014 | 12/3/14 | 12/10/14 | Yes | * | * | ||||
White et al., 2014 | 12/3/14 | 1/30/15 | Yes | * | * | * | |||
WHO Ebola Response Team, 2015 | 12/14/14 | 12/24/14 | No | * | * | ||||
Chowell et al., 2014 | 12/17/14 | 1/21/15 | Yes | * | |||||
Siettos et al., 2014 | 12/21/14 | 3/9/15 | Yes | * | * | ||||
Park et al., 2015 | 12/26/14 | 6/18/15 | No | * | |||||
Nadhem and Nejib, 2015 | 12/30/14 | 6/14/15 | Yes | * | |||||
Camacho et al., 2015 | 1/18/15 | 2/10/15 | Yes | * | * | ||||
Carroll et al., 2015 | 1/31/15 | 6/17/15 | No | * | * | ||||
Bellan et al., 2015 | 2/9/15 | 4/15/15 | Yes | * | * | ||||
Barbarossa et al., 2015 | 2/13/15 | 7/21/15 | Yes | * | * | * | |||
Kugelman et al., 2015 | 2/14/15 | 6/12/15 | No | * | |||||
Cleaton et al., 2015 | 2/28/15 | 9/3/15 | Yes | * | |||||
Wang and Zhong, 2015 | 3/18/15 | 3/24/15 | Yes | * | |||||
Toth et al., 2015 | 3/31/15 | 7/14/15 | Yes | * | * | ||||
Dong et al., 2015 | 4/3/15 | 9/5/15 | Yes | * | * | ||||
Browne et al., 2015 | 4/12/15 | 5/14/15 | Yes | * | * | ||||
Zinszer et al., 2015 | 5/13/15 | 9/1/15 | Yes | * |
The number of publications with models to estimate R increased rapidly early in the epidemic, along with those including intervention, forecasting, and regional and international spread models; the growth rate of publications with phylogenetic modeling applications and clinical trial models increased later in the epidemic (Figure 2).

Cumulative number of modeling applications by date of most recent EVD data used.
The figure includes 125 modeling applications across the 66 publications.
Of the 125 models reported across the studies, 74% included mechanistic assumptions about disease transmission (e.g., compartmental, agent-based, or phylogenetic models), while 26% were purely phenomenological (Supplementary file 2).
Data sources
For 54 (82%) of the 66 publications, the only EVD data used was pre-existing and publicly-available (Table 1). Typically, these were aggregate case data posted online by the WHO or affected countries, or Ebola virus genetic data released previously during the epidemic. Twelve studies used original EVD epidemiological data (Baize et al., 2014; WHO Ebola Response Team, 2014; 2015; Faye et al., 2015; Yamin et al., 2014) or genomic data (Baize et al., 2014; Gire et al., 2014; Simon-Loriere et al., 2015; Tong et al., 2015; Hoenen et al., 2015; Park et al., 2015; Carroll et al., 2015; Kugelman et al., 2015).
Examples of additional data used for some modeling applications include official reports of social mobilization efforts (Fast et al., 2015), media reports of case clusters (Cleaton et al., 2015), media reports of events that may curtail or aggravate transmission (Majumder et al., 2014), and international air travel data (Gomes et al., 2014; Poletto et al., 2014; Read et al., 2015; Bogoch et al., 2015; Rainisch et al., 2015; Cope et al., 2014). Several studies incorporated spatial data on EVD cases into models of regional EVD spread (Gire et al., 2014; Merler et al., 2015; Rainisch et al., 2015; Tong et al., 2015; Carroll et al., 2015; Zinszer et al., 2015).
Data and results sharing
Of the 12 studies that collected original EVD data, 9 released those data before or at the time of publication (8 with Ebola virus genetic data deposited in GenBank (Baize et al., 2014; Gire et al., 2014; Simon-Loriere et al., 2015; Tong et al., 2015; Hoenen et al., 2015; Park et al., 2015; Carroll et al., 2015; Kugelman et al., 2015) and 1 with detailed epidemiological data in the online publication (Yamin et al., 2014). Many publications used results from the WHO Ebola Response Team investigations (WHO Ebola Response Team, 2014; 2015) (for example, estimates of the generation time, case fatality rate, or other epidemiological parameters as model inputs), but the detailed epidemiological data from these studies, to date, are not publicly available.
Accumulation of shared EVD data over successive studies was evident especially in the phylogenetic analyses. For example, all phylogenetic studies published after release of the initial Ebola virus sequences by (Baize et al., 2014) (Guinea) and (Gire et al., 2014) (Sierra Leone) incorporated those sequence data.
Across all studies, the publication lag (defined as date of most recent EVD data used to date of online publication) was almost 3 months (median [interquartile range] = 85 [30–157] days). The lag varied across modeling applications, and was considerably shorter in studies that included models to estimate R (median = 58 days for publications with R estimation versus 118 days for others) or to forecast (median = 50 versus 125 days) (Figure 3).

Publication lag by type of modeling application.
The vertical red and turquoise lines indicate the median lag for publications including and not including, respectively, the type of modeling application.
Lags were longest for studies with phylogenetic and clinical trials applications (median = 125 and 108 days, respectively), although there were fewer publications with these models.
Modeling results: R and forecasts
Forty-one publications characterized epidemic dynamics using epidemiological (N=36), genomic (N=4), or news report data (N=1). Twenty-four of these provided estimates of the basic reproduction number (R0) for Guinea, Liberia, Sierra Leone, or West Africa, using epidemiological or genomic data (Figure 4, Supplementary file 3).

R0 estimates by type of model input data.
Aggregate, case counts released by the WHO or Ministries of Health; Line-level, individual-level data from epidemiological investigations; Genomic, Ebola virus sequence data. The Figure excludes an outlier estimate of 8.33 for Sierra Leone (Fisman et al., 2014).
There were 16 country-specific estimates of R0 for Guinea, Liberia, or Sierra Leone that used EVD epidemiological data (aggregate or line-level) and provided 95% confidence or credible intervals (CIs). Median CI width was about 85% smaller for models that used cumulative EVD counts (N=11 models in 5 publications) than for models that used disaggregated EVD case data, such as weekly counts (N=5 models in 3 publications) (Figure 5).

R0 estimates and CIs by type of epidemiological input data.
Disaggregated data typically were weekly counts. Top row: Vertical lines indicate 95% CIs. Bottom row: Horizontal bars indicate median CI width.
Although CIs were also narrower for models when deterministic rather than stochastic methods were used to estimate parameter uncertainty, all of the deterministic results came from a single study (Figure 6).

R0 estimates and CIs by model fitting method.
Top row: Vertical lines indicate 95% CIs. Bottom row: Horizontal bars indicate median CI width.
Fifteen publications provided numerical forecasts of cumulative EVD incidence for West African countries. Of 22 models that assumed no additional response measures beyond those implemented at the time (i.e., 'status quo' assumptions), 18 overestimated the future number of cases (Figure 7, Supplementary file 4).

Accuracy of cumulative incidence forecasts.
Accuracy is shown as the ratio of predicted incidence to incidence subsequently reported by the WHO. 'Dampening' refers to various approaches to restrict the growth of forecasted incidence over time. Top row: Accuracy by date of forecast. Bottom row: Accuracy by forecast lead time ('Horizon'). The Figure excludes one forecast with horizon > 1 year (Fisman and Tuite, 2014).
In multivariate analysis, forecast error was lower for forecasts made later in the outbreak (14% reduction in mean absolute percentage error [MAPE] per week, P<0.001), higher for forecasts with longer time horizons (29% increase in MAPE per week, P<0.01), and lower for forecasts that used decay terms, spatially heterogeneous contact patterns, or other methods that served to constrain projected incidence growth (90% reduction in MAPE, P<0.01). Country and number of parameters in the model were not statistically significant predictors of forecast accuracy.
Discussion
We identified 66 modeling publications during approximately 18 months of the EVD response that assessed trends in the intensity of transmission, effectiveness of control measures, future case counts, regional and international spreading risk, Ebola virus phylogenetic relationships and recent evolutionary dynamics, and feasibility of clinical trials in West Africa. We found a heavy dependence on public data for EVD modeling, and identified factors that might have influenced model performance. To our knowledge, this review is one of the most comprehensive assessments of mathematical modeling applied to a single real-world public health emergency.
An important caveat of our review is that it only captures published results. We are aware of additional EVD epidemiological investigations and modeling not yet published. Some modelers providing direct support to operational response efforts have not published results, possibly because of operational demands.
Also, we could not account comprehensively for the sources of variation across studies. For example, studies that estimated R0 using the same data sources at about the same time reported varied results. Such variation may, in part, reflect the problem of identifiability, with different R0 estimates possible for models that perform equally well depending on other parameter values (Weitz and Dushoff, 2015). Ideally, an investigation into this heterogeneity would include implementation of models in a common testing environment.
Our review suggests several possible steps for improving the application of epidemiological modeling during public health emergencies. First, agreement on community best practices could improve the quality of modeling support to decision-makers. For example, our analysis is consistent with simulation studies showing underestimation of uncertainty in estimating R0 with cumulative (as opposed to disaggregated) incidence data, and supports the recommendation to use disaggregated data and stochastic models (King et al., 2015). Additionally, incidence forecasts provided reasonable prospective estimates several weeks forward in time during the initial phase; however, given available data and methodologies these forecasts became progressively more inaccurate as they projected dynamics beyond several weeks. Validation of incidence forecasts against other relevant data, such as hospital admissions and contacts identified, also could provide evidence that the assumptions are sound.
The 2014 onwards ebola outbreak in West Africa clearly highlights the need for a better understanding of how increasing awareness of severe infections within a community decreases their transmissibility even in the absence of specific interventions. Advancing methodological approaches to capture this effect, such as dampening approaches, might help account for behavioral changes, interventions, contact heterogeneity, or other factors that can be expected in a public health emergency which likely will improve forecasting accuracy. Establishing best practices within the community will allow decision-makers the ability to more quickly accept methodologies and results that have been generated via these best practices. Hence, decisions based on these results can happen more quickly.
Second, modeling coordination could facilitate direct comparison of modeling results, identifying issues on which diverse approaches agree and areas of greater uncertainty. Epidemiological modelers might learn from comparison initiatives in modeling of influenza (Centers for Disease Control and Prevention, 2013), dengue (US Department of Commerce), and HIV (HIV modeling consortium); and in other fields such as climate forecasting Intergovernmental Panel on Climate Change, 2010). For epidemiological application, an ensemble approach should preserve methodological diversity to exploit the full range of state-of-the-art modeling methods, but include enough standardization to enable cross-model comparison. Establishing an initial architecture for a coordinated, ensemble effort now could assist the response to EVD, and future public health emergencies.
Perhaps most importantly, outbreak modeling efforts would be much more fruitful if data and analytical results could be made available more quickly to all interested parties (Yozwiak et al., 2015). The publication timelines for academic journals typically will not be consistent with decision-making needs during public health emergencies like the EVD epidemic, where the epidemiological situation was highly dynamic and the usefulness of data and forecasts time-constrained. Establishing mechanisms for modelers without special access to the official epidemiological teams to share interim results would expand the evidence base for response decision-making. Ideally, data should be made available online in machine-readable form to facilitate use in analyses. Modelers and other analysts expended enormous effort during the EVD epidemic transcribing data posted online in pdf documents.
New norms for data-sharing during public health emergencies (World Health Organization, 2015) would remove the most obvious hurdle for model comparison. The current situation where groups either negotiate bilaterally with individual countries or work exclusively with global health and development agencies is understandable, but highly ineffective. The EVD outbreak highlights again – after the 2003 Severe Acute Respiratory Syndrome epidemic and 2009 influenza A (H1N1) pandemic – that an independent, well-resourced global data observatory could greatly facilitate the public health response in many ways, not least of which would be the enablement of rapid, high quality, and easily comparable disease-dynamic studies.
Materials and methods
For this review, we adapted the PRISMA methodology (Moher et al., 2009) to identify quantitative modeling studies of the 2013-present West Africa EVD epidemic. We searched PubMed on September 24, 2015, for publications in English since December 1, 2013, using the term ‘Ebola’ in any field. We reviewed all returned abstracts and selected ones for confirmatory, full-text review that mentioned use of quantitative models to characterize or predict epidemic dynamics or evaluate interventions. We included studies that met this criterion in full-text review.
We excluded studies of clinical prediction models, viral or physiological function models, ecological niche models, animal reservoir models, and publications that did not use data from the 2013-present West Africa EVD epidemic.
For included publications, we recorded the geographic settings, date of most recent EVD data used and date of publication, type of EVD data used, questions the models addressed, modeling approaches, and key results, including estimates of the basic reproduction number (R0) and forecasts of future EVD incidence provided in the main text of the publications. To assess forecast accuracy, we compared predictions of models made under ‘status quo’ assumptions (i.e., without explicit inclusion of additional interventions or behavioral changes) to EVD incidence data subsequently released by the WHO (World Health Organization, 2016), using the WHO figures dated soonest after the forecast target date.
References
-
Ebola superspreadingThe Lancet. Infectious Diseases 15:507–508.https://doi.org/10.1016/S1473-3099(15)70135-0
-
Emergence of zaire ebola virus disease in guineaThe New England Journal of Medicine 371:1418–1425.https://doi.org/10.1056/NEJMoa1404505
-
Statistical power and validity of ebola vaccine trials in sierra leone: a simulation study of trial design and analysisThe Lancet. Infectious Diseases 15:703–710.https://doi.org/10.1016/S1473-3099(15)70139-8
-
Modeling contact tracing in outbreaks with application to ebolaJournal of Theoretical Biology 384:33–49.https://doi.org/10.1016/j.jtbi.2015.08.004
-
Modelling the effect of early detection of ebolaThe Lancet. Infectious Diseases 15:148–149.https://doi.org/10.1016/S1473-3099(14)71084-9
-
Characterizing ebola transmission patterns based on internet news reportsClinical Infectious Diseases civ748.https://doi.org/10.1093/cid/civ748
-
Evaluation of ebola spreading in west africa and decision of optimal medicine delivery strategies based on mathematical modelsInfection, Genetics and Evolution 36:35–40.https://doi.org/10.1016/j.meegid.2015.09.003
-
Ebola cases and health system demand in liberiaPLoS Biology 13:e1002056.https://doi.org/10.1371/journal.pbio.1002056
-
Chains of transmission and control of ebola virus disease in conakry, guinea, in 2014: an observational studyThe Lancet. Infectious Diseases 15:320–326.https://doi.org/10.1016/S1473-3099(14)71075-8
-
Estimating the basic reproductive ratio for the ebola outbreak in liberia and sierra leoneInfectious Diseases of Poverty 4:13.https://doi.org/10.1186/s40249-015-0043-3
-
Avoidable errors in the modelling of outbreaks of emerging pathogens, with special reference to ebolaProceedings of the Royal Society B: Biological Sciences 282:20150347.https://doi.org/10.1098/rspb.2015.0347
-
Evaluation of the benefits and risks of introducing ebola community care centers, sierra leoneEmerging Infectious Diseases 21:393–399.https://doi.org/10.3201/eid2103.141892
-
Monitoring of ebola virus makona evolution through establishment of advanced genomic capability in liberiaEmerging Infectious Diseases 21:1135–1143.https://doi.org/10.3201/eid2107.150522
-
Dynamics and control of ebola virus transmission in montserrado, liberia: a mathematical modelling analysisThe Lancet. Infectious Diseases 14:1189–1195.https://doi.org/10.1016/S1473-3099(14)70995-8
-
Estimating the future number of cases in the ebola epidemic--liberia and sierra leone, 2014-2015Morbidity and Mortality Weekly Report. Surveillance Summaries (Washington, D.C. : 2002) 63 Suppl 3:1–14.
-
Early transmission dynamics of ebola virus disease (eVD), West Africa, march to august 2014Euro Surveillance : Bulletin Européen Sur Les Maladies Transmissibles = European Communicable Disease Bulletin 19:20894.https://doi.org/10.2807/1560-7917.ES2014.19.36.20894
-
Estimating ebola treatment needs, united statesEmerging Infectious Diseases 21:1273–1275.https://doi.org/10.3201/eid2107.150286
-
Regional spread of ebola virus, west africa, 2014Emerging Infectious Diseases 21:444–447.https://doi.org/10.3201/eid2103.141845
-
Estimates of outbreak risk from new introductions of ebola with immediate and delayed transmission controlEmerging Infectious Diseases 21:1402–1408.https://doi.org/10.3201/eid2108.150170
-
Ebola outbreak in west africa: real-time estimation and multiple-wave predictionMathematical Biosciences and Engineering 12:1055–1063.https://doi.org/10.3934/mbe.2015.12.1055
-
Ebola virus disease in West Africa--the first 9 months of the epidemic and forward projectionsThe New England Journal of Medicine 371:1481–1495.https://doi.org/10.1056/NEJMoa1411100
-
West African ebola epidemic after one year--slowing but not yet under controlThe New England Journal of Medicine 372:584–587.https://doi.org/10.1056/NEJMc1414992
-
Modeling the transmission dynamics of ebola virus disease in liberiaScientific Reports 5:13857.https://doi.org/10.1038/srep13857
-
Effect of ebola progression on transmission and control in liberiaAnnals of Internal Medicine 162:11–17.https://doi.org/10.7326/M14-2255
-
The velocity of ebola spread in parts of west africaThe Lancet. Infectious Diseases 15:1005–1007.https://doi.org/10.1016/S1473-3099(15)00234-0
Decision letter
-
Mark JitReviewing Editor; London School of Hygiene & Tropical Medicine, and Public Health England, United Kingdom
In the interests of transparency, eLife includes the editorial decision letter and accompanying author responses. A lightly edited version of the letter sent to the authors after peer review is shown, indicating the most substantive concerns; minor comments are not usually included.
Thank you for submitting your work entitled "Improving epidemiological modeling coordination for Ebola and other public health emergencies" for peer review at eLife. Your submission has been favorably evaluated by Prabhat Jha (Senior editor), and three reviewers, one of whom (Mark Jit) is a member of our Board of Reviewing Editors. Another of the reviewers, Marco Ajelli, has also agreed to reveal their identity.
The reviewers have discussed the reviews with one another and the Reviewing editor has drafted this decision to help you prepare a revised submission.
The reviewers all agreed that this is a timely, comprehensive and useful review of modelling studies that have described and/or forecasted the Ebola epidemic in West Africa. We all enjoyed reading the paper, and agreed that it made some important points about good modelling practice in general. In particular, we agree that data sharing and complete access to best available epidemiological data (especially patient databases) for all modelling groups is of paramount importance in the case of future outbreaks of infectious diseases.
We believe that the way the information and discussion was presented could be improved, and by doing so would enhance the influence that the paper may have. In particular:
1) More detail could be provided on how Ebola modelling was actually used to support decision making both internationally and for individual national governments. Many of the recommendations in the Discussion are predicated on the premise that the modelling results were actually useful to inform policy (or if they weren't, could be made useful by improving the modelling process). Some historical discussion on what happened in practice would help here. Currently only a single US-centric reference is provided.
2) The paper needs a proper Materials and methods section providing details about the search terms/strategy, inclusion/exclusion criteria and information extracted from the papers. Given that this is a review, compliance with PRISMA guidelines would normally be expected. If a non-standard approach to searching was used, this needs to be justified and the method explained.
3) Each section of the paper seems to contain a mixture of (i) data directly extracted from the reviewed papers, (ii) direct inferences from these data, (iii) more general discussion points around the inferences and (iv) broader discussion (e.g. recommendations for standardisation and funding for future outbreak response). While each section is helpful, the way the paper is written makes it difficult to distinguish the provenance of the different points made. For instance, there are statements like "Modelers welcomed the release of these data, but expended considerable effort curating them for analysis", "accurate epidemic forecasting for Ebola proved to be quite challenging, especially for long time horizons, because the behavior of people infected and at risk was unpredictable", "incomplete information from the field can make it difficult to determine if failure to end the epidemic reflects sub-optimal implementation of effective measures, or over-optimistic assumptions about their effectiveness" where it is unclear whether these are statements drawn from the reviewed papers, or based on the authors' own experiences.
Using a traditional separation into Results and Discussion may also help this as most scientists are used to the idea of drawing immediate inferences from data, and then suggesting broader implications of these inferences. The Discussion section may be lengthy and structured if necessary. The broader discussion around funding etc. which is not really an extrapolation from the data (however welcome the points may be) should probably be given a section of its own.
4) Table 2 shows a very large variation in the baseline scenario forecast, spanning orders of magnitude. The authors seem to suggest that this stems from different assumptions used about epidemic growth forecasting e.g. random mixing, incidence decay or spatial/sociodemographic structuring. It would be useful if you could explicitly structure the presentation of results to show whether these different assumptions had an effect on the outbreak size. This should at minimum be done in the visual presentation but if some sort of meta-analysis/meta-regression could help that would be even better.
5) Additional information about the spatial structure (if any) of the models reviewed would be useful. Spatial structure in these models ranges from homogeneous (non-spatial models) to detailed agent-based models such as the one developed by Merler et al. Lancet ID.
6) Some models employed single streams of data (e.g., all case counts by country) while other models employed multiple data streams (e.g., all cases and cases among health-care workers such as the multi-type branching process model by Drake et al.). This would be useful to highlight.
7) The paper makes the point that there are differences between the results of various modelling groups because of different methodologies used. However, in some cases modelling methodology has been improperly used, and this would be useful to highlight. In particular the following areas could be given attention:
a) Parameter estimation. In Box 1, you have underlined that the different models analyzed in this review have a remarkably different number of "unknown" parameters to be estimated. This represents a very important topic that could be given more space in this manuscript). For example:
A clear distinction should be made between estimated and imputed parameters. For instance the number of estimated parameters in Pandey et al. (2014) is 4, while the other parameters are somehow imputed from the literature. That said, still the 4 unknown parameters estimated in Pandey et al. (2014) are not univocally identifiable by using as input the dataset considered in that work.
In Rivers et al. (2014) about 10 (surely not all unequivocally identifiable) parameters have been estimated on the basis of case counts only.
b) Uncertainty. A good modelling practice (which is not always followed in the reviewed papers) is the assessment of the uncertainty due to i) the stochasticity of the infection transmission process, and ii) the uncertainty in parameter estimates. In particular, as regards the uncertainty in parameter estimates, methods providing estimates of the parameters distribution (e.g., MCMC) should be preferred with respect to methods providing point estimate of the parameter values only (e.g., MSE, ML). This would avoid (or at least limit) the spread in the community of biased findings never confirmed on the field or by subsequent studies (e.g., the estimate that the main driver of the Ebola epidemic was the transmission during traditional burials found by Pandey et al. (2014) which was based on a point estimate made through a MSE fitting procedure).
c) Model validation. Another major point is that most (though not all) the reviewed modelling studies lack in model validation against other epidemiologically relevant quantities (e.g., ETU admissions, number of individuals included in contact tracing). Validation against other quantities is a key indicator to understand whether the model is able to explain the reason of an observed pattern of infection spread or it is just able to reproduce case counts "by chance". In fact, although it is welcome to show that the model is in agreement with the next few incidence points after stopping model calibration, modelers should not limit model validation to that.
8) It would also be good to clarify that the reviewed modelling studies can be divided in two categories: the ones aimed at explaining the observed pattern, for instance by accounting for actual data on intervention measures, and those that do not provide any explanation for the observed pattern (e.g., the model in Fisman, Khoo and Tuite (2014)). In fact, the point here is not to fit some data points, but to understand the mechanisms behind the observed infection spread as their knowledge is key for predictions and preparedness.
9) The importance for modelling groups of having access to online epidemiological data in an easy to export format (e.g., CSV file) could be highlighted. In fact, the Ministries of Health of the three most affected countries provided PDF files only, some of which were images, thus implying remarkable efforts by modelers to readily use such datasets. Moreover, the case of Guinea was even worse as, at the beginning of the epidemic, there was not an official website of the Guinean Ministry of Health with their reports uploaded.
https://doi.org/10.7554/eLife.09186.015Author response
1) More detail could be provided on how Ebola modelling was actually used to support decision making both internationally and for individual national governments. Many of the recommendations in the Discussion are predicated on the premise that the modelling results were actually useful to inform policy (or if they weren't, could be made useful by improving the modelling process). Some historical discussion on what happened in practice would help here. Currently only a single US-centric reference is provided.
We recast the manuscript as more of a standard scoping literature review, with less emphasis on anecdotal evidence of the practical utility of modeling during the outbreak. We mostly have only our personal experiences to draw on for such evidence; the revision draws much more heavily on the peer-reviewed literature (please see response to Comment 2 below).
2) The paper needs a proper Materials and methods section providing details about the search terms/strategy, inclusion/exclusion criteria and information extracted from the papers. Given that this is a review, compliance with PRISMA guidelines would normally be expected. If a non-standard approach to searching was used, this needs to be justified and the method explained.
We revised the manuscript as a PRISMA-compliant literature review.
3) Each section of the paper seems to contain a mixture of (i) data directly extracted from the reviewed papers, (ii) direct inferences from these data, (iii) more general discussion points around the inferences and (iv) broader discussion (e.g. recommendations for standardisation and funding for future outbreak response). While each section is helpful, the way the paper is written makes it difficult to distinguish the provenance of the different points made. For instance, there are statements like "Modelers welcomed the release of these data, but expended considerable effort curating them for analysis", "accurate epidemic forecasting for Ebola proved to be quite challenging, especially for long time horizons, because the behavior of people infected and at risk was unpredictable", "incomplete information from the field can make it difficult to determine if failure to end the epidemic reflects sub-optimal implementation of effective measures, or over-optimistic assumptions about their effectiveness" where it is unclear whether these are statements drawn from the reviewed papers, or based on the authors' own experiences. Using a traditional separation into Results and Discussion may also help this as most scientists are used to the idea of drawing immediate inferences from data, and then suggesting broader implications of these inferences. The Discussion section may be lengthy and structured if necessary. The broader discussion around funding etc. which is not really an extrapolation from the data (however welcome the points may be) should probably be given a section of its own.
We agree that the original organization was non-standard and confusing. The revision is structured as a standard research paper, with Introduction, Materials and methods, Results, and Discussion. The broader implications and recommendations are included in the Discussion.
4) Table 2 shows a very large variation in the baseline scenario forecast, spanning orders of magnitude. The authors seem to suggest that this stems from different assumptions used about epidemic growth forecasting e.g. random mixing, incidence decay or spatial/sociodemographic structuring. It would be useful if you could explicitly structure the presentation of results to show whether these different assumptions had an effect on the outbreak size. This should at minimum be done in the visual presentation but if some sort of meta-analysis/meta-regression could help that would be even better.
We replaced the table with a figure that identifies several factors that may have influenced forecasting results in a clearer and more compelling way. Figure 6 shows the association between more accurate forecasts and earlier forecasts, shorter time horizons, and approaches that constrained predicted epidemic growth (for which we suggest the term “dampening”). We also performed a regression to quantify these effects, reported in the Results.
5) Additional information about the spatial structure (if any) of the models reviewed would be useful. Spatial structure in these models ranges from homogeneous (non-spatial models) to detailed agent-based models such as the one developed by Merler et al. Lancet ID.
We identified modeling studies that incorporated spatial data (see Results, subsection “Data sources”).
6) Some models employed single streams of data (e.g., all case counts by country) while other models employed multiple data streams (e.g., all cases and cases among health-care workers such as the multi-type branching process model by Drake et al.). This would be useful to highlight.
We added two paragraphs in the Results section on sources of data (Results, subsection “Data sources”) and hope this addresses the spirit of the excellent comment, which we interpret as a call for more thorough characterization of the types of model input data.
7) The paper makes the point that there are differences between the results of various modelling groups because of different methodologies used. However, in some cases modelling methodology has been improperly used, and this would be useful to highlight. In particular the following areas could be given attention: a) Parameter estimation. In Box 1, you have underlined that the different models analyzed in this review have a remarkably different number of "unknown" parameters to be estimated. This represents a very important topic that could be given more space in this manuscript). For example: A clear distinction should be made between estimated and imputed parameters. For instance the number of estimated parameters in Pandey et al. (2014) is 4, while the other parameters are somehow imputed from the literature. That said, still the 4 unknown parameters estimated in Pandey et al. (2014) are not univocally identifiable by using as input the dataset considered in that work. In Rivers et al. (2014) about 10 (surely not all unequivocally identifiable) parameters have been estimated on the basis of case counts only.
We provided the number of parameters for models estimating R, and distinguished between estimated and imputed values (Supplementary file 3).
b) Uncertainty. A good modelling practice (which is not always followed in the reviewed papers) is the assessment of the uncertainty due to i) the stochasticity of the infection transmission process, and ii) the uncertainty in parameter estimates. In particular, as regards the uncertainty in parameter estimates, methods providing estimates of the parameters distribution (e.g., MCMC) should be preferred with respect to methods providing point estimate of the parameter values only (e.g., MSE, ML). This would avoid (or at least limit) the spread in the community of biased findings never confirmed on the field or by subsequent studies (e.g., the estimate that the main driver of the Ebola epidemic was the transmission during traditional burials found by Pandey et al. (2014) which was based on a point estimate made through a MSE fitting procedure).
We added a detailed analysis of how use of cumulative versus disaggregated input data, and deterministic versus stochastic approaches to parameter estimation, may have influenced the reported uncertainty of R estimates (see Results, subsection “Modeling results: R and forecasts” as well as Figure 4 and Figure 5).
c) Model validation. Another major point is that most (though not all) the reviewed modelling studies lack in model validation against other epidemiologically relevant quantities (e.g., ETU admissions, number of individuals included in contact tracing). Validation against other quantities is a key indicator to understand whether the model is able to explain the reason of an observed pattern of infection spread or it is just able to reproduce case counts "by chance". In fact, although it is welcome to show that the model is in agreement with the next few incidence points after stopping model calibration, modelers should not limit model validation to that.
We added this excellent point to the Discussion (paragraph four):
“Validation of incidence forecasts against other relevant data, such as hospital admissions and contacts identified, also could provide evidence that the assumptions are sound.”
8) It would also be good to clarify that the reviewed modelling studies can be divided in two categories: the ones aimed at explaining the observed pattern, for instance by accounting for actual data on intervention measures, and those that do not provide any explanation for the observed pattern (e.g., the model in Fisman, Khoo and Tuite (2014)). In fact, the point here is not to fit some data points, but to understand the mechanisms behind the observed infection spread as their knowledge is key for predictions and preparedness.
We added this distinction to our characterization of the models (see Results, subsection “Overview of modeling applications”):
“Of the 125 models reported across the studies, 74% included mechanistic assumptions about disease transmission (e.g., compartmental, agent-based, or phylogenetic models), while 26% were purely phenomenological (Supplementary file 2).”
9) The importance for modelling groups of having access to online epidemiological data in an easy to export format (e.g., CSV file) could be highlighted. In fact, the Ministries of Health of the three most affected countries provided PDF files only, some of which were images, thus implying remarkable efforts by modelers to readily use such datasets. Moreover, the case of Guinea was even worse as, at the beginning of the epidemic, there was not an official website of the Guinean Ministry of Health with their reports uploaded.
We’re grateful for the chance to address this very important point, and added the recommendation to the Discussion (paragraph seven):
“Ideally, data should be made available online in machine-readable form to facilitate use in analyses. Modelers and other analysts expended enormous effort during the EVD epidemic transcribing data posted online in PDF documents.”
https://doi.org/10.7554/eLife.09186.016Article and author information
Author details
Funding
No external funding was received for this work.
Acknowledgements
We thank the reviewers for excellent comments that improved the manuscript. The views expressed are those of the authors and do not necessarily reflect the views of any part of the US Government.
Reviewing Editor
- Mark Jit, London School of Hygiene & Tropical Medicine, and Public Health England, United Kingdom
Publication history
- Received: June 2, 2015
- Accepted: November 19, 2015
- Accepted Manuscript published: December 8, 2015 (version 1)
- Version of Record published: February 16, 2016 (version 2)
Copyright
© 2015, Chretien et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 6,193
- Page views
-
- 1,005
- Downloads
-
- 97
- Citations
Article citation count generated by polling the highest count across the following sources: Scopus, Crossref, PubMed Central.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Epidemiology and Global Health
Local cervical cancer epidemiological data essential to project the context-specific impact of cervical cancer preventive measures are often missing. We developed a framework, hereafter named Footprinting, to approximate missing data on sexual behaviour, human papillomavirus (HPV) prevalence, or cervical cancer incidence, and applied it to an Indian case study. With our framework, we (1) identified clusters of Indian states with similar cervical cancer incidence patterns, (2) classified states without incidence data to the identified clusters based on similarity in sexual behaviour, (3) approximated missing cervical cancer incidence and HPV prevalence data based on available data within each cluster. Two main patterns of cervical cancer incidence, characterized by high and low incidence, were identified. Based on the patterns in the sexual behaviour data, all Indian states with missing data on cervical cancer incidence were classified to the low-incidence cluster. Finally, missing data on cervical cancer incidence and HPV prevalence were approximated based on the mean of the available data within each cluster. With the Footprinting framework, we approximated missing cervical cancer epidemiological data and made context-specific impact projections for cervical cancer preventive measures, to assist public health decisions on cervical cancer prevention in India and other countries.
-
- Epidemiology and Global Health
- Microbiology and Infectious Disease
Background:
Dog-mediated rabies is endemic across Africa causing thousands of human deaths annually. A One Health approach to rabies is advocated, comprising emergency post-exposure vaccination of bite victims and mass dog vaccination to break the transmission cycle. However, the impacts and cost-effectiveness of these components are difficult to disentangle.
Methods:
We combined contact tracing with whole-genome sequencing to track rabies transmission in the animal reservoir and spillover risk to humans from 2010-2020, investigating how the components of a One Health approach reduced the disease burden and eliminated rabies from Pemba Island, Tanzania. With the resulting high-resolution spatiotemporal and genomic data we inferred transmission chains and estimated case detection. Using a decision tree model we quantified the public health burden and evaluated the impact and cost-effectiveness of interventions over a ten-year time horizon.
Results:
We resolved five transmission chains co-circulating on Pemba from 2010 that were all eliminated by May 2014. During this period, rabid dogs, human rabies exposures and deaths all progressively declined following initiation and improved implementation of annual islandwide dog vaccination. We identified two introductions to Pemba in late 2016 that seeded re-emergence after dog vaccination had lapsed. The ensuing outbreak was eliminated in October 2018 through reinstated islandwide dog vaccination. While post-exposure vaccines were projected to be highly cost-effective ($256 per death averted), only dog vaccination interrupts transmission. A combined One Health approach of routine annual dog vaccination together with free post-exposure vaccines for bite victims, rapidly eliminates rabies, is highly cost-effective ($1657 per death averted) and by maintaining rabies freedom prevents over 30 families from suffering traumatic rabid dog bites annually on Pemba island.
Conclusions:
A One Health approach underpinned by dog vaccination is an efficient, cost-effective, equitable and feasible approach to rabies elimination, but needs scaling up across connected populations to sustain the benefits of elimination, as seen on Pemba, and for similar progress to be achieved elsewhere.
Funding:
Wellcome [207569/Z/17/Z, 095787/Z/11/Z, 103270/Z/13/Z], the UBS Optimus Foundation, the Department of Health and Human Services of the National Institutes of Health [R01AI141712] and the DELTAS Africa Initiative [Afrique One-ASPIRE/DEL-15-008] comprising a donor consortium of the African Academy of Sciences (AAS), Alliance for Accelerating Excellence in Science in Africa (AESA), the New Partnership for Africa's Development Planning and Coordinating (NEPAD) Agency, Wellcome [107753/A/15/Z], Royal Society of Tropical Medicine and Hygiene Small Grant 2017 [GR000892] and the UK government. The rabies elimination demonstration project from 2010-2015 was supported by the Bill & Melinda Gates Foundation [OPP49679]. Whole-genome sequencing was partially supported from APHA by funding from the UK Department for Environment, Food and Rural Affairs (Defra), Scottish government and Welsh government under projects SEV3500 & SE0421.