Summary

epiworldRcalibrate is an R package that provides fast, data-driven calibration of agent-based epidemic models built with epiworldR (Meyer and Vega Yon 2023). Calibration — the process of finding model parameters that reproduce observed epidemic dynamics — is traditionally performed using computationally expensive simulation-based methods such as Approximate Bayesian Computation (ABC), which can require minutes to hours per run. epiworldRcalibrate addresses this bottleneck by implementing DeepIMC (Deep Inverse Mapping Calibration), a pretrained Bidirectional Long Short-Term Memory (BiLSTM) neural network (Najafzadehkhoei et al. 2025) that estimates SIR model parameters from a 60-day incidence time series in seconds. The package manages all required Python dependencies automatically through reticulate (Ushey, Allaire, and Tang 2024), requiring no manual environment configuration from the user. Pre-computed ABC results are also shipped with the package, enabling researchers to benchmark DeepIMC estimates against full Bayesian posterior distributions without re-running expensive simulations.

Introduction

Agent-based models (ABMs) are widely used in infectious disease epidemiology because they capture individual-level behaviors and contact heterogeneity that compartmental models cannot represent (Railsback and Grimm 2019). However, a persistent challenge in applied ABM work is calibration: identifying the parameter values — transmission probability, contact rate, and recovery rate — that make the model’s simulated incidence match observed surveillance data.

The dominant calibration approach, Approximate Bayesian Computation (ABC), draws candidate parameters from a prior, simulates the model, and accepts draws whose output is sufficiently close to the observed data under a chosen discrepancy metric (Toni et al. 2009; Marjoram et al. 2003). This provides posterior distributions over parameters but requires thousands of forward simulations and can be prohibitively expensive when calibration must be repeated across locations, intervention scenarios, or forecasting horizons. An alternative line of work uses machine learning to build surrogate models that approximate the forward mapping from parameters to simulated trajectories (Lamperti, Roventini, and Sani 2018; Angione, Silverman, and Yaneske 2022). While surrogates accelerate individual forward passes, they still require iterative sampling at calibration time.

epiworldRcalibrate takes a different approach, implementing the inverse mapping directly: the DeepIMC network learns to map observed incidence trajectories back to the underlying model parameters. Once trained offline, the network produces calibrated parameter estimates via a single forward pass at inference time — no simulation is required. In a comprehensive simulation study of 5,000 epidemic scenarios, DeepIMC reduced calibration time from approximately 77 seconds (ABC–LFMCMC) to 2.35 seconds while achieving lower parameter recovery error and tighter, well-calibrated prediction intervals (Najafzadehkhoei et al. 2025).

Statement of Need

epiworldRcalibrate addresses the need for fast, accessible calibration of ABMs built with epiworldR. It targets researchers and public health practitioners who need rapid parameter estimates during an active outbreak, researchers conducting simulation studies requiring repeated calibration across many scenarios, and educators teaching infectious disease modeling who want students to focus on model interpretation rather than calibration machinery.

Existing R packages for ABC calibration — such as abc (Csilléry, François, and Blum 2012) — require the user to supply a forward simulator and run it thousands of times per calibration, making them slow by design and dependent on the user’s model implementation. epiworldRcalibrate is, to our knowledge, the first R package to provide a pretrained deep learning model for SIR calibration, enabling near-instant parameter estimation from incidence data with no simulation required at inference time.

Methods

BiLSTM architecture

The DeepIMC calibration model is a BiLSTM neural network (Schuster and Paliwal 1997) implemented in PyTorch and interfaced to R via reticulate (Ushey, Allaire, and Tang 2024). BiLSTMs process sequential input in both the forward and backward temporal directions simultaneously, allowing the network to exploit the full shape of the epidemic curve — including the growth phase, peak timing, and tail decay — when estimating parameters. This bidirectional context is particularly valuable for calibration from complete trajectories, where the tail is as informative about transmission dynamics as the early rise.

The network consists of three stacked BiLSTM layers with 160 hidden units per direction and a dropout rate of 0.5 applied between layers. The final forward and backward hidden states are concatenated with two normalized scalar inputs — population size and recovery rate — to form a 322-dimensional feature vector. This is passed through two fully connected layers (322 → 64 → 3) with ReLU activation in the first layer. Three output heads produce epidemiologically constrained estimates: a Sigmoid activation bounds transmission probability ptran to \((0, 1)\), and Softplus activations enforce positivity for contact rate crate and basic reproduction number \(R_0\). All hyperparameters — including learning rate (\(2.77 \times 10^{-4}\)), number of layers, hidden units, and dropout — were selected via Optuna-based tuning.

The network is trained with a composite loss function combining mean squared error with an epidemiological consistency penalty:

\[\mathcal{L}(\phi) = \mathbb{E}\left[\|\hat{\theta}_\phi(Y) - \theta\|_2^2\right] + \lambda \left(\hat{R}_{0,\phi}(Y)\cdot\gamma - \hat{p}_{\text{tran},\phi}(Y)\cdot\hat{c}_{\text{rate},\phi}(Y)\right)^2\]

where \(\gamma\) is the recovery rate and \(\lambda = 1.77 \times 10^{-4}\) was tuned jointly with the other hyperparameters. This penalty encourages the predicted parameters to satisfy the theoretical identity \(R_0 \times \text{recov} = \text{ptran} \times \text{crate}\), acting as a light regulariser that enforces epidemiological coherence without dominating the training signal.

Training data

The network was trained entirely on synthetic SIR simulations generated with epiworldR. For each simulation, parameters were drawn from the following distributions:

  • Population size \(N \sim \text{Uniform}(5{,}000,\ 10{,}000)\)
  • Recovery rate \(p_{\text{recov}} \sim \text{Uniform}(0.071,\ 0.25)\)
  • Initial prevalence \(\sim \text{Uniform}(100,\ 2{,}000) / N\)
  • Contact rate \(c_{\text{rate}} \sim \text{Uniform}(5,\ 15)\)
  • Basic reproduction number \(R_0 \sim \text{Uniform}(1,\ 5)\)
  • Transmission probability \(p_{\text{tran}} = R_0 \times p_{\text{recov}} / c_{\text{rate}}\), constrained to \([0, 1]\)

Each simulation produced a 60-day daily incidence trajectory, which served as the sole sequence input to the network. Population size and recovery rate were included as additional scalar covariates, consistent with the assumption — standard in applied epidemic modeling — that these quantities are known or reliably estimated from external sources. Input features were normalised with MinMax scaling; the fitted scalers are bundled with the model weights and applied identically at inference time. The model was trained with batch size 64 for up to 100 epochs with early stopping based on validation loss.

Python environment management

The BiLSTM model requires PyTorch, NumPy, scikit-learn, and joblib — Python libraries not natively available in R. A common barrier to adopting Python-backed R packages is the complexity of environment setup: users may need to install a specific Python version, create a virtual environment, and resolve package conflicts with existing installations. epiworldRcalibrate eliminates this barrier through reticulate (Ushey, Allaire, and Tang 2024), using the py_require() interface introduced in reticulate >= 1.41, which is backed by uv — a fast, self-contained Python package manager that resolves and installs all required packages into a dedicated virtual environment without requiring a system Python installation.

From the user’s perspective, setup requires a single function call, run once after installing the package:

epiworldRcalibrate::setup_python_deps(force = TRUE)

This function declares the required packages and a Python version constraint (>=3.11,<3.12) to reticulate::py_require(), initializes the managed environment, and verifies that each package can be successfully imported. In subsequent R sessions, setup_python_deps() with force = FALSE (the default) performs only the verification step. Python is never initialized automatically during normal package use: calibration functions call an internal check that raises an informative error if setup has not been completed, directing the user to setup_python_deps(). The pretrained model weights and fitted scalers are bundled under inst/models/, so no network access is required after initial setup.

ABC benchmark

Pre-computed ABC results obtained via Likelihood-Free MCMC (LFMCMC) (Marjoram et al. 2003) as implemented in epiworldR are shipped as the abc_calibration_params dataset. Each LFMCMC chain ran for 2,000 iterations, discarding the first 1,000 as burn-in, with parameters proposed via multiplicative log-normal perturbations and discrepancy measured by Euclidean distance between simulated and observed trajectories with an adaptive tolerance \(\varepsilon = 0.05 \cdot \|S_{\text{obs}}\|_2\). These pre-computed results allow users to compare DeepIMC point estimates against full posterior distributions without re-running the multi-hour procedure. The generating script is available in data-raw/abc_calibration_results.R.

Key Features

Built on the epiworldR simulation framework (Meyer and Vega Yon 2023) and the reticulate Python interface (Ushey, Allaire, and Tang 2024), epiworldRcalibrate provides the following features:

Comparison with Existing Methods

Feature epiworldRcalibrate (DeepIMC) ABC (abc package)
Calibration time ~2 seconds ~77 seconds (simple SIR)
Uncertainty quantification Point estimate Full posterior
Requires simulation at inference No Yes
Pretrained model included Yes No
Automatic Python management Yes No
epiworldR integration Native Manual

Acknowledgements

This work was supported by cooperative agreement CDC-RFA-FT-23-0069 from the CDC’s Center for Forecasting and Outbreak Analytics. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the Centers for Disease Control and Prevention.

References

Angione, Claudio, Eric Silverman, and Elisabeth Yaneske. 2022. “Using Machine Learning as a Surrogate Model for Agent-Based Simulations.” PLoS ONE 17 (2): e0263150. https://doi.org/10.1371/journal.pone.0263150.
Csilléry, Katalin, Olivier François, and Michael G. B. Blum. 2012. “Abc: An r Package for Approximate Bayesian Computation (ABC).” Methods in Ecology and Evolution 3 (3): 475–79. https://doi.org/10.1111/j.2041-210X.2011.00179.x.
Lamperti, Francesco, Andrea Roventini, and Amir Sani. 2018. “Agent-Based Model Calibration Using Machine Learning Surrogates.” Journal of Economic Dynamics and Control 90: 366–89. https://doi.org/10.1016/j.jedc.2018.03.011.
Marjoram, Paul, John Molitor, Vincent Plagnol, and Simon Tavaré. 2003. “Markov Chain Monte Carlo Without Likelihoods.” Proceedings of the National Academy of Sciences 100 (26): 15324–28. https://doi.org/10.1073/pnas.0306899100.
Meyer, Derek, and George G. Vega Yon. 2023. “epiworldR: Fast Agent-Based Epi Models.” Journal of Open Source Software 8 (90): 5781. https://doi.org/10.21105/joss.05781.
Najafzadehkhoei, Sima, George G. Vega Yon, Bernardo Modenesi, and Derek S. Meyer. 2025. “Generalized Machine Learning for Fast Calibration of Agent-Based Epidemic Models.” arXiv Preprint arXiv:2509.07013. https://doi.org/10.48550/arXiv.2509.07013.
Railsback, Steven F., and Volker Grimm. 2019. Agent-Based and Individual-Based Modeling: A Practical Introduction. 2nd ed. Princeton, NJ: Princeton University Press.
Schuster, Mike, and Kuldip K. Paliwal. 1997. “Bidirectional Recurrent Neural Networks.” IEEE Transactions on Signal Processing 45 (11): 2673–81. https://doi.org/10.1109/78.650093.
Toni, Tina, David Welch, Natalja Strelkowa, Andreas Ipsen, and Michael P. H. Stumpf. 2009. “Approximate Bayesian Computation Scheme for Parameter Inference and Model Selection in Dynamical Systems.” Journal of the Royal Society Interface 6 (31): 187–202. https://doi.org/10.1098/rsif.2008.0172.
Ushey, Kevin, J. J. Allaire, and Yuan Tang. 2024. Reticulate: Interface to Python. https://CRAN.R-project.org/package=reticulate.