Microsimulation at Scale for Chronic Disease Modelling:  Executing 100 Million Individual Life-Course Simulations in 100 Seconds

Microsimulation at Scale for Chronic Disease Modelling: Executing 100 Million Individual Life-Course Simulations in 100 Seconds

Odhrán McConnell  ( HealthLumen )  —  “Microsimulation at Scale for Chronic Disease Modelling: Executing 100 Million Individual Life-Course Simulations in 100 Seconds”  (joint work with: Simon Lande)
July 1, 2026, 0:00 am TBC TBC
Conference presentation

Microsimulation is a uniquely powerful technique for chronic disease modelling because it simulates outcomes at the level of the individual over time, capturing heterogeneity, history-dependent progression, multimorbidity, and complex clinical pathways that cohort averages cannot. In an era when chronic diseases account for the majority of global mortality and impose escalating pressure on health systems, decisions about their prevention, treatment, pricing, and resource allocation carry profound long-term clinical and financial consequences. Consequently, accurate long-horizon modelling of these diseases has become central to policy, reimbursement, and investment decisions. Historically, however, microsimulation has been constrained by computational performance. Statistical precision requires large, simulated populations to reduce Monte Carlo error, and probabilistic sensitivity analysis multiplies this burden through repeated parameter sampling. Many models built in spreadsheets or high-level languages require hours or days to run, limiting scenario exploration, delaying iteration, and reducing their practical utility in time-sensitive decision environments. To address these limitations, a legacy microsimulation stack was rebuilt into a high-performance platform capable of executing 100 million life-course simulations in approximately 100 seconds. Performance gains were achieved through several core engineering innovations. The microsimulation core was implemented in modern C++, enabling direct control over memory allocation, cache locality, and execution flow. Compared with interpreted (e.g. Python, R) or spreadsheet-based environments, compiled C++ dramatically reduces runtime overhead and enables predictable, deterministic execution, strengthening validation processes and supporting regulatory-grade transparency and auditability. Memory architecture was optimised to maximise Central Processing Unit (CPU) cache efficiency and minimise allocation costs. Modelled individuals’ attributes, state transitions, and event processes were encoded in compact, structured formats, allowing large virtual populations to be simulated without performance degradation. The engine exploited modern multi-core CPU architectures through multi-threading, allowing independent patient simulations to run concurrently. Because individual life trajectories are largely independent within Monte Carlo microsimulation, the model parallelises naturally, enabling near-linear scaling with available cores. Beyond single-machine performance, the system supports horizontal scaling via containerised simulation instances, allowing elastic expansion across the infrastructure based on workload demand, without reliance on specialised high-performance computing clusters. The platform includes integrated pipelines for data ingestion, preprocessing, simulation execution, and post-processing. Outputs are automatically aggregated into epidemiological, and economic metrics, including incidence, prevalence, costs, and healthcare resource use outcomes, ready for decision analysis. A user-facing interface abstracts technical complexity, allowing domain experts to configure scenarios and execute simulations without interacting directly with the code or infrastructure. The entire platform is securely hosted in the cloud, allowing for easy set up and access anywhere in the world. The system is comprised of cross-cloud components that allow it to be hosted in any of the major cloud providers.
These advances represent a fundamental shift in capability: complex simulations once requiring hours or days may now be completed in seconds, enabling real-time exploration of uncertainty, and rapid scenario iteration to expedite decision-making. Microsimulation can therefore operate at the scale and speed demanded by modern policy, reimbursement, and investment strategies, amid growing chronic disease complexity and multimorbidity.