Investigating Complex HPV Dynamics Using Emulation and History Matching (2024)

[type=editor,orcid=0000-0003-2825-3651]

\cormark

[1]

1]organization=Department of Mathematical Sciences,addressline=Durham University,city=Durham,postcode=DH1 3LE,country=UK

[orcid = 0000-0002-8479-1860]

[orcid = 0000-0002-7587-9182]

[orcid = 0000-0002-9161-9946][orcid = 0000-0002-0216-9913][orcid = 0000-0003-0437-7916][orcid = 0000-0003-4410-6635][orcid = 0000-0003-1409-8531]

2]organization=Institute for Disease Modelling,addressline=Bill and Melinda Gates Foundation,city=Seattle,country=USA

3]organization=London School of Hygiene and Tropical Medicine,city = London,country = UK

\cortext

[cor1]Principal corresponding author

Andrew Iskauskasandrew.iskauskas@durham.ac.uk[  Jamie A. Cohen  Danny Scarponi  Ian Vernon  Michael Goldstein  Daniel Klein  Richard G. White  Nicky McCreesh[[

Abstract

The study of transmission and progression of human papillomavirus (HPV) is crucial for understanding the incidence of cervical cancers, and has been identified as a priority worldwide. The complexity of the disease necessitates a detailed model of HPV transmission and its progression to cancer; to infer properties of the above we require a careful process that can match to imperfect or incomplete observational data. In this paper, we describe the HPVsim simulator to satisfy the former requirement; to satisfy the latter we couple this stochastic simulator to a process of emulation and history matching using the R package hmer. With these tools, we are able to obtain a comprehensive collection of parameter combinations that could give rise to observed cancer data, and explore the implications of the variability of these parameter sets as it relates to future health interventions.

keywords:

emulation\sephistory matching\sephuman papillomavirus\sepmodelling

1 Introduction

Human papillomavirus (HPV) is the most common sexually-transmitted infection, and is the dominant contributor to the incidence of cervical cancer globally [1]. In recent years, the World Health Organisation (WHO) has identified HPV as a priority and introduced the Global Strategy to Accelerate the Elimination of Cervical Cancer [2]. In particular, attention has been focused on the delivery of effective one-dose vaccines [3] that could make the elimination of HPV more attainable for many countries, not least those with the highest burden of cervical cancer.

To evaluate the benefits and drawbacks of different vaccine roll-out schemes in a particular country or region, without intensifying screening processes, we must have an understanding of the transmission, dynamics, and progression of HPV within the population of interest. A common approach to making such determinations is via a complex computer model – an approach widely used in epidemiology, including for modelling COVID-19 [4], Human Immunodeficiency Virus (HIV) [5] and tuberculosis [6]. To this end, the HPVsim model [7] has been developed. This agent-based model allows for flexible and detailed simulations of HPV transmission through a population network using structured sexual networks, co-transmitting HPV genotypes, B- and T-cell mediated immunity, and high-resolution disease natural history. The model can also be used to apply interventions, such as the rollout of different vaccination strategies, to represent possible future policy suggestions and assess their potential impact and cost-effectiveness.

The detail of HPVsim allows it to accurately reflect the intricate dynamics of HPV, but comes at a cost. Before intervention strategies can be evaluated, we must first determine the combinations of parameters within the simulator that can give rise to the observed reality of HPV and cervical cancer epidemiology in the country or region in question. The high fidelity and depth of the model translates to a multitude of different parameters and choices of network structure and, without a systematic methodology for assessing the ‘suitability’ of a given collection of parameters, any subsequent inference or prediction will be imperfect or incorrect. It is also necessary to accept that neither the model nor the real-world observations are expected to be perfect representations of the underlying reality of HPV progression and the true HPV and cervical cancer burden, respectively; any process of matching parameter sets to observations must account for this.

The structure of an agent-based model gives rise to an additional consideration – even for a fixed choice of parameters the simulator is inherently stochastic, and repeated evaluations will result in a different prediction for the quantities of interest. While we could deem the variation due to stochasticity as subdominant to other effects, if we wish to find a complete collection of viable parameter choices upon which predictions and policy can be built then we should consider the stochastic behaviour as integral to the simulator output.

Various methodologies exist to match complex models to data [8, 9, 10, 11, 12, 13, 14, 15, 16], each with their own advantages and drawbacks. We choose here to apply the history matching framework [17], which allows for an inherent understanding of model and observational uncertainty. Suitably applied, it can be used to find the complete space of parameter combinations that can give rise to observational data, in contrast to approaches like optimisation which seek to find a single ‘best’ match. Using the history matching framework in conjunction with emulators [18] – fast statistical approximations to model output – we may explore the high-dimensional parameter space in a computationally feasible manner.

2 The HPVsim Model

Human Papillomavirus (HPV) is one of the most common sexually transmitted infections worldwide, with an estimated 80%percent8080\%80 % of people infected with it at least once in their lifetime [19]. Most infections are believed to clear of their own accord within 2222 years of onset [20], but some infections persist and cause cells to transform into pre-cancerous lesions. Unchecked, these cells can develop into invasive cancers, particularly vagin*l, anal, penile, and cervical cancers. Of real concern is the link between HPV and cervical cancer, as it is estimated that over 99%percent9999\%99 % of cervical cancer cases present with some form of HPV infection [21], contributing to over half a million new cases annually.

Despite the clear correlation between HPV and cervical cancers, the progression of pre-cancerous cells transforming to stages of cervical intraepithelial neoplasia (CIN) and finally to cancer can not be reliably observed due to long dwelltimes between infection, onset of pre-cancer, and invasion; the influence of HPV on this transition is therefore not necessarily straightforward. It is therefore necessary to use complex computer models (henceforth referred to as simulators) to represent the natural history of the disease, in order to obtain estimates of cancers caused by HPV. Such simulators are more accessible – and, in some cases, more reliable – than observational studies of the disease; when used judiciously they can be a powerful tool to use to inform decision making for future policy.

One such simulator is HPVsim [7], built under the umbrella framework of the Starsim models [22]. It allows for a detailed agent-based representation of the population dynamics of a country without requiring that each individual be modelled explicitly; gives an appropriate and flexible contact structure for HPV transmission; and allows a detailed specification of parameters that influence the natural history of the disease. Different HPV genotypes can be modelled separately – a crucial aspect of modelling the disease, where different genotypes have been observed to contribute at varying severity to pre-cancerous lesions, CIN stages, and cancers [23]. Intervention strategies, including adolescent screening and prophylactic vaccination, can be easily implemented; the process of natural clearance and reactivation of HPV within an individual can also be included. The combination of features within the model makes HPVsim flexible enough to accurately represent the dynamics of the disease and its progression to cancer within national or sub-national settings.

An in-depth description of the structure of HPVsim [7] and a technical description of its usage [24] are beyond the scope of this article; we briefly describe the high level structure here. Individuals are classified by a sexual contact structure, which accounts for individuals partaking in marital, casual, and one-off relationships. Each of these classes of relationship has distinguishing characteristics such as average duration, probability of concurrency, and probability of condom usage, all of which may contribute to the likelihood of contracting HPV. Once HPV has been contracted by an individual, HPVsim models the corresponding natural history dependent on the genotype contracted and agent characteristics, such as prior immunity; in this, the infection has a probability of transforming cells and an individual may progress from pre-cancerous to having neoplastic and finally cancerous cells. An individual may clear the infection without external intervention; in this case, cells already transformed may remain and the individual may contract HPV again in their lifetime but with lower probability, depending on parameter choices regarding initial and waning immunity and cross-genotype immunities. Some examples of pathways that may occur within an individual in the HPVsim model are shown in Figure1, showing relationships and disease progression for three individuals within the model.

Investigating Complex HPV Dynamics Using Emulation and History Matching (1)

To avoid simulating each individual in the population of the country in question, and given the comparatively high prevalence of HPV within a community, a dynamic rescaling of the population is performed at the point where an individual joins the cancer pathway; one individual comes to represent a number of individuals with (initially) the same characteristics, all of whom then progress through the natural history dynamics at their own rate. This multiscale approach elegantly balances the needs of individual based modelling and the computational demands of running large populations, allowing for high granularity of results with efficient simulator run-time.

The flexibility of such a model coupled with the relative paucity of reliable data on HPV prevalence or progression to cancer, particularly in low- to middle-income countries (LMICs), poses a problem of identifiability. If we wish to find a collection of parameters that, when provided to HPVsim to produce simulated outputs, would give rise to the observed reality, the dimensionality of the input parameter space can vastly outstrip that of the available data. It therefore becomes crucial that we have a good understanding of parameters that materially affect the disease dynamics. Furthermore, even if we can identify a subset of the parameters at our disposal to investigate, there remain two questions to answer:

  • How can we best explore the (reduced) parameter space?

  • By considering only a subset of parameters, are we removing parts of parameter space that could in fact represent reality?

The former question is common in the modelling community, particularly since many methods of calibration suffer from what is referred to as ‘the curse of dimensionality’ [25] which might preclude a computationally viable in-depth exploration of the parameter space we wish to consider. The latter question is more subtle, but of paramount importance if we wish to perform post-hoc analysis on the model once matched to data (for example, considering interventions) – if we do not or can not consider the effect of having fixed parameters that could have been varied, we may have ignored parameter combinations that represent the true dynamics of the disease. Any further analysis would therefore be at risk of misrepresenting the projected effects of our intervention strategies and result in non-robust inference: this could be devastating in the context of national health strategies.

One further complication is that HPVsim is inherently stochastic; if we wish to robustly explore the parameter space we must be aware that multiple evaluations, termed repetitions or realisations, will be required at each parameter combination to understand this source of variability. For complex models such as HPVsim, obtaining even a modest number of realisations at each parameter combination can be time- and resource-consuming and compounds the issue of exploring the space fully. We therefore apply a methodology capable of accounting for the two questions posed while handling the disconnect between the finite collection of realisations and the observed reality, in a computationally feasible time.

3 Emulation and History Matching

In this section we briefly outline the mathematical framework of emulation, the process of history matching, and the effect of stochasticity on this approach. More details on the background to emulation can be found in [17, 26, 27, 28]. Details about the framework, including the specific application of emulation and history matching to stochastic models, are provided in Appendix A.

An emulator [18, 29, 30] is a statistical surrogate to a complex simulator output, which given a relatively small ensemble of runs from the simulator in question can provide predictions of the simulator output at any unseen point in the parameter space. In this application, we construct Bayes linear emulators [31], whose prior specification requires only a statement about expectations, variances, and covariances; given these specifications and data D𝐷Ditalic_D from simulator evaluations, an emulator can provide posterior predictions, 𝔼D[f(x)]subscript𝔼𝐷delimited-[]𝑓𝑥\mathbb{E}_{D}[f(x)]blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ] for the output f(x)𝑓𝑥f(x)italic_f ( italic_x ) at any unseen point x𝑥xitalic_x, as well as the uncertainty in the prediction VarD[f(x)]subscriptVar𝐷delimited-[]𝑓𝑥\text{Var}_{D}[f(x)]Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ]. Computing the adjusted expectation and variance corresponds simply to matrix multiplication and a single offline computation of a matrix inverse, making their evaluation orders of magnitude faster than simulator evaluations from any but the most simple models.

Having created emulators for the outputs of interest, our aim is to leverage their computational efficiency to explore the parameter space and identify potentially suitable points for matching to observational data. We must first decide what it means for a point to be ‘suitable’ in this context; to this end, we define an implausibility measure I(x)𝐼𝑥I(x)italic_I ( italic_x ):

I2(x)=(𝔼D[g(x)]z)2VarD[g(x)]+Var[e]+Var[ϵ(x)].superscript𝐼2𝑥superscriptsubscript𝔼𝐷delimited-[]𝑔𝑥𝑧2subscriptVar𝐷delimited-[]𝑔𝑥Vardelimited-[]𝑒Vardelimited-[]italic-ϵ𝑥I^{2}(x)=\frac{(\mathbb{E}_{D}[g(x)]-z)^{2}}{\text{Var}_{D}[g(x)]+\text{Var}[e%]+\text{Var}[\epsilon(x)]}.italic_I start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ( italic_x ) = divide start_ARG ( blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g ( italic_x ) ] - italic_z ) start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT end_ARG start_ARG Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g ( italic_x ) ] + Var [ italic_e ] + Var [ italic_ϵ ( italic_x ) ] end_ARG .(1)

Here, the random quantities e𝑒eitalic_e and ϵ(x)italic-ϵ𝑥\epsilon(x)italic_ϵ ( italic_x ) represent the observational uncertainty and the model discrepancy respectively; to wit, the extent to which we are uncertain about the accuracy of the real-world observation, z𝑧zitalic_z, made and the (possibly x𝑥xitalic_x-dependent) extent to which we believe our simulator may not be representative of real life [30]. A large value of I(x)𝐼𝑥I(x)italic_I ( italic_x ) suggests that, despite any uncertainties inherent in the model or in the emulation at the parameter combination x𝑥xitalic_x, the prediction is extremely unlikely to give rise to an acceptable match to observational data. Conversely, a small implausibility at x𝑥xitalic_x can be due to two main reasons: either the prediction is close to the observation, suggesting an acceptable match to data; or the uncertainties (particularly the emulator uncertainty VarD[g(x)]subscriptVar𝐷delimited-[]𝑔𝑥\text{Var}_{D}[g(x)]Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g ( italic_x ) ]) are large at the point, suggesting a region of parameter space worthy of further investigation. Such points are termed non-implausible.

The process of history matching leverages the concept of implausibility, and the fact that emulators require relatively few simulator evaluations to train, as follows. At each ‘wave’ k𝑘kitalic_k of history matching, we use a collection of simulator evaluations drawn from our current non-implausible region (denoted 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT) to train emulators to outputs of interest. These emulators are then used to calculate implausibility across the whole of 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT, resulting in a smaller non-implausible space 𝒳k+1subscript𝒳𝑘1\mathcal{X}_{k+1}caligraphic_X start_POSTSUBSCRIPT italic_k + 1 end_POSTSUBSCRIPT. This new non-implausible space is sampled from, forming the new ensemble of points on which to perform simulator evaluations. This process continues until we reach a pre-agreed stopping condition; at this point, our final non-implausible region represents all parameter combinations that could give rise to the observational data, given our uncertainties about the observational data and simulator itself. The formal statement of the history matching process, as well as extensions to considering multiple outputs and the effect of stochastic simulators upon the implausibility measure, can be found in Supplementary Material B.

The process of determining prior specifications for emulators of complex simulators is seldom intuitive, and performing robust sampling from geometrically non-trivial non-implausible regions is often difficult. In this work we have taken full advantage of the hmer package [32], which automates the process of emulator training and adjustment with respect to data via emulator_from_data(), validates trained emulators for suitability via validation_diagnostics(), and carefully proposes space-filling designs from the non implausible region via generate_new_design(). It also allows for intuitive visualisation of multiple different features of the trained emulator predictions, the proposed non-implausible region, and evolving features of the input and output spaces across multiple waves. More details of the functionality of the hmer package may be found in [33], as well as in the examples and vignettes included in the package itself; where visualisations have been generated using the package in this work, we quote the function directly.

4 Results

4.1 Problem Specification

The HPVsim model can provide a wealth of information about any feature of the disease, its spread through a population, and the progression to cancerous lesions within an individual. However, observational data available is not of the same breadth or depth, so we must determine what data is suitable for matching our simulator evaluations to. We consider as observations of interest the numbers of new cancer cases recorded in 2020202020202020 in the country of interest, aggregated by age, as well as the measured distribution of HPV genotypes in the population for patients presenting with high-grade lesions (termed the “CIN3” state) and cancers [34, 35]. These data were deemed to be reliable and sufficient for providing a meaningful history match, particularly as the eventual goal of the analysis of HPVsim simulations includes an analysis of genotype acquisition within the population and future cancer cases. A common clinical differentiation between different HPV genotypes motivated a further focus and partial amalgamation of genotypes of interest: expert elicitation suggested that a dozen genotypes were of interest, accounting for over 95%percent9595\%95 % of HPV cases [34]. Two genotypes, HPV16 and HPV18, were considered distinct while five further genotypes (HPV 31, 33, 45, 52 and 58) were collected into a ‘high impact’ class, HPVhi5; the remaining five genotypes of interest were collected into an ‘other high risk’ class, HPVohr. With this specification of genotype classes, 22222222 outputs of interest were identified for which reliable observational data was available, along with an appropriate understanding of their observational uncertainty. Details of this observational data, as well as any uncertainties therein, may be found in Supplementary Material C.2. Initial pilot simulations suggested a tension between reported early- and late-age cancers, compared to the possible output of HPVsim (and by extension, the disease dynamics); in the absence of strong beliefs regarding systematic bias in the data-collection, we deemed it appropriate to increase the observational uncertainty for these outputs.

We must also determine the most important parameters to include in our analysis, from the large array of parameters available for adjustment in HPVsim. After expert consideration to highlight the most influential parameters to the simulator, and sensitivity analyses to determine those with the greatest impact on model behaviour, the candidate set of parameters was reduced to a collection of 33333333 whose inclusion would be critical to a robust investigation of the model behaviour. For those parameters which remained unvaried, fixed values were determined via expert elicitation or inferred from demographic surveys [36] and appropriate model discrepancy included to account for their exclusion. The initial non-implausible space, 𝒳0subscript𝒳0\mathcal{X}_{0}caligraphic_X start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT, therefore consisted of a 33333333dimensional hypercube whose extent corresponded to the largest physically reasonable range for each of the parameters considered. The initial parameter space is detailed in Supplementary Material C.1.

4.2 Emulation and History Matching

4.2.1 Emulator Training

At each wave, an ensemble of 495495495495 parameter sets in a space-filling design were provided to HPVsim for evaluation; of these, 330330330330 were used to train the emulators in accordance with arguments presented in [37], with the remaining 165165165165 used for validating the trained emulators. Due to the philosophy behind variance emulation and history matching, we were able to modify our emulation strategy throughout the waves without concerns about jeopardising our final inference. At early waves, anticipating volatile (and therefore computationally expensive) simulator behaviour, we demanded only 16161616 realisations at each parameter combination and focused only on emulating the mean and variance response, accepting that this would result in a conservative prediction of uncertainty. As the non-implausible space was reduced and the simulator behaviour became more stable, we increased the number of realisations at a point to 32323232, and at later waves still to 50505050.

To validate the specifications for our emulators, particularly the global response to the inputs, emulator diagnostics described in [33] were performed using the hold-out validation set, alongside a handful of ‘acceptable’ simulator runs obtained via pilot studies and hand-fitting; where problems were observed in a particular emulated output, we were at liberty to interrogate the structure of the offending emulator and adjust the prior specification accordingly. An example of diagnostic plots produced for each emulator is shown in Figure2. Corrective measures applied included inflation of the prior uncertainty; modification of the global emulator response; and inclusion of additional simulator runs in problematic regions of space. Quantities whose emulator could not adequately represent the output at a given wave, even after modification, were removed from the current wave and reincorporated at the successive wave.

Investigating Complex HPV Dynamics Using Emulation and History Matching (2)

4.2.2 History Matching

Having created a collection of emulators for the relevant outputs at a wave, we applied the process of history matching. At early waves, we used a somewhat conservative second-maximum of univariate implausibilities, letting at most one output be far from the corresponding target, with a cutoff of I2M(x)3subscript𝐼2𝑀𝑥3I_{2M}(x)\leq 3italic_I start_POSTSUBSCRIPT 2 italic_M end_POSTSUBSCRIPT ( italic_x ) ≤ 3; at later waves we required maximum implausibility to satisfy this cutoff. At the final wave, we incorporated relationships between emulated outputs using a multivariate measure. In searching for acceptable matches to observation, we leveraged the vastly improved computational efficiency of the emulators, evaluating no fewer than 10,0001000010{,}00010 , 000 candidate points during the point proposal stage at each wave. We followed the procedure detailed in [33] to search for a space-filling non-implausible sample of points, summarised in Supplementary Material B.

Where we were unable to generate sufficiently many candidate points from this process, and hence we might be concerned about over-sampling of a small region of the true non-implausible region, the implausibility cutoff was allowed to be relaxed so as to allow for a representative sample. Candidate points generated in this ‘relaxed’ framework were then sub-selected to provide a representative sample at a lower cutoff and used as a proposal for generating a larger collection of points. This process could be repeated many times before we obtained a space-filling design of non-implausible points at the desired cutoff, ensuring a more accurate identification of the complete non-implausible region111This ‘annealing’ process is similar in spirit to that proposed in [38], though less computationally expensive and somewhat less robust in identifying disconnected regions. However, for this application, the method described above proved adequate for identifying the non-implausible space..

Our decision on how many waves were required for this problem was motivated by the aims of the analysis. There are two connected but different goals of the analysis at hand: the first is of interpretation of the non-implausible region, and in particular the most influential or restricted parameters; the second is to make meaningful statements about the distribution of different genotype classes within the population and the trend of future cancer cases, for which no observational data exists. To our first aim, we wished to propose sufficiently many points that, when evaluated by the simulator, gave rise to matches to observational data – this was not guaranteed to occur and a secondary condition was that the emulator uncertainties had a subdominant effect in the denominator of the implausibility measure (suggesting that further waves of emulation would not materially affect the proposal of points). To address the second aim, we also evaluated a sample of the points proposed at each wave with respect to those unobservable quantities of interest. Where these quantities showed no substantial difference between waves or across the current non-implausible space, indicating that our further inference about these quantities would not be affected by a change in structure of the non-implausible region, we could consider this aim of the history match fulfilled.

4.3 Analysis of Results

Before focusing on the eventual structure of the final non-implausible region, we first discuss some interpretable aspects of the simulator extracted from the emulation process. We can interrogate the active variables for each output and well as the strength of effect that each input has on the outputs in Figure3, as generated by hmer::effect_strength().

Investigating Complex HPV Dynamics Using Emulation and History Matching (3)

We may note some features of the outputs. Firstly, the choice of active variables coincides in many cases with physical intuition we might have about the disease: for example, transmission rate beta and episomal duration (the average duration of infection prior to clearance, control, or transformation) de_16 have a greater influence on the resulting early-age cancers than in later ages, while parameters governing latent control and reactivation of HPV (control and reactive) tend to more severely impact the resulting cancers in older age groups. The relative strengths-of-effect also agree with intuition: for example, a higher beta results in more cancer cases, while a higher control reduces the total number of cancers in an age group. Some aspects are more surprising: the relative paucity of impact that genotype-specific relative transmission rates rb have on cancers suggests that deviations from the overall transmissibility do not materially affect the resulting profile of cancers, perhaps suggesting that a combination of severity rate (the rate of transformation of cells in infected individuals) sr, relative transmissibility rb, and transformation probability tp may be more directly informative.

In all, considering the stopping conditions described above, 16161616 waves of emulation and history matching were required. At this stage, the non-implausible region was unlikely to be restricted any further by additional waves, over 25%percent2525\%25 % of points proposed by the emulators resulted in complete matches to data (even in the absence of any model discrepancy), and all projected quantities of interest were stable between waves. To interpret the results most effectively, we now turn to visualisation of the final non-implausible space. We first consider the progression of the outputs towards the observational targets as we progress through the waves: Figures4 and 5 shows the output runs, coloured by wave, as generated by hmer::simulator_plot(). For ease of visual analysis, we have split the outputs into two plots: one for cancer totals and one for proportions. Due to high output variability, the total number of cancers per age group is presented on a log scale.

Investigating Complex HPV Dynamics Using Emulation and History Matching (4)
Investigating Complex HPV Dynamics Using Emulation and History Matching (5)

With increasing wave number, the ensemble of simulator runs arising from non-implausible points at that wave is tightening toward the targets, shown as error bars. This reduction of the output space is particularly noticeable in Figure4, where the extrema of model behaviour in the first and final wave differs by a factor of up to 400400400400. The conflict anticipated in pilot studies is borne out by the early- and late-stage cancer totals, which are tied to the lower ends of their respective targets. Figure5 demonstrates that there is more flexibility in the proportions of various HPV genotypes contributing to cancers and high-grade lesions, but the behaviour of genotype 16161616 seems to be the most restricted beyond that imposed by the observational data. Note that in these plots, each individual line corresponds to a single parameter set, where we have aggregated using the mean of the realisations. There is therefore additional stochastic variability not displayed here, but this representation serves well in showing the progression of the history match and the broad trends of the simulator response.

To try to gain further insight into the output structure of the model, we might consider looking at pairs of observations rather than each individually. This information is presented in Figure6 for an informative subset of the outputs and provided using hmer::wave_values() – in each panel of the plot not on the main diagonal, the results of two simulator outputs for a given parameter set are plotted as a point, with a target window overlaid corresponding to the observational data. In those plots above the diagonal, the plot is ‘zoomed in’ to the region of interest, so that the panel is centred on a square corresponding to the rescaled target window; below the main diagonal, the results are presented raw. We immediately see the extreme reduction in simulator behaviour as the history matching waves progress, to the extent that the observational targets for cancer cases are almost too small to see in the raw plots. We also see strong correlations between cancers in neighbouring age groups, attenuating as we consider correlations between age groups with larger separation. The conflict between early-age cancers and those at later age groups becomes more stark, as we see that simulator runs are only just able to overlap with the target windows in a small corner. This feature is especially stark when considering simultaneously matching to cancers at ages 25-3025-3025\text{-}3025 - 30 and 35-4035-4035\text{-}4035 - 40. In the targets where we consider proportions of genotypes, we see clear structure; particularly a hard diagonal bound on the value of pairs of outputs which deal with the same type of proportion (either cancers or CIN3), since these proportions must sum to no more than 1111; between different types of proportions, we see a broadly correlated relationship for a given genotype (for example, cancer_16 and cin3_16), with little dependence between genotypes. This sheds light on the nature of the cross-genotype immunity structure we could have specified: were we to have had strong views on whether contracting and clearing HPV16 affords protection against contracting HPV18 (say) at some later point in life, we could have stated this within the fixed parameters of HPVsim; the results we have obtained here do not give any strong suggestion of the necessity of such a statement.

Investigating Complex HPV Dynamics Using Emulation and History Matching (6)

We can examine two-dimensional projections of the input space, rather than the output space, using hmer::wave_points(). In the interests of ease of inspection, a truncated collection of parameters is shown in Figure7 to highlight which input parameters have been most restricted across the waves of history matching, and to where.

Investigating Complex HPV Dynamics Using Emulation and History Matching (7)

We can see that, even at early waves where the simulator behaviour was subject to a large amount of volatility, the overall severity rate of the disease (governed by sev_dist) was necessarily low. This is a consequence of the wide ranges placed on the parameters initially: while the highest values of severity were not unphysical, they were very unlikely to give rise to the observations we intended to match to and this was quickly borne out. We also see that the probability of reactivation of the disease and of latently controlling the disease (reactive and control, respectively) have found themselves heavily restricted by our final wave: in the absence of preventative measures and in light of observational data, we must conclude that the chance of reactivation of latently controlled HPV is low, and that a large number of infected individuals clear the disease of their own accord. This seems sensible in the context of HPV, where we know that an extremely high proportion of the population contract it, but comparatively few progress to cancer. It is possible, however, that introducing measures such as vaccination or screening into the HPVsim natural history would have modified these statements – for this LMIC country in question, we would not consider this result to be incompatible with the true natural history of the disease due to limited testing and control strategies implemented. The genotype-specific parameters also show some interesting features: we see in particular that the individual genotypical severity rates show a large amount of dissimilarity with HPV16 having the largest severity of all genotypes under consideration; the other individual genotype, HPV18, seems to have one of the lowest severity profiles but its probability of transformation (tp_18) is much higher than the aggregated genotypes HPVhi5 and HPVohr, partially justifying its singular inclusion.

We may find rough bounds on the volume of space that remains after performing the waves of history matching: on the basis of the proportion of non-implausible points in 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT that remain non-implausible according to the wave k𝑘kitalic_k emulators, we estimate a final volume equal to 1×1016similar-toabsent1superscript1016\sim 1\times 10^{-16}∼ 1 × 10 start_POSTSUPERSCRIPT - 16 end_POSTSUPERSCRIPT; a rough upper bound on the space reduction can be given by considering the minimum enclosing hyperrectangle for wave 16161616 and compare it to our initial ranges, finding a volume ratio of 1.2×10141.2superscript10141.2\times 10^{-14}1.2 × 10 start_POSTSUPERSCRIPT - 14 end_POSTSUPERSCRIPT.

Investigating Complex HPV Dynamics Using Emulation and History Matching (8)

Figure8 shows the reduction in space over the course of the waves of history matching. We may note the ‘spike’ in space reduction in the final wave after a period of decreasing gains in previous waves: this is due to the inclusion of a multivariate implausibility measure in wave 16161616, forcing relationships between outputs. In fact, the inclusion of covariance information between outputs was a large driver in the satisfaction of our stopping conditions; to verify this, a further ensemble of emulators was generated from the proposal at the end of wave 16161616 but was unable to reduce the space to any significant extent. We also considered the emulator uncertainty at the final wave compared to the other sources of uncertainty, namely observational error and model discrepancy: the prior emulator variances are 4%percent44\%4 % the size of the other uncertainties on average, with range [0.2%,12.4%]percent0.2percent12.4[0.2\%,12.4\%][ 0.2 % , 12.4 % ]. At this stage, the emulators are contributing very little to the overall uncertainty in the implausibility measure (1), and so further refinement of the emulators was extremely unlikely to lead to more accurate results. We return to the final stopping condition shortly.

The information presented in Figure8 is worthy of further emphasis: to obtain a single point from the final non-implausible space were we to search at random from the initial volume, we would require around a quadrillion model evaluations (multiplied by the requisite number of realisations at each parameter set). The entire process of history matching and emulation required the evaluation in HPVsim of 8,00080008{,}0008 , 000 points in parameter space, including those that contributed to validation sets, resulting in 180,000180000180{,}000180 , 000 model evaluations in total including repetitions. Of all points proposed, 802802802802 result in output consistent with all observed data - an appreciable yield of 10%percent1010\%10 %. In the final wave of history matching, over half of the points proposed were consistent with observation and model discrepancy, and we may therefore use the emulators at the final wave to generate arbitrarily many non-implausible parameter sets at a speed far greater than would be possible using the simulator alone. The reason we have required so few simulator evaluations is due to the burden of computation and prediction taken on by the emulators. Only because emulator prediction of the mean response is around 107superscript10710^{7}10 start_POSTSUPERSCRIPT 7 end_POSTSUPERSCRIPT times faster than the equivalent HPVsim simulator evaluation, accounting for the need to perform realisations, were we able to effect a robust examination of the input parameter space.

4.4 Comparison to Other Methods

Of course, other methodologies exist to match complex models to data, and in fact HPVsim incorporates an optimisation algorithm [8]; we conclude this section with a comparison between our approach and that of the ‘native’ HPVsim approach.

Even with this comparatively inexpensive method, an optimisation run required between 1,00010001{,}0001 , 000 and 10,0001000010{,}00010 , 000 simulator evaluations to find a local maxima of the goodness-of-fit measure, even before stochasticity is taken into account. Were we to assume that multiple seeds of the optimisation algorithm would result in an equivalent exploration of the non-implausible region, and supposing that we take the last 10101010 points proposed by optimisation as a proxy for the space of interest, we would need to evaluate the simulator at a minimum of 80,0008000080{,}00080 , 000 parameter locations; in order to use no more simulator evaluations than the history matching process required, we could therefore perform no more than 2222 realisations at each parameter combination, in contrast to the minimum of 16161616 that we were able to evaluate with the history matching approach. Furthermore, the requirement of targeting a goodness-of-fit measure can provide proposals with good measure statistics that, under inspection, have unwanted behaviour when considered in the context of matching the model to data. We see this demonstrated in Figures9 and 10, where the two methods are compared directly.

Investigating Complex HPV Dynamics Using Emulation and History Matching (9)
Investigating Complex HPV Dynamics Using Emulation and History Matching (10)

We can see that, when considering the age-aggregated cancers in Figure9, those points proposed using history matching and those proposed via optimisation have somewhat different characteristics. Particularly for early-age groups, the behaviour of each group diverges: while the history matched proposals have been constrained to the lower ends of the target bounds in early- and late-age cancers so as to not overestimate mid-age cancer cases, the optimisation procedure (and the goodness-of-fit measure used) has prioritised targeting the centres of the late-age cancer cases, at the expense of those targets for cancers in anyone aged below 35353535. In some cases, this disconnect is quite stark, even on a logarithmic scale (in particular, those for age groups 20-2520-2520\text{-}2520 - 25 and 25-3025-3025\text{-}3025 - 30). In fact, those points proposed by history matching surpassed the points generated via optimisation even with reference to the goodness-of-fit measure used by the optimiser, highlighting the risks inherent in using a method which can gravitate to local minima, even given different starting seeds.

In the targeted proportions shown in Figure10, the difference is less striking when considering whether proposals lie inside the target window, but there is one aspect worthy of note. While the history matching proposals are spread widely around their respective target regions, those obtained via optimisation are comparatively tightly clustered and often to one extreme of the region that we know a posteriori would result in acceptable matches. Were we to rely solely on the information provided to us by optimisation, we would be at risk of making overly restrictive statements about the relative contributions of each genotype to cancers, which could result in non-robust statements about the effectiveness of vaccination strategies were we to consider a single genotype in isolation.

The above considerations, and particularly those arising from Figure10, highlight the key conceptual difference between history matching and optimisation. Whereas optimisation seeks to find a ‘good’ parameter combination for fitting to observational data, history matching seeks to find all possible combinations that match observations. We would therefore not expect the optimisation algorithm to span the full space of possible matches (even with multiple restarts with different seeds), as it is task to which it is not designed. The primary strength of optimisation is the speed at which one might find ‘good’ parameter combinations, but we see here that even this cannot be depended on: to stand a chance of achieving an collection of acceptable parameter combinations of comparable size, accounting for stochastic repetition, we would have required at least an order of magnitude more simulator evaluations. Furthermore, there is no guarantee that those points generated would be representative of the full space of good matches, liable as the algorithm is to settle in local minima of some loss function.

5 Inference and Prediction for HPV

Often, a key aim of disease modelling is to make informed statements about the future trend of the disease (as well as the possible spread of outcomes) so as to make robust recommendations about the potential efficacy of intervention campaigns; we may also wish to investigate properties of the natural history of the disease that would be impossible, or at least extremely hard, to do via real-world studies. We detail here how the results of the history matching process helps with both of these aims.

We first consider the acquisition and progression of HPV and its transformation to cancer. Due to the agent-based nature of HPVsim, we may interrogate the progression of each individual from the moment they contract HPV to their clearance of the disease or transformation to cancer, thus building up a demographic study of the ages at which HPV is contracted and the time taken to progress through each of the cervical intraepithelial neoplastic stages. This information could be of extreme value, potentially in conjunction with targeted longitudinal studies, to evaluate the optimal ages at which screening or vaccination strategies should be focused. The results are shown in Figures11 and 12.

Investigating Complex HPV Dynamics Using Emulation and History Matching (11)
Investigating Complex HPV Dynamics Using Emulation and History Matching (12)

We can see that the impact of each of the genotype classes on the progression of HPV to cancer are similar in shape, but there are some key differences between them. The spreads of possible stage acquisitions for the individual genotypes, HPV16 and HPV18, are much tighter than those for the 5555-genotype classes HPVhi5 and HPVohr. This is unlikely to be due to the fact that the 5555-genotype classes consist of multiple different strains of HPV, since HPVsim models those classes in the same fashion as it does the individual genotypes; moreover the flexibility in our targets used (and described in Supplementary Material C) for these genotype classes is no less restrictive than for the individual genotypes. We may conclude that, in order to obtain the observed data we have available to us, the individual genotype classes have a much more restricted possible progression path within the population.

We may also consider general behaviours between these genotypes, particularly in considering the most likely length of time spent in each state. While each HPV genotype reaches peak prevalence in the population at around the same time, at the ages of 15-1715-1715\text{-}1715 - 17, there is a marked difference in the equivalent age for cancer. Table1 shows the most common times spent in each state.

GenotypePrecinCIN1CIN2CIN3Total
HPV161111444455551111111121212121
HPV184444555588881111111128282828
HPVhi52222444466661717171729292929
HPVohr2222555566661515151528282828

While there is little material difference between genotypes when considering time spent in a particular state, the cumulative effect of the more severe genotypes is significant. In particular, the modal age at which cancer is present for an individual infected with HPV16 is 36363636, in comparison with 43-4643-4643\text{-}4643 - 46 for all other genotypes. If we consider the ‘precin’ and low-grade lesion (CIN1) states as being manageable (via pharmaceutical treatment or self-clearance), then this suggests that it is paramount that any screening procedures focus on the time spent in these stages: namely, ages 15-2415-2415\text{-}2415 - 24. Once an individual has reached CIN2, it is unlikely that an individual will self-clear and prevent progression to later stages and, eventually, cancer. In these circ*mstances more involved, expensive, and painful treatments are required, making early detection paramount [39].

As mentioned, it is possible for individuals who contract HPV to clear the disease of their own accord. In this case, we may be interested in the time taken to clear the disease: in particular, whether the clearance time is lower than the time taken to reach the CIN2 stage. Figure13 shows the proportion of individuals that clear HPV in a certain number of years, collected in quarter-year intervals, as well as the cumulative proportion of infected individuals that clear the disease over time.

Investigating Complex HPV Dynamics Using Emulation and History Matching (13)

We may note that the peak of disease clearance is around 6666 months, common to all genotypes under consideration; after 1111 year, approximately 50%percent5050\%50 % of infected individuals have cleared their infection. This is in line with observational studies on self-clearance of HPV, which suggests a median clearance time of 13.513.513.513.5 months and 52%percent5252\%52 % of cases cleared within this period [20]222This data is an appropriate comparator as it ignores any external contributions from HIV positivity or treatment pathways, in alignment with our modelling scenario.. One interesting distinguishing feature is in the cumulative clearance rate from each genotype: we see that HPV16 is more likely to persist beyond 6666 years than the other genotypes. Coupled with the comparatively short gestation time seen in Table1 we might consider this genotype to be the critical strain of interest, and a potential focus for any further study.

As a final example, we briefly consider the range of possibilities in the future. We do not make definitive statements save to highlight the breadth of possible outcomes given the observational data we have had access to. Figure14 is derived from running HPVsim at each of the final proposal points across time, and considering total number of cancer cases at each year from 2010201020102010 to 2030203020302030.

Investigating Complex HPV Dynamics Using Emulation and History Matching (14)

We note that all points proposed, based as they were on age-aggregated targets, nevertheless also fall within the overall total number of cancers for 2020202020202020, as demonstrated by the vertical error-bar. This needn’t have been the case: for example, had some of the parameter combinations resulted in agreement with all observational data but always as an over-estimate of the means, then we may have seen trajectories above this bar (a point we return to shortly). We may see, nevertheless, that there remains a large spread on this higher-level statistic, and that spread propagates to any predictions we might make in the future. Averaging across predictions at each of 2020202020202020 and 2030203020302030, there is little difference (12,3001230012{,}30012 , 300 versus 12,2751227512{,}27512 , 275), and the proportion of predictions that suggest a decrease in cancer cases over the ten year period is 56%similar-toabsentpercent56\sim 56\%∼ 56 %, giving no strong indication about future behaviour. The projected change is in the range [5.7%,4.5%]percent5.7percent4.5[-5.7\%,4.5\%][ - 5.7 % , 4.5 % ], corresponding to between 600600600600 fewer and 580580580580 more cancer cases in 2030203020302030 compared to 2020202020202020. This suggests a final range of total cancers in 2030203020302030 of [9,900,14,000]990014000[9{,}900,14{,}000][ 9 , 900 , 14 , 000 ]. The two extremal predictions are almost directly opposite, and if taken in isolation would result in entirely different conclusions, and yet they are of equal validity given the information available to us333For reference, the equivalent analysis on the optimisation runs mentioned in the preceding section suggests a projected change in the range [4%,1.7%]percent4percent1.7[-4\%,1.7\%][ - 4 % , 1.7 % ] corresponding to between 450450450450 fewer and 250250250250 more cancer cases in 2030203020302030, with 30%percent3030\%30 % of proposed points suggesting an overall increase in numbers of cancer cases during this time.. Were we to consider using HPVsim to make predictions of efficacy of interventions, the range of possible baseline outcomes must be taken into account in order to robustly predict the effect of any collection of vaccination and screening strategies, and certainly when considering cost-effectiveness of those strategies.

Given that the range that the predictions can span is almost certainly of critical importance for a robust analysis, we might consider further whether the points obtained through history matching represent the full range of behaviours. There may be a section of a boundary in the high-dimensional space which is not included in this ensemble of final runs which would give rise to a more extreme prediction – since we are exploring a 33333333dimensional space we would not expect the ensemble of 500500500500 points here to explore all corners and boundaries of the non-implausible space and know that extreme behaviour is likely to fall on the boundary. We can, however, use these points as a training set for emulating the progression of total cancer cases over time and thence seek to find the projected maximum and minimum via the emulator, rather than the simulator.

Figure15 gives us an indication of the behaviour of the total number of cancer cases across the entirety of the non-implausible space. In order to aid the investigation of this, we have used our final wave of emulators to propose 5,00050005{,}0005 , 000 points from the non-implausible region. We then created an emulator for the simulator prediction of cancer cases in 2030203020302030 using the 500500500500 runs above, and used this trained emulator to predict over the larger ensemble of parameters. In contrast with the equivalent task using HPVsim, the process of proposal, training, and prediction required less than an hour’s computational time.

Investigating Complex HPV Dynamics Using Emulation and History Matching (15)

This gives us some degree of intuition into the drivers of high numbers of cancer cases in 2030203020302030: in particular, high reactivation rate reactive has an obvious effect on increasing the numbers of cancers in the future as we would expect. There are also interesting features here which may bear consideration; in particular, the negative correlation between cancer cases in 2030203020302030 and both duration of cancer (dc) and individual-level severity (sev_dist). Such features may be related to behavioural changes in sexual contact when individuals suffer from cancer for a longer period of time, or where they are infected with more severe cases of HPV and thus progress to diagnosable cases more quickly. Fundamentally, however, we find that over this much larger space of acceptable parameter combinations the predictions of change in cancer cases from 2020202020202020 to 2030203020302030 is widened further, but not excessively so: according to our trained emulator, projected changes range from 7.5%percent7.5-7.5\%- 7.5 % to +5.8%percent5.8+5.8\%+ 5.8 %, corresponding to between 800800800800 fewer and 750750750750 more cases; the overall range of possible total cancers in 2030203020302030 is [9,600,14,100]960014100[9{,}600,14{,}100][ 9 , 600 , 14 , 100 ]. Fundamentally, this potential difference of 4,50045004{,}5004 , 500 cancer cases will have a huge effect on any decision making we would wish to perform and model using a simulator such as HPVsim. One may see that decisions made about clinical interventions or the focus of new screening programs would drastically change were we to consider only one of the extremes of predicted behaviour; the framework of emulation and history matching allows clinicians and economists to be aware of the complete space of possible effects any intervention may have, and make appropriately informed decisions as a result.

6 Discussion

Models such as HPVsim are invaluable in investigating complex infections at a national level, allowing for an exploration of the natural history of a disease that is not possible from observational studies. Any shortfalls in collected data that we do possess can, in many cases, be mitigated by inference from such a model; parameter combinations that provide matches to this observational data can be used in a variety of ways to analyze physical characteristics of the disease, national demographics, or consider the potential effects of interventions. However, the complexity inherent in using such a model for a thoughtful analysis of a disease can preclude a meaningful and thorough exploration of the parameter space, particularly when the parameter space in question is of high dimension.

We have argued here that emulation and history matching addresses the problems in matching a complex model to observational data efficiently. The application of this framework to HPVsim, facilitated by the R package hmer, allowed us to start with very uninformative parameter ranges (restricted only by physical impossibility, rather than expert judgement on their true values) and substantially reduce it to find a large collection of matches to observed data with comparatively few simulator evaluations: we required fewer than 200,000200000200{,}000200 , 000 simulator evaluations in total to identify an acceptable subspace approximately 17171717 orders of magnitude smaller than our original space. In the process of performing our waves of emulation, we obtained over 800800800800 parameter combinations whose simulator evaluations provided matches to our data, given the underlying uncertainties in observation, model, and stochasticity; having performed the 16161616 waves of emulation we can generate many more such parameter combinations in a short space of time. In contrast, performing optimisation in the same setting required many more simulator evaluations to obtain a handful of final proposal points; these final parameter combinations were still outperformed by those proposed by history matching in almost all cases, both when visually analyzing these proposals and by the optimisation scheme’s own goodness-of-fit measure. More importantly, the parameter sets proposed by history matching implicitly take into account the imperfections inherent in any simulation task, which is non-trivial (or in some cases, impossible) to include in other methodologies such as optimisation.

In the context of HPVsim, the large collection of parameter sets proposed from history matching comprise a representative collection of the complete parameter space that could give rise to our physical reality; this space can be used to make further inference about unobservable characteristics of the disease, make meaningful comparisons between different genotypes therein, and form a basis for robust prediction of future trends and response to intervention. We have identified a collection of factors that would help with assessing the implications of future behavioural changes while ensuring the validity of any inference drawn. In this work we have focused on the natural history of HPV in the community, particularly the acquisition, clearance, and progression of HPV to cancerous cells, but one could apply this process to any possible output of the simulator. We have also considered the output space of possible cancer cases in the future, in the absence of intervention, in order to highlight the range of different baseline scenarios that should be considered in an intervention scenario. Of most interest is the fact that our output space is split almost half-and-half as to whether we expect annual cancer cases to decrease or increase in 2030203020302030 compared to our observations in 2020202020202020; this is of key importance were we to consider applying interventions and analysing their effect.

The eventual uses of the calibrated model can be instructive in ways other than that described here. While we considered the dwelltime of genotypes as being unobservable in Section5, one could see such quantities as in-principle observable given new studies of the disease. One might use the information gained here to identify observational quantities that would be most informative in further reducing the non-implausible region to aid further study design. This concept of using emulation to guide experimental design has been touched upon in other settings [40], but would be of paramount importance here. Large-scale population studies of disease are time-consuming and expensive and, frequently, those quantities that are most accurately observed are not necessarily those that we would select for divining properties of the disease. By using this framework to inform data-gathering campaigns, we may be able to provide concrete statements about the most important information required to characterise the disease in a given country. We hope to explore this avenue of research in future works.

Such avenues of exploration and investigation notwithstanding, the work presented here provides a blueprint for robust examination of complex, stochastic, computer models and a means by which we can interrogate the final space of acceptable matches. With the increase in availability of computational resources, agent-based models such as HPVsim are common in modelling communities, and their use is likely to increase in the future. We hope that the techniques demonstrated in this work provide a framework for modelling communities to perform comprehensive analyses of their simulators, make confident statements about inferential quantities of interest, and provide robust predictions and recommendations for policy makers that can materially improve the landscape of global health.

Acknowledgements

AI, DS, and NM would like to acknowledge the support and funding provided by the Wellcome Trust grant 218261/Z/19/Z; AI is supported by Durham University’s Willmore Fellowship. RGW is funded by the Wellcome Trust (310728/Z/24/Z, 218261/Z/19/Z), NIH (1R01AI147321-01, G-202303-69963, R-202309-71190), EDTCP (RIA208D-2505B), UK MRC (CCF17-7779 via SET Bloomsbury), ESRC (ES/P008011/1), BMGF (INV-004737, INV-035506), and the WHO (2020/985800-0).

References

  • Burd [2003]E.M. Burd,Human papillomavirus and cervical cancer,Clinical microbiology reviews 16(2003) 1–17.
  • Brisson and Drolet [2019]M.Brisson, M.Drolet,Global elimination of cervical cancer as a publichealth problem,The Lancet Oncology 20(2019) 319–321.
  • WHO [2022]WHO, One-dose human papillomavirus (HPV)vaccine offers solid protection against cervical cancer,https://www.who.int/news/item/11-04-2022-one-dose-human-papillomavirus-(hpv)-vaccine-offers-solid-protection-against-cervical-cancer,2022. Accessed: 2023-01-13.
  • Aylett-Bullock etal. [2021]J.Aylett-Bullock, C.Cuesta-Lazaro,A.Quera-Bofarull, M.Icaza-Lizaola,A.Sedgewick, H.Truong,A.Curran, E.Elliott,T.Caulfield, K.Fong, etal.,JUNE: Open-source individual-based epidemiologysimulation,Royal Society open science 8(2021) 210506.
  • Andrianakis etal. [2017]I.Andrianakis, N.McCreesh,I.Vernon, T.J. McKinley,J.E. Oakley, R.N. Nsubuga,M.Goldstein, R.G. White,Efficient history matching of a high dimensionalindividual-based HIV transmission model,SIAM/ASA Journal on Uncertainty Quantification5 (2017) 694–719.
  • Scarponi etal. [2023]D.Scarponi, A.Iskauskas,R.A. Clark, I.Vernon,T.J. McKinley, M.Goldstein,C.Mukandavire, A.Deol,C.Weerasuriya, R.Bakker,R.G. White, N.McCreesh,Demonstrating multi-country calibration of atuberculosis model using new history matching and emulation package - hmer,Epidemics (2023)100678.
  • Stuart etal. [2023]R.M. Stuart, J.A. Cohen,C.C. Kerr, R.G. Abeysuriya,M.Zimmerman, D.W. Rao,M.C. Boudreau, D.J. Klein,HPVsim: An agent-based model of HPV transmissionand cervical disease,medRxiv (2023)2023–02.
  • Akiba etal. [2019]T.Akiba, S.Sano,T.Yanase, T.Ohta,M.Koyama,Optuna: A next-generation hyperparameter optimizationframework,in: Proceedings of the 25th ACM SIGKDDinternational conference on knowledge discovery & data mining,2019, pp. 2623–2631.
  • Panovska-Griffiths etal. [2023]J.Panovska-Griffiths, T.Bayley,T.Ward, A.Das,L.Imeneo, C.Kerr,S.Maskell,Machine learning assisted calibration of stochasticagent-based models for pandemic outbreak analysis,Nature Portfolio, to appear(2023).
  • Gibson and Renshaw [1998]G.J. Gibson, E.Renshaw,Estimating parameters in stochastic compartmentalmodels using markov chain methods,Mathematical Medicine and Biology: A Journal of theIMA 15 (1998) 19–40.
  • O’Neill and Roberts [1999]P.D. O’Neill, G.O. Roberts,Bayesian inference for partially observed stochasticepidemics,Journal of the Royal Statistical Society A162 (1999) 121–129.
  • Jewell etal. [2009]C.P. Jewell, T.Kypraios,P.Neal, G.O. Roberts,Bayesian analysis for emerging infectious diseases,Bayesian analysis 4(2009) 465–496.
  • McKinley etal. [2009]T.McKinley, A.R. Cook,R.Deardon,Inference in epidemic models without likelihoods,The International Journal of Biostatistics5 (2009).
  • McKinley etal. [2018]T.J. McKinley, I.Vernon,I.Andrianakis, N.McCreesh,J.E. Oakley, R.N. Nsubuga,M.Goldstein, R.G. White,Approximate bayesian computation and simulation-basedinference for complex stochastic epidemic models,Statistical science 33(2018) 4–18.
  • Toni etal. [2009]T.Toni, D.Welch,N.Strelkowa, A.Ipsen,M.P. Stumpf,Approximate bayesian computation scheme for parameterinference and model selection in dynamical systems,Journal of the Royal Society Interface6 (2009) 187–202.
  • Kennedy and O’Hagan [2001]M.C. Kennedy, A.O’Hagan,Bayesian calibration of computer models,Journal of the Royal Statistical Society B63 (2001) 425–464.
  • Craig etal. [1997]P.S. Craig, M.Goldstein,A.H. Seheult, J.A. Smith,Pressure matching for hydrocarbon reservoirs: A casestudy in the use of bayes linear strategies for large computer experiments,in: Case studies in Bayesian statistics,Springer-Verlag, 1997, pp.37–93.
  • Vernon etal. [2014]I.Vernon, M.Goldstein,R.Bower,Galaxy formation: Bayesian history matching for theobservable universe,Statistical science (2014)81–90.
  • Chesson etal. [2014]H.W. Chesson, E.F. Dunne,S.Hariri, L.E. Markowitz,The estimated lifetime probability of acquiring humanpapillomavirus in the united states,Sexually Transmitted Diseases 41(2014) 660.
  • Adebamowo etal. [2018]S.N. Adebamowo, A.Famooto,E.O. Dareng, O.Olawande,O.Olaniyan, R.Offiong,C.A. Adebamowo, A.R.G. aspart oftheH3AfricaConsortium,Clearance of type-specific, low-risk, and high-riskcervical human papillomavirus infections in HIV-negative and HIV-positivewomen,Journal of Global Oncology 4(2018) JGO–17.
  • Walboomers etal. [1999]J.M. Walboomers, M.V. Jacobs,M.M. Manos, F.X. Bosch,J.A. Kummer, K.V. Shah,P.J. Snijders, J.Peto,C.J. Meijer, N.Muñoz,Human papillomavirus is a necessary cause of invasivecervical cancer worldwide,The Journal of Pathology 189(1999) 12–19.
  • Kerr etal. [2021]C.C. Kerr, R.M. Stuart,D.Mistry, R.G. Abeysuriya,K.Rosenfeld, G.R. Hart,R.C. Núñez, J.A. Cohen,P.Selvaraj, B.Hagedorn, etal.,Covasim: an agent-based model of COVID-19 dynamicsand interventions,PLOS Computational Biology 17(2021) e1009149.
  • Centre [2021]H.I. Centre, Human papillomavirus andrelated diseases report: Nigeria, 2021.Accessed: 2023-03-15.
  • for DiseaseModelling [2023]I.for DiseaseModelling, HPVsimdocumentation, 2023. Accessed: 2023-06-20.
  • Fernández-Martínez andFernández-Muñiz [2020]J.L. Fernández-Martínez,Z.Fernández-Muñiz,The curse of dimensionality in inverse problems,Journal of Computational and Applied Mathematics369 (2020) 112571.
  • Vernon etal. [2018]I.Vernon, J.Liu,M.Goldstein, J.Rowe,J.Topping, K.Lindsey,Bayesian uncertainty analysis for complex systemsbiology models: Emulation, global parameter searches and evaluation of genefunctions,BMC systems biology 12(2018) 1–29.
  • Santner etal. [2003]T.J. Santner, B.J. Williams,W.I. Notz, B.J. Williams,The Design and Analysis of Computer Experiments,volume1, Springer-Verlag,2003.
  • Bowman and Woods [2016]V.E. Bowman, D.C. Woods,Emulation of multivariate simulators using thin-platesplines with application to atmospheric dispersion,SIAM/ASA Journal on Uncertainty Quantification4 (2016) 1323–1344.
  • Jackson and Woods [2019]S.E. Jackson, D.C. Woods,Bayes linear emulation of simulator networks,arXiv preprint arXiv:1910.08003(2019).
  • Goldstein etal. [2013]M.Goldstein, A.Seheult,I.Vernon,Assessing model adequacy,Environmental Modelling: Finding Simplicity inComplexity (2013) 435–449.
  • Goldstein and Wooff [2007]M.Goldstein, D.Wooff,Bayes Linear Statistics: Theory and Methods, volume716, John Wiley & Sons,2007.
  • Iskauskas and McKinley [2023]A.Iskauskas, T.J. McKinley,hmer package, 2023. URL: https://CRAN.R-project.org/package=hmer, accessed:2024-06-27.
  • Iskauskas etal. [2024]A.Iskauskas, I.Vernon,M.Goldstein, D.Scarponi,N.McCreesh, T.J. McKinley,R.G. White,Emulation and history matching using the hmerpackage,Journal of Statistical Software109 (2024) 1–48.
  • for Researchon Cancer [2020]I.A. for Researchon Cancer, Global cancerobservatory: Cancer today,https://gco.iarc.fr/today/, 2020.Accessed 2023-03-13.
  • Bruni etal. [2019]L.Bruni, G.Albero,B.Serrano, M.Mena,D.Gómez, J.Muñoz, etal.,Human papillomavirus and related diseases report,nigeria,L’Hospitalet de Llobregat (2019).
  • USAID [2023]USAID, Demographic and health surveysprogram, 2023. Accessed: 2023-10-17.
  • Loeppky etal. [2009]J.L. Loeppky, J.Sacks,W.J. Welch,Choosing the sample size of a computer experiment: Apractical guide,Technometrics 51(2009) 366–376.
  • Williamson and Vernon [2013]D.Williamson, I.Vernon,Implausibility driven evolutionary monte carlo forefficient generation of uniform and optimal designs for multi-wave computerexperiments,arXiv preprint arXiv:1309.3520(2013).
  • Marth etal. [2017]C.Marth, F.Landoni,S.Mahner, M.McCormack,A.Gonzalez-Martin, N.Colombo,Cervical cancer: ESMO clinical practice guidelinesfor diagnosis, treatment and follow-up,Annals of Oncology 28(2017) iv72–iv83.
  • Jackson [2018]S.E. Jackson, Design of Physical SystemExperiments Using Bayes Linear Emulation and History Matching Methodologywith Application to Arabidopsis Thaliana, Ph.D. thesis, Durham University,2018.
  • Vernon and Goldstein [2023]I.Vernon, M.Goldstein,Bayes linear emulation and history matching ofstochastic systems biology models,In preparation (2023).
  • Rasmussen [2003]C.E. Rasmussen,Gaussian processes in machine learning,in: Summer School on Machine Learning,Springer-Verlag, 2003, pp.63–71.
  • Andrianakis and Challenor [2012]I.Andrianakis, P.G. Challenor,The effect of the nugget on gaussian processemulators of computer models,Computational Statistics & Data Analysis56 (2012) 4215–4228.
  • Maatouk etal. [2015]H.Maatouk, O.Roustant,Y.Richet,Cross-validation estimations of hyper-parameters ofgaussian processes with inequality constraints,Procedia Environmental Sciences27 (2015) 38–44.
  • Conti etal. [2009]S.Conti, J.P. Gosling,J.E. Oakley, A.O’Hagan,Gaussian process emulation of dynamic computercodes,Biometrika 96(2009) 663–676.
  • Vernon etal. [2010]I.Vernon, M.Goldstein,R.Bower,Galaxy formation: A bayesian uncertainty analysis,Bayesian analysis 5(2010) 619–669.
  • Brynjarsdottir and O’Hagan [2014]J.Brynjarsdottir, A.O’Hagan,Learning about physical parameters: The importance ofmodel discrepancy,Inverse problems 30(2014) 114007.

Appendix A Bayes Linear Emulation of Stochastic Models

A.1 Bayes Linear Emulation

Here we provide details of the process behind emulation, particularly as applied to emulating the stochasticity of agent-based models. Further details may be found in [41, 31].

Consider a simulation of a real-world process y𝑦yitalic_y which takes a set of input parameters, described as a vector x𝑥xitalic_x of length d𝑑ditalic_d, and returns a set of m𝑚mitalic_m outputs {fi(x)}i=1,,msubscriptsubscript𝑓𝑖𝑥𝑖1𝑚\{f_{i}(x)\}_{i=1,\dots,m}{ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) } start_POSTSUBSCRIPT italic_i = 1 , … , italic_m end_POSTSUBSCRIPT. A univariate emulator is a fast statistical approximation of one of the simulator outputs, built using a comparatively small collection of simulator runs, which provides an expected value for the simulator output at any unseen points x𝑥xitalic_x in the parameter space along with a corresponding estimate of the uncertainty of the prediction, reflecting our beliefs about the simulator in question.

Concretely, we create a prior representation of a given simulator output fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) in emulator form as

gi(x)=j=1piβijhij(xAi)+ui(xAi)+wi(x).subscript𝑔𝑖𝑥superscriptsubscript𝑗1subscript𝑝𝑖subscript𝛽𝑖𝑗subscript𝑖𝑗subscript𝑥subscript𝐴𝑖subscript𝑢𝑖subscript𝑥subscript𝐴𝑖subscript𝑤𝑖𝑥g_{i}(x)=\sum_{j=1}^{p_{i}}\beta_{ij}h_{ij}(x_{A_{i}})+u_{i}(x_{A_{i}})+w_{i}(%x).italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) = ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_p start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUPERSCRIPT italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT italic_h start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) + italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) + italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) .(2)

Here xAisubscript𝑥subscript𝐴𝑖x_{A_{i}}italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT, where Ai{1,,d}subscript𝐴𝑖1𝑑A_{i}\subseteq\{1,\dots,d\}italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ⊆ { 1 , … , italic_d }, are the set of ‘active variables’ for output fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ); that is, the components of the input vector x𝑥xitalic_x that are most influential in determining the behaviour of fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ). The hijsubscript𝑖𝑗h_{ij}italic_h start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT are pisubscript𝑝𝑖p_{i}italic_p start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT known simple functions of the xAisubscript𝑥subscript𝐴𝑖x_{A_{i}}italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT, with βijsubscript𝛽𝑖𝑗\beta_{ij}italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT the corresponding coefficients – together these two terms define a regression surface encoding the global response of fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) to the inputs. ui(xAi)subscript𝑢𝑖subscript𝑥subscript𝐴𝑖u_{i}(x_{A_{i}})italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) is a second-order weakly stationary process which captures residual variation in the active variables and can be seen as governing the local behaviour of the simulator output. While the regression functions hijsubscript𝑖𝑗h_{ij}italic_h start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT are considered known, we view their coefficients and the second-order process as being unknown and therefore treat them as random variables. We assume the following covariance structure for ui(xAi)subscript𝑢𝑖subscript𝑥subscript𝐴𝑖u_{i}(x_{A_{i}})italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ):

Cov[ui(xAi),ui(xAi)]=(1δi)σi2r(xAi,xAi).Covsubscript𝑢𝑖subscript𝑥subscript𝐴𝑖subscript𝑢𝑖subscriptsuperscript𝑥subscript𝐴𝑖1subscript𝛿𝑖superscriptsubscript𝜎𝑖2𝑟subscript𝑥subscript𝐴𝑖subscriptsuperscript𝑥subscript𝐴𝑖\text{Cov}[u_{i}(x_{A_{i}}),u_{i}(x^{\prime}_{A_{i}})]=(1-\delta_{i})\sigma_{i%}^{2}r(x_{A_{i}},x^{\prime}_{A_{i}}).Cov [ italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) , italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) ] = ( 1 - italic_δ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ) italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT italic_r ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) .

r(x,x)𝑟𝑥superscript𝑥r(x,x^{\prime})italic_r ( italic_x , italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) is a correlation function, suitably chosen depending on our beliefs about the output (common examples and their usage can be found in [42]); our choice often depends on how ‘smooth’ we expect the output response to be with respect to our input parameters and the extent to which neighbouring points in parameter space influence the point x𝑥xitalic_x in question. Our overall variance σi2subscriptsuperscript𝜎2𝑖\sigma^{2}_{i}italic_σ start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT represents our prior uncertainty about our predictions, and δi[0,1]subscript𝛿𝑖01\delta_{i}\in[0,1]italic_δ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ∈ [ 0 , 1 ] accounts for the residual effect of the inactive variables on the simulator output. This is accounted for in the ‘nugget’ term wi(x)subscript𝑤𝑖𝑥w_{i}(x)italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ), which has correlation structure

Cov[wi(x),wi(x)]=σi2δiIx=xCovsubscript𝑤𝑖𝑥subscript𝑤𝑖superscript𝑥superscriptsubscript𝜎𝑖2subscript𝛿𝑖subscript𝐼𝑥superscript𝑥\text{Cov}[w_{i}(x),w_{i}(x^{\prime})]=\sigma_{i}^{2}\delta_{i}I_{x=x^{\prime}}Cov [ italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) ] = italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT italic_δ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT italic_I start_POSTSUBSCRIPT italic_x = italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT end_POSTSUBSCRIPT

with Ix=xsubscript𝐼𝑥superscript𝑥I_{x=x^{\prime}}italic_I start_POSTSUBSCRIPT italic_x = italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT end_POSTSUBSCRIPT is an indicator function taking the value 1111 if x=x𝑥superscript𝑥x=x^{\prime}italic_x = italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT and 00 otherwise.

If we further assume that the regression coefficients βijsubscript𝛽𝑖𝑗\beta_{ij}italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT, second-order process uisubscript𝑢𝑖u_{i}italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT, and nugget term wisubscript𝑤𝑖w_{i}italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT are mutually uncorrelated, that 𝔼[ui(x)]=𝔼[wi(x)]=0𝔼delimited-[]subscript𝑢𝑖𝑥𝔼delimited-[]subscript𝑤𝑖𝑥0\mathbb{E}[u_{i}(x)]=\mathbb{E}[w_{i}(x)]=0blackboard_E [ italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] = blackboard_E [ italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] = 0 for all x𝑥xitalic_x, and that Var[βij]=0Vardelimited-[]subscript𝛽𝑖𝑗0\text{Var}[\beta_{ij}]=0Var [ italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT ] = 0 [30], then we see that our emulator prediction at a point x𝑥xitalic_x is determined by the regression surface, with variance in that prediction given by σ2superscript𝜎2\sigma^{2}italic_σ start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT:

𝔼[gi(x)]=j=1pi𝔼[βij]hij(xAi),Var[gi(x)]=(1δi)σi2+δiσi2=σi2.formulae-sequence𝔼delimited-[]subscript𝑔𝑖𝑥superscriptsubscript𝑗1subscript𝑝𝑖𝔼delimited-[]subscript𝛽𝑖𝑗subscript𝑖𝑗subscript𝑥subscript𝐴𝑖Vardelimited-[]subscript𝑔𝑖𝑥1subscript𝛿𝑖superscriptsubscript𝜎𝑖2subscript𝛿𝑖superscriptsubscript𝜎𝑖2superscriptsubscript𝜎𝑖2\mathbb{E}[g_{i}(x)]=\sum_{j=1}^{p_{i}}\mathbb{E}[\beta_{ij}]h_{ij}(x_{A_{i}})%,\hskip 14.22636pt\text{Var}[g_{i}(x)]=(1-\delta_{i})\sigma_{i}^{2}+\delta_{i}%\sigma_{i}^{2}=\sigma_{i}^{2}.blackboard_E [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] = ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_p start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUPERSCRIPT blackboard_E [ italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT ] italic_h start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) , Var [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] = ( 1 - italic_δ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ) italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT + italic_δ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT = italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT .

The choice of a zero-variance regression surface decouples the global behaviour of the simulator from the weakly-stationary process, allowing the latter to focus on representing the variation of the residuals, providing a clear separation between the global and local behaviour and allowing for a convenient representation of our beliefs about the output via the regression surface and our uncertainty about those beliefs.

Before we can update the emulated structure with respect to data, we must complete the a priori specifiction for the quantities βijsubscript𝛽𝑖𝑗\beta_{ij}italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT, ui(xAi)subscript𝑢𝑖subscript𝑥subscript𝐴𝑖u_{i}(x_{A_{i}})italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUBSCRIPT italic_A start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ), and wi(x)subscript𝑤𝑖𝑥w_{i}(x)italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ). This can be done in a variety of ways: for instance, if one is willing and able to specify full distributions for these quantities, we could then use maximum likelihood (ML) or maximum a posteriori (MAP) estimates to determine plug-in values for their hyperparameters [43], or use cross-validation [44]. However, we may not be able (or willing) to make full distributional specifications for these quantities, whether due to a lack of prior knowledge required for such a specification or a lack of faith in such statements. It is rare, however, that we are similarly hampered in making statements about the second-order structure of the system – expectations and (co)variances – particularly in light of the emulator structure described above. The Bayes linear framework requires only these quantities, and we proceed accordingly. Explicitly, then, we focus on specifying the quantities 𝔼[βij]𝔼delimited-[]subscript𝛽𝑖𝑗\mathbb{E}[\beta_{ij}]blackboard_E [ italic_β start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT ], σisubscript𝜎𝑖\sigma_{i}italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT, δisubscript𝛿𝑖\delta_{i}italic_δ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT, and any hyperparameters that govern the behaviour of the correlation function r(x,x)𝑟𝑥superscript𝑥r(x,x^{\prime})italic_r ( italic_x , italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ). These quantities can be determined via a full a priori specification or by using pragmatic plug-in estimates [27, 16, 42].

We may now use data obtained from simulator runs to update our beliefs in light of data. Suppose that we have data obtained from running the simulator at a series of points (x(1),x(2),,x(n))superscript𝑥1superscript𝑥2superscript𝑥𝑛(x^{(1)},x^{(2)},\dots,x^{(n)})( italic_x start_POSTSUPERSCRIPT ( 1 ) end_POSTSUPERSCRIPT , italic_x start_POSTSUPERSCRIPT ( 2 ) end_POSTSUPERSCRIPT , … , italic_x start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT ), resulting in a collection of outputs444For a stochastic simulator, we would obtain multiple output values for a given input parameter x𝑥xitalic_x; we will discuss this point momentarily.

Di=(fi(x(1)),,fi(x(n))).subscript𝐷𝑖subscript𝑓𝑖superscript𝑥1subscript𝑓𝑖superscript𝑥𝑛D_{i}=\left(f_{i}(x^{(1)}),\dots,f_{i}(x^{(n)})\right).italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = ( italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ( 1 ) end_POSTSUPERSCRIPT ) , … , italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT ) ) .

If we had a full distributional statement about our emulator quantities, then we would produce a posterior estimate using Bayes’ Theorem; the equivalent in our case is given by the Bayes linear update formulae [31]. The Bayes-adjusted emulator prediction for gi(x)subscript𝑔𝑖𝑥g_{i}(x)italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) at new points x𝑥xitalic_x, xsuperscript𝑥x^{\prime}italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT is given by

𝔼Di[gi(x)]=subscript𝔼subscript𝐷𝑖delimited-[]subscript𝑔𝑖𝑥absent\displaystyle\mathbb{E}_{D_{i}}[g_{i}(x)]=blackboard_E start_POSTSUBSCRIPT italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] =𝔼[gi(x)]+Cov[gi(x),Di]Var[Di]1(Di𝔼[Di]),𝔼delimited-[]subscript𝑔𝑖𝑥Covsubscript𝑔𝑖𝑥subscript𝐷𝑖Varsuperscriptdelimited-[]subscript𝐷𝑖1subscript𝐷𝑖𝔼delimited-[]subscript𝐷𝑖\displaystyle\mathbb{E}[g_{i}(x)]+\text{Cov}[g_{i}(x),D_{i}]\text{Var}[D_{i}]^%{-1}(D_{i}-\mathbb{E}[D_{i}]),blackboard_E [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] + Cov [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] Var [ italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] start_POSTSUPERSCRIPT - 1 end_POSTSUPERSCRIPT ( italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT - blackboard_E [ italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] ) ,(3)
CovDi[gi(x),gi(x)]=subscriptCovsubscript𝐷𝑖subscript𝑔𝑖𝑥subscript𝑔𝑖superscript𝑥absent\displaystyle\text{Cov}_{D_{i}}[g_{i}(x),g_{i}(x^{\prime})]=Cov start_POSTSUBSCRIPT italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) ] =Cov[gi(x),gi(x)]Cov[gi(x),Di]Var[Di]1Cov[Di,gi(x)]Covsubscript𝑔𝑖𝑥subscript𝑔𝑖superscript𝑥Covsubscript𝑔𝑖𝑥subscript𝐷𝑖Varsuperscriptdelimited-[]subscript𝐷𝑖1Covsubscript𝐷𝑖subscript𝑔𝑖superscript𝑥\displaystyle\text{Cov}[g_{i}(x),g_{i}(x^{\prime})]-\text{Cov}[g_{i}(x),D_{i}]%\text{Var}[D_{i}]^{-1}\text{Cov}[D_{i},g_{i}(x^{\prime})]Cov [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) ] - Cov [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] Var [ italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] start_POSTSUPERSCRIPT - 1 end_POSTSUPERSCRIPT Cov [ italic_D start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) ](4)

where here 𝔼D[gi(x)]subscript𝔼𝐷delimited-[]subscript𝑔𝑖𝑥\mathbb{E}_{D}[g_{i}(x)]blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] is the expectation or prediction for the output in question, updated with respect to the data D𝐷Ditalic_D, CovD[gi(x),gi(x)]subscriptCov𝐷subscript𝑔𝑖𝑥subscript𝑔𝑖superscript𝑥\text{Cov}_{D}[g_{i}(x),g_{i}(x^{\prime})]Cov start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT ) ] is the adjusted covariance, and VarD[g(x)]=CovD[gi(x),gi(x)]subscriptVar𝐷delimited-[]𝑔𝑥subscriptCov𝐷subscript𝑔𝑖𝑥subscript𝑔𝑖𝑥\text{Var}_{D}[g(x)]=\text{Cov}_{D}[g_{i}(x),g_{i}(x)]Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g ( italic_x ) ] = Cov start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) , italic_g start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] is the adjusted variance corresponding to this prediction.

We therefore obtain a prediction and corresponding uncertainty for the value of fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) at a point x𝑥xitalic_x whose simulator output has not been evaluated. The uncertainty depends on our prior beliefs and the proximity of data obtained to our unseen point; indeed one can show that for a deterministic simulator, the posterior prediction at a ‘training’ point x(k)superscript𝑥𝑘x^{(k)}italic_x start_POSTSUPERSCRIPT ( italic_k ) end_POSTSUPERSCRIPT is identical to the simulator output f(x(k))𝑓superscript𝑥𝑘f(x^{(k)})italic_f ( italic_x start_POSTSUPERSCRIPT ( italic_k ) end_POSTSUPERSCRIPT ) with vanishing posterior variance in the absence of a nugget term. Crucially, the quantities (3) and (4) are extremely fast to evaluate, so the emulators allow us to perform extensive exploration of the simulator’s behaviour over the input space.

As a final note, were we to have assumed normal and Gaussian process priors for β𝛽\betaitalic_β and u(x)𝑢𝑥u(x)italic_u ( italic_x ), respectively, then the approach here bears close similarity to Gaussian process emulation [45]. However, Gaussian process emulation requires invoking additional distributional assumptions that may be difficult to justify, force stricter and more complex diagnostics to be satisfied, and materially affect our final inference. The Bayes linear approach allows us to circumvent these concerns and provide meaningful statements about the behaviour of our simulator without invoking potentially unjustified assumptions about the underlying distribution.

A.2 Covariance Emulation

As alluded to, the process of adjusting our emulators above presupposes that we obtain a single output value from application of the simulator to each parameter set x(i)superscript𝑥𝑖x^{(i)}italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT. For a deterministic model, this is all that is required to understand the output surface at the points of interest; for a stochastic model, however, repeated evaluations of the simulator at the same parameter set will not be identical. We therefore propose a modification to the framework presented above.

Suppose that we have two simulator outputs fa(x)subscript𝑓𝑎𝑥f_{a}(x)italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) and fb(x)subscript𝑓𝑏𝑥f_{b}(x)italic_f start_POSTSUBSCRIPT italic_b end_POSTSUBSCRIPT ( italic_x ) for a parameter set x𝑥xitalic_x. Performing n𝑛nitalic_n evaluations of the simulator at x𝑥xitalic_x gives rise to realisations (fa(1)(x),fa(2)(x),fa(n)(x))subscriptsuperscript𝑓1𝑎𝑥subscriptsuperscript𝑓2𝑎𝑥subscriptsuperscript𝑓𝑛𝑎𝑥(f^{(1)}_{a}(x),f^{(2)}_{a}(x),\dots f^{(n)}_{a}(x))( italic_f start_POSTSUPERSCRIPT ( 1 ) end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) , italic_f start_POSTSUPERSCRIPT ( 2 ) end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) , … italic_f start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ), and similarly for output b𝑏bitalic_b. For each output, we assume that realisations are second-order exchangeable, so that

fa(k)(x)=(fa(x))+k(fa(x)),k=1,2,,n.formulae-sequencesubscriptsuperscript𝑓𝑘𝑎𝑥subscript𝑓𝑎𝑥subscript𝑘subscript𝑓𝑎𝑥𝑘12𝑛f^{(k)}_{a}(x)=\mathcal{M}(f_{a}(x))+\mathcal{R}_{k}(f_{a}(x)),\hskip 5.69054%ptk=1,2,\dots,n.italic_f start_POSTSUPERSCRIPT ( italic_k ) end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) = caligraphic_M ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) + caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) , italic_k = 1 , 2 , … , italic_n .

(fa(x))subscript𝑓𝑎𝑥\mathcal{M}(f_{a}(x))caligraphic_M ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) can be thought of as the true mean of the output fa(x)subscript𝑓𝑎𝑥f_{a}(x)italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) that would be observed were we to perform infinitely many evaluations at the point x𝑥xitalic_x, while k(fa(x))subscript𝑘subscript𝑓𝑎𝑥\mathcal{R}_{k}(f_{a}(x))caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) represents the residual variability due to the kthsuperscript𝑘𝑡k^{th}italic_k start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT realisation from this mean. We may extend the concept of second-order exchangeability to the residuals; for outputs a𝑎aitalic_a and b𝑏bitalic_b we define

k(fa(x))k(fb(x))Qab(k)(x)=(Qab(x))+k(Qab(x)),subscript𝑘subscript𝑓𝑎𝑥subscript𝑘subscript𝑓𝑏𝑥subscriptsuperscript𝑄𝑘𝑎𝑏𝑥subscript𝑄𝑎𝑏𝑥subscript𝑘subscript𝑄𝑎𝑏𝑥\mathcal{R}_{k}(f_{a}(x))\mathcal{R}_{k}(f_{b}(x))\equiv Q^{(k)}_{ab}(x)=%\mathcal{M}(Q_{ab}(x))+\mathcal{R}_{k}(Q_{ab}(x)),caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_b end_POSTSUBSCRIPT ( italic_x ) ) ≡ italic_Q start_POSTSUPERSCRIPT ( italic_k ) end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) = caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) + caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) ,

where now Qabsubscript𝑄𝑎𝑏Q_{ab}italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT represents the true covariance between outputs a𝑎aitalic_a and b𝑏bitalic_b. Note that, if a=b𝑎𝑏a=bitalic_a = italic_b then Qab(k)(x)=(k(fa(x)))2subscriptsuperscript𝑄𝑘𝑎𝑏𝑥superscriptsubscript𝑘subscript𝑓𝑎𝑥2Q^{(k)}_{ab}(x)=\left(\mathcal{R}_{k}(f_{a}(x))\right)^{2}italic_Q start_POSTSUPERSCRIPT ( italic_k ) end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) = ( caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) ) start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT and the specifications are identical to those proposed in [41].

It is not feasible to observe the true mean from the simulator, nor the true covariance matrix: for finitely many realisations at the point x𝑥xitalic_x we may obtain the sample means and covariances. Let the sample covariance obtained from n𝑛nitalic_n realisations be denoted qab,nsubscript𝑞𝑎𝑏𝑛q_{ab,n}italic_q start_POSTSUBSCRIPT italic_a italic_b , italic_n end_POSTSUBSCRIPT: then computation similar to that given in [31] yields the following relation:

qab,n(x)=1nk=1n((Qab(x))+k(Qab(x)))1n(n1)klk(fa(x))l(fb(x))(Qab(x))+Tab(x).subscript𝑞𝑎𝑏𝑛𝑥1𝑛superscriptsubscript𝑘1𝑛subscript𝑄𝑎𝑏𝑥subscript𝑘subscript𝑄𝑎𝑏𝑥1𝑛𝑛1subscript𝑘𝑙subscript𝑘subscript𝑓𝑎𝑥subscript𝑙subscript𝑓𝑏𝑥subscript𝑄𝑎𝑏𝑥subscript𝑇𝑎𝑏𝑥q_{ab,n}(x)=\frac{1}{n}\sum_{k=1}^{n}\left(\mathcal{M}(Q_{ab}(x))+\mathcal{R}_%{k}(Q_{ab}(x))\right)-\frac{1}{n(n-1)}\sum_{k\neq l}\mathcal{R}_{k}(f_{a}(x))%\mathcal{R}_{l}(f_{b}(x))\equiv\mathcal{M}(Q_{ab}(x))+T_{ab}(x).italic_q start_POSTSUBSCRIPT italic_a italic_b , italic_n end_POSTSUBSCRIPT ( italic_x ) = divide start_ARG 1 end_ARG start_ARG italic_n end_ARG ∑ start_POSTSUBSCRIPT italic_k = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT ( caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) + caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) ) - divide start_ARG 1 end_ARG start_ARG italic_n ( italic_n - 1 ) end_ARG ∑ start_POSTSUBSCRIPT italic_k ≠ italic_l end_POSTSUBSCRIPT caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) caligraphic_R start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_b end_POSTSUBSCRIPT ( italic_x ) ) ≡ caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) + italic_T start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) .

Hence the sample covariance is related to the true covariance (Qab(x))subscript𝑄𝑎𝑏𝑥\mathcal{M}(Q_{ab}(x))caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) via the additional term Tab(x)subscript𝑇𝑎𝑏𝑥T_{ab}(x)italic_T start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ), where explicitly

Tab(x)=1nk=1nk(Qab(x))1n(n1)klk(fa(x))k(fb(x))subscript𝑇𝑎𝑏𝑥1𝑛superscriptsubscript𝑘1𝑛subscript𝑘subscript𝑄𝑎𝑏𝑥1𝑛𝑛1subscript𝑘𝑙subscript𝑘subscript𝑓𝑎𝑥subscript𝑘subscript𝑓𝑏𝑥T_{ab}(x)=\frac{1}{n}\sum_{k=1}^{n}\mathcal{R}_{k}(Q_{ab}(x))-\frac{1}{n(n-1)}%\sum_{k\neq l}\mathcal{R}_{k}(f_{a}(x))\mathcal{R}_{k}(f_{b}(x))italic_T start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) = divide start_ARG 1 end_ARG start_ARG italic_n end_ARG ∑ start_POSTSUBSCRIPT italic_k = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) - divide start_ARG 1 end_ARG start_ARG italic_n ( italic_n - 1 ) end_ARG ∑ start_POSTSUBSCRIPT italic_k ≠ italic_l end_POSTSUBSCRIPT caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_b end_POSTSUBSCRIPT ( italic_x ) )

can be thought of as a correction term that reconciles the two quantities. Under the assumptions of second-order exchangeability; namely that residuals are mutually uncorrelated and uncorrelated with the true mean and covariance, we find that the additional variability due to finite sample size is

VTab(x)Var[Tab(x)]=1nCab,R(C)(x)+1n(n1)(Cov[(Qaa(x)),(Qbb(x))]+Caa(x)Cbb(x)+Cab,M(x)+Cab(x)2),subscriptsuperscript𝑉𝑎𝑏𝑇𝑥Vardelimited-[]subscript𝑇𝑎𝑏𝑥1𝑛subscript𝐶𝑎𝑏𝑅𝐶𝑥1𝑛𝑛1Covsubscript𝑄𝑎𝑎𝑥subscript𝑄𝑏𝑏𝑥subscript𝐶𝑎𝑎𝑥subscript𝐶𝑏𝑏𝑥subscript𝐶𝑎𝑏𝑀𝑥subscript𝐶𝑎𝑏superscript𝑥2V^{ab}_{T}(x)\equiv\text{Var}[T_{ab}(x)]=\frac{1}{n}C_{ab,R(C)}(x)+\frac{1}{n(%n-1)}\left(\text{Cov}[\mathcal{M}(Q_{aa}(x)),\mathcal{M}(Q_{bb}(x))]\right.%\left.+C_{aa}(x)C_{bb}(x)+C_{ab,M}(x)+C_{ab}(x)^{2}\right),italic_V start_POSTSUPERSCRIPT italic_a italic_b end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_T end_POSTSUBSCRIPT ( italic_x ) ≡ Var [ italic_T start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ] = divide start_ARG 1 end_ARG start_ARG italic_n end_ARG italic_C start_POSTSUBSCRIPT italic_a italic_b , italic_R ( italic_C ) end_POSTSUBSCRIPT ( italic_x ) + divide start_ARG 1 end_ARG start_ARG italic_n ( italic_n - 1 ) end_ARG ( Cov [ caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_a end_POSTSUBSCRIPT ( italic_x ) ) , caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_b italic_b end_POSTSUBSCRIPT ( italic_x ) ) ] + italic_C start_POSTSUBSCRIPT italic_a italic_a end_POSTSUBSCRIPT ( italic_x ) italic_C start_POSTSUBSCRIPT italic_b italic_b end_POSTSUBSCRIPT ( italic_x ) + italic_C start_POSTSUBSCRIPT italic_a italic_b , italic_M end_POSTSUBSCRIPT ( italic_x ) + italic_C start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ) ,

where for ease of notation we have defined

Cab(x)subscript𝐶𝑎𝑏𝑥\displaystyle C_{ab}(x)italic_C start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x )=𝔼[(Qab(x)],\displaystyle=\mathbb{E}[\mathcal{M}(Q_{ab}(x)],= blackboard_E [ caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ] ,
Cab,M(x)subscript𝐶𝑎𝑏𝑀𝑥\displaystyle C_{ab,M}(x)italic_C start_POSTSUBSCRIPT italic_a italic_b , italic_M end_POSTSUBSCRIPT ( italic_x )=Var[(Qab(x)],\displaystyle=\text{Var}[\mathcal{M}(Q_{ab}(x)],= Var [ caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ] ,
Cab,R(C)(x)subscript𝐶𝑎𝑏𝑅𝐶𝑥\displaystyle C_{ab,R(C)}(x)italic_C start_POSTSUBSCRIPT italic_a italic_b , italic_R ( italic_C ) end_POSTSUBSCRIPT ( italic_x )=Var[k(Qab(x))].absentVardelimited-[]subscript𝑘subscript𝑄𝑎𝑏𝑥\displaystyle=\text{Var}[\mathcal{R}_{k}(Q_{ab}(x))].= Var [ caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ) ] .

Note that, for increasing n𝑛nitalic_n, the contribution of the correction term decreases; in the limit as n𝑛n\to\inftyitalic_n → ∞, the correction term vanishes as we would expect. Furthermore, the nature of the correction term is dominated by the residuals on the covariance structure, k(Qab(x))subscript𝑘subscript𝑄𝑎𝑏𝑥\mathcal{R}_{k}(Q_{ab}(x))caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_Q start_POSTSUBSCRIPT italic_a italic_b end_POSTSUBSCRIPT ( italic_x ) ), not the residuals on the mean surface; though both play a part for modest n𝑛nitalic_n.

Given determinations about the relevant prior quantities, and with the corresponding emulators for the mean and variance constructed using those priors, the update formulae presented in (3) and (4) follow straightforwardly, save for one key aspect. For simplicity, consider only two simulator outputs and suppose that we have data points (x(1),,x(m))superscript𝑥1superscript𝑥𝑚(x^{(1)},\dots,x^{(m)})( italic_x start_POSTSUPERSCRIPT ( 1 ) end_POSTSUPERSCRIPT , … , italic_x start_POSTSUPERSCRIPT ( italic_m ) end_POSTSUPERSCRIPT ) upon each of which we perform nlsubscript𝑛𝑙n_{l}italic_n start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT repetitions, l=1,,m𝑙1𝑚l=1,\dots,mitalic_l = 1 , … , italic_m giving rise to sample means Du=(μu(1),,μu(m))subscript𝐷𝑢superscriptsubscript𝜇𝑢1superscriptsubscript𝜇𝑢𝑚D_{u}=(\mu_{u}^{(1)},\dots,\mu_{u}^{(m)})italic_D start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT = ( italic_μ start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( 1 ) end_POSTSUPERSCRIPT , … , italic_μ start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ( italic_m ) end_POSTSUPERSCRIPT ) and sample (co)variances Suv=(qn1uv(x1),,qnmuv(xm))subscript𝑆𝑢𝑣superscriptsubscript𝑞subscript𝑛1𝑢𝑣subscript𝑥1superscriptsubscript𝑞subscript𝑛𝑚𝑢𝑣subscript𝑥𝑚S_{uv}=(q_{n_{1}}^{uv}(x_{1}),\dots,q_{n_{m}}^{uv}(x_{m}))italic_S start_POSTSUBSCRIPT italic_u italic_v end_POSTSUBSCRIPT = ( italic_q start_POSTSUBSCRIPT italic_n start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT ( italic_x start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ) , … , italic_q start_POSTSUBSCRIPT italic_n start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT ( italic_x start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT ) ) for u,v{a,b}𝑢𝑣𝑎𝑏u,v\in\{a,b\}italic_u , italic_v ∈ { italic_a , italic_b }. Let VM,iuv=Var[(Quv(x(i)))]superscriptsubscript𝑉𝑀𝑖𝑢𝑣Vardelimited-[]subscript𝑄𝑢𝑣superscript𝑥𝑖V_{M,i}^{uv}=\text{Var}[\mathcal{M}(Q_{uv}(x^{(i)}))]italic_V start_POSTSUBSCRIPT italic_M , italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT = Var [ caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_u italic_v end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT ) ) ] and CM,ijuv=Cov[(Quv(x(i))),(Quv(x(j)))]superscriptsubscript𝐶𝑀𝑖𝑗𝑢𝑣Covsubscript𝑄𝑢𝑣superscript𝑥𝑖subscript𝑄𝑢𝑣superscript𝑥𝑗C_{M,ij}^{uv}=\text{Cov}[\mathcal{M}(Q_{uv}(x^{(i)})),\mathcal{M}(Q_{uv}(x^{(j%)}))]italic_C start_POSTSUBSCRIPT italic_M , italic_i italic_j end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT = Cov [ caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_u italic_v end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ( italic_i ) end_POSTSUPERSCRIPT ) ) , caligraphic_M ( italic_Q start_POSTSUBSCRIPT italic_u italic_v end_POSTSUBSCRIPT ( italic_x start_POSTSUPERSCRIPT ( italic_j ) end_POSTSUPERSCRIPT ) ) ], then for each element of the covariance matrix we seek to emulate we have

Var[Suv]=(VM,1uv+VT,1uvCM,12uvCM,1muvCM,21uvVM,2uv+VT,2uvCM,2muvCM,1muvCM,2muvVM,muv+VT,muv).Vardelimited-[]subscript𝑆𝑢𝑣matrixsuperscriptsubscript𝑉𝑀1𝑢𝑣superscriptsubscript𝑉𝑇1𝑢𝑣superscriptsubscript𝐶𝑀12𝑢𝑣superscriptsubscript𝐶𝑀1𝑚𝑢𝑣superscriptsubscript𝐶𝑀21𝑢𝑣superscriptsubscript𝑉𝑀2𝑢𝑣superscriptsubscript𝑉𝑇2𝑢𝑣superscriptsubscript𝐶𝑀2𝑚𝑢𝑣superscriptsubscript𝐶𝑀1𝑚𝑢𝑣superscriptsubscript𝐶𝑀2𝑚𝑢𝑣superscriptsubscript𝑉𝑀𝑚𝑢𝑣superscriptsubscript𝑉𝑇𝑚𝑢𝑣\text{Var}[S_{uv}]=\begin{pmatrix}V_{M,1}^{uv}+V_{T,1}^{uv}&C_{M,12}^{uv}&%\dots&C_{M,1m}^{uv}\\C_{M,21}^{uv}&V_{M,2}^{uv}+V_{T,2}^{uv}&\dots&C_{M,2m}^{uv}\\\vdots&\vdots&\ddots&\vdots\\C_{M,1m}^{uv}&C_{M,2m}^{uv}&\cdots&V_{M,m}^{uv}+V_{T,m}^{uv}\end{pmatrix}.Var [ italic_S start_POSTSUBSCRIPT italic_u italic_v end_POSTSUBSCRIPT ] = ( start_ARG start_ROW start_CELL italic_V start_POSTSUBSCRIPT italic_M , 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT + italic_V start_POSTSUBSCRIPT italic_T , 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL start_CELL italic_C start_POSTSUBSCRIPT italic_M , 12 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL start_CELL … end_CELL start_CELL italic_C start_POSTSUBSCRIPT italic_M , 1 italic_m end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL end_ROW start_ROW start_CELL italic_C start_POSTSUBSCRIPT italic_M , 21 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL start_CELL italic_V start_POSTSUBSCRIPT italic_M , 2 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT + italic_V start_POSTSUBSCRIPT italic_T , 2 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL start_CELL … end_CELL start_CELL italic_C start_POSTSUBSCRIPT italic_M , 2 italic_m end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL end_ROW start_ROW start_CELL ⋮ end_CELL start_CELL ⋮ end_CELL start_CELL ⋱ end_CELL start_CELL ⋮ end_CELL end_ROW start_ROW start_CELL italic_C start_POSTSUBSCRIPT italic_M , 1 italic_m end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL start_CELL italic_C start_POSTSUBSCRIPT italic_M , 2 italic_m end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL start_CELL ⋯ end_CELL start_CELL italic_V start_POSTSUBSCRIPT italic_M , italic_m end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT + italic_V start_POSTSUBSCRIPT italic_T , italic_m end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_u italic_v end_POSTSUPERSCRIPT end_CELL end_ROW end_ARG ) .

With this information in hand, we may perform the Bayes linear update (3) to obtain a posterior prediction for the elements of the covariance matrix. Once obtained, the mean data matrix D𝐷Ditalic_D obtains a similar corrective term based on this posterior prediction: let Vu(x)superscriptsubscript𝑉𝑢𝑥V_{u}^{*}(x)italic_V start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∗ end_POSTSUPERSCRIPT ( italic_x ) be the adjusted prediction of variance for output u𝑢uitalic_u at point x𝑥xitalic_x. Then

Var[Du]Var[Du]+diag(1n1Vu(x1),,1nmVu(xm)),Vardelimited-[]subscript𝐷𝑢Vardelimited-[]subscript𝐷𝑢diag1subscript𝑛1superscriptsubscript𝑉𝑢subscript𝑥11subscript𝑛𝑚superscriptsubscript𝑉𝑢subscript𝑥𝑚\text{Var}[D_{u}]\to\text{Var}[D_{u}]+\text{diag}\left(\frac{1}{n_{1}}V_{u}^{*%}(x_{1}),\dots,\frac{1}{n_{m}}V_{u}^{*}(x_{m})\right),Var [ italic_D start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT ] → Var [ italic_D start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT ] + diag ( divide start_ARG 1 end_ARG start_ARG italic_n start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_ARG italic_V start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∗ end_POSTSUPERSCRIPT ( italic_x start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ) , … , divide start_ARG 1 end_ARG start_ARG italic_n start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_ARG italic_V start_POSTSUBSCRIPT italic_u end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∗ end_POSTSUPERSCRIPT ( italic_x start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT ) ) ,

where diag()diag\text{diag}(\cdot)diag ( ⋅ ) represents the diagonal matrix whose diagonal entries are the argument \cdot. The process of updating the mean predictions is unchanged save for this adjustment to the data variance matrix, and we may use the resulting emulators in the normal fashion.

There are two additional considerations when applying covariance emulation to a simulator. Firstly, we must determine second-order specifications for both the mean and variance emulator: this incorporates not just first- and second-order quantities, but also fourth-order quantities since for example Cab,R(C)(x)subscript𝐶𝑎𝑏𝑅𝐶𝑥C_{ab,R(C)}(x)italic_C start_POSTSUBSCRIPT italic_a italic_b , italic_R ( italic_C ) end_POSTSUBSCRIPT ( italic_x ) represents the variance of the residuals of a covariance. These higher-order quantities are less intuitive than those required to specify an emulator for the simulator output itself; while some can be determined in much the same way as for a mean emulator, given data, some cannot. A number of arguments about how to determine these quantities can be found in [31]. Secondly, our implausibility measure is naturally modified by this change due to its dependence on our variance: in the multivariate form presented in (7) we obtain an extra term arising from covariance matrix Var[k(fa(x))]Vardelimited-[]subscript𝑘subscript𝑓𝑎𝑥\text{Var}[\mathcal{R}_{k}(f_{a}(x))]Var [ caligraphic_R start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ( italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT ( italic_x ) ) ].

Finally, one may note that this is not a true multivariate emulation; we are instead individually emulating the elements of a covariance matrix. For some applications, this could result in predicted covariance matrices that are not semi-positive definite, and we must be careful in such situations555Indeed, even if we were to simply emulate variances and ignore the off-diagonal terms of the covariance matrix, we may still predict negative variance at certain points in parameter space.. One could treat such ill-posed predictions as diagnostic warnings on those parts of parameter space where they occur, effectively refusing to make statements about those parts of parameter space in the given wave; employ techniques to recast an ill-posed matrix as a valid covariance matrix, for example, performing an eigendecomposition and replacing any negative eigenvalues with 00; or in some cases a transformation to log-covariance can be appropriate666Though one must be careful about the connection between the uncertainty in the prediction of the mean of the simulator output and the prediction of the mean of the log-variance.. For systems with large numbers of outputs, this may also result in a prohibitively large number of emulators, since a system with n𝑛nitalic_n outputs would result in the construction of n+12n(n+1)𝑛12𝑛𝑛1n+\frac{1}{2}n(n+1)italic_n + divide start_ARG 1 end_ARG start_ARG 2 end_ARG italic_n ( italic_n + 1 ) emulators. Depending on the structure of the model in question, one might choose to focus on a subset of output covariances: for example, if we have good reason to believe that two outputs are (close to) uncorrelated, then no additional information would be gained from constructing a covariance emulator for that combination.

Intricacies notwithstanding, the framework presented above allows us to construct a natural hierarchy capable of encapsulating the output response and stochasticity over the parameter space, feeding informed statements about heteroskedasticity in the space of interest into posterior predictions of uncertainty on the outputs of interest.

Appendix B The History Matching Framework

The emulators defined in Supplementary Material A have the benefit of being extremely fast to evaluate in comparison with the simulator they represent. The drawback of such an approach is that, by definition, the emulators are statistical approximations of the simulator output; while they naturally provide a quantification of the uncertainty about any prediction made, this remains a source of uncertainty. However, this is by no means the only source of uncertainty in our calibration problem; a common aim when using simulators is to answer the following question, which implicitly describes those external sources of uncertainty and which we will further unpack shortly.

Given observed data corresponding to a simulator output, what combinations of input parameters could give rise to simulator output consistent with this observation?

To answer this question, we apply the history matching approach [17], which aims to find the space of acceptable matches via complementarity. For further discussion of this choice and a comparison with other approaches, see [46, 14, 26].

We must first create a link between an observation of a real-world process and our emulator. Let us denote one aspect of the real-world process, represented by simulator output fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ), by yisubscript𝑦𝑖y_{i}italic_y start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT. Observations of yisubscript𝑦𝑖y_{i}italic_y start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT are seldom made perfectly; for example, in epidemiological models it is common to expect that case numbers for a disease are subject to under-reporting or over-dispersion. We therefore link the observation zisubscript𝑧𝑖z_{i}italic_z start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT of the process to the process itself via

zi=yi+ei,subscript𝑧𝑖subscript𝑦𝑖subscript𝑒𝑖z_{i}=y_{i}+e_{i},italic_z start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = italic_y start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT + italic_e start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ,

with eisubscript𝑒𝑖e_{i}italic_e start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT some random quantity representing the observational error structure. Similarly, we would not expect the simulator to be a perfect representation of the physical reality it models: we link the two via

yi=fi(x)+ϵi(x),subscript𝑦𝑖subscript𝑓𝑖𝑥subscriptitalic-ϵ𝑖𝑥y_{i}=f_{i}(x)+\epsilon_{i}(x),italic_y start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) + italic_ϵ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ,

where ϵi(x)subscriptitalic-ϵ𝑖𝑥\epsilon_{i}(x)italic_ϵ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) represents the (in general, x𝑥xitalic_x-dependent) ‘model discrepancy’ for output i𝑖iitalic_i at a parameter set x𝑥xitalic_x which seeks to describe the deficiency between the simulator and reality [30, 47, 46]. We also have a natural means by which to link the simulator output to the emulators as described in the preceding section; we therefore have a chain of statements that connect the observational data available to us to the emulator predictions.

This uncertainty structure allows us to approach the problem of finding acceptable parameter sets in a markedly different way to methods such as optimisation. Rather than seeking points in parameter space whose simulated outputs are likely to be good matches to the observational data (via, for example, maximising a goodness-of-fit measure), we instead focus on removing parts of parameter space that are highly unlikely to give rise to a good match. By ‘highly unlikely’ here we mean that even accounting for all uncertainties in the system (emulator, observational, and model uncertainty), our prediction has a negligible probability of giving rise to our observation. We can therefore systematically and iteratively remove the unsuitable parts of parameter space, improve the emulators’ predictions over the remaining space, and repeat the process until we arrive at the full parameter space of interest.

Concretely, we now define an implausibility measure [18] for observations z=(z1,,zm)𝑧subscript𝑧1subscript𝑧𝑚z=(z_{1},\dots,z_{m})italic_z = ( italic_z start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , … , italic_z start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT ) corresponding to outputs f1(x),,fm(x)subscript𝑓1𝑥subscript𝑓𝑚𝑥f_{1}(x),\dots,f_{m}(x)italic_f start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ( italic_x ) , … , italic_f start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT ( italic_x ). The implausibility for a single output fi(x)subscript𝑓𝑖𝑥f_{i}(x)italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) can be written as

Ii2(x)=(𝔼D[fi(x)]zi)2VarD[fi(x)]+Var[ei]+Var[ϵi(x)]+𝔼S[gv(x)],subscriptsuperscript𝐼2𝑖𝑥superscriptsubscript𝔼𝐷delimited-[]subscript𝑓𝑖𝑥subscript𝑧𝑖2subscriptVar𝐷delimited-[]subscript𝑓𝑖𝑥Vardelimited-[]subscript𝑒𝑖Vardelimited-[]subscriptitalic-ϵ𝑖𝑥subscript𝔼𝑆delimited-[]subscript𝑔𝑣𝑥I^{2}_{i}(x)=\frac{(\mathbb{E}_{D}[f_{i}(x)]-z_{i})^{2}}{\text{Var}_{D}[f_{i}(%x)]+\text{Var}[e_{i}]+\text{Var}[\epsilon_{i}(x)]+\mathbb{E}_{S}[g_{v}(x)]},italic_I start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) = divide start_ARG ( blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] - italic_z start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ) start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT end_ARG start_ARG Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] + Var [ italic_e start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] + Var [ italic_ϵ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] + blackboard_E start_POSTSUBSCRIPT italic_S end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT ( italic_x ) ] end_ARG ,(5)

where 𝔼D[fi(x)]subscript𝔼𝐷delimited-[]subscript𝑓𝑖𝑥\mathbb{E}_{D}[f_{i}(x)]blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] and VarD[fi(x)]subscriptVar𝐷delimited-[]subscript𝑓𝑖𝑥\text{Var}_{D}[f_{i}(x)]Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] are the prediction and variance for the mean response, and 𝔼S[gv(x)]subscript𝔼𝑆delimited-[]subscript𝑔𝑣𝑥\mathbb{E}_{S}[g_{v}(x)]blackboard_E start_POSTSUBSCRIPT italic_S end_POSTSUBSCRIPT [ italic_g start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT ( italic_x ) ] symbolically represents the prediction of the stochasticity, relative to sample data S𝑆Sitalic_S. If Ii(x)subscript𝐼𝑖𝑥I_{i}(x)italic_I start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) is large, then we may think it unlikely that we would obtain an acceptable match to observed data were we to run the simulator at the point x𝑥xitalic_x; such a point is termed implausible. Conversely, if I(x)𝐼𝑥I(x)italic_I ( italic_x ) is small then we are unable to rule out the possibility that x𝑥xitalic_x would give rise to a good match to observational data, given the information available to us; x𝑥xitalic_x is correspondingly deemed non-implausible. Note that the implausibility I(x)𝐼𝑥I(x)italic_I ( italic_x ) can be small for two reasons: either the prediction 𝔼D[fi(x)]subscript𝔼𝐷delimited-[]subscript𝑓𝑖𝑥\mathbb{E}_{D}[f_{i}(x)]blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] is close to the observation z𝑧zitalic_z, suggesting a good match to data; or else the uncertainties (particularly VarD[fi(x)]subscriptVar𝐷delimited-[]subscript𝑓𝑖𝑥\text{Var}_{D}[f_{i}(x)]Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ]) are large, suggesting that we currently do not have enough information to rule out a good match to data and that further investigation at this point would be useful. The implausibility measure therefore acts as a measure of both exploration and exploitation, indicating regions of parameter space whose consideration would improve our understanding of the simulator. A point is deemed implausible only if an simulator evaluation at this point has low probability of matching to observational data and inspection of the neighbourhood of the point would provide no additional insight.

In the presence of multiple outputs, a natural approach is to demand that a point be non-implausible with respect to all observations: this gives rise to a maximum implausibility measure

IM(x)=maxi{Ii(x)}i=1,,mI_{M}(x)=\max_{i}\{I_{i}(x)\}_{i=1,\dots,m}italic_I start_POSTSUBSCRIPT italic_M end_POSTSUBSCRIPT ( italic_x ) = roman_max start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT { italic_I start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) } start_POSTSUBSCRIPT italic_i = 1 , … , italic_m end_POSTSUBSCRIPT(6)

from which other, less restrictive, measures such as second-maximum I2M(x)subscript𝐼2𝑀𝑥I_{2M}(x)italic_I start_POSTSUBSCRIPT 2 italic_M end_POSTSUBSCRIPT ( italic_x ) and n𝑛nitalic_nth maximum implausibilities follow. We may combine these measures at will; for example, allowing at most one output to be slightly further from the data, but all to be within a certain closeness: we would therefore wish to satisfy a constraint IM(x)c1subscript𝐼𝑀𝑥subscript𝑐1I_{M}(x)\leq c_{1}italic_I start_POSTSUBSCRIPT italic_M end_POSTSUBSCRIPT ( italic_x ) ≤ italic_c start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT on the maximum implausibility and an additional constraint I2M(x)c2subscript𝐼2𝑀𝑥subscript𝑐2I_{2M}(x)\leq c_{2}italic_I start_POSTSUBSCRIPT 2 italic_M end_POSTSUBSCRIPT ( italic_x ) ≤ italic_c start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT on the second maximum, with c1>c2subscript𝑐1subscript𝑐2c_{1}>c_{2}italic_c start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT > italic_c start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT, for example.

We may also combine the multiple emulator implausibilities more directly, via a multivariate measure:

I2(x)=(𝔼D[f(x)]z)(VarD[f(x)]+Var[ei]+Var[ϵi(x)]+𝔼S[fv(x)])1(𝔼D[f(x)]z)superscript𝐼2𝑥superscriptsubscript𝔼𝐷delimited-[]𝑓𝑥𝑧topsuperscriptsubscriptVar𝐷delimited-[]𝑓𝑥Vardelimited-[]subscript𝑒𝑖Vardelimited-[]subscriptitalic-ϵ𝑖𝑥subscript𝔼𝑆delimited-[]subscript𝑓𝑣𝑥1subscript𝔼𝐷delimited-[]𝑓𝑥𝑧I^{2}(x)=(\mathbb{E}_{D}[f(x)]-z)^{\top}\left(\text{Var}_{D}[f(x)]+\text{Var}[%e_{i}]+\text{Var}[\epsilon_{i}(x)]+\mathbb{E}_{S}[f_{v}(x)]\right)^{-1}(%\mathbb{E}_{D}[f(x)]-z)\\italic_I start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ( italic_x ) = ( blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ] - italic_z ) start_POSTSUPERSCRIPT ⊤ end_POSTSUPERSCRIPT ( Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ] + Var [ italic_e start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ] + Var [ italic_ϵ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ] + blackboard_E start_POSTSUBSCRIPT italic_S end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT ( italic_x ) ] ) start_POSTSUPERSCRIPT - 1 end_POSTSUPERSCRIPT ( blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ] - italic_z )(7)

Here, the quantity 𝔼D[f(x)]subscript𝔼𝐷delimited-[]𝑓𝑥\mathbb{E}_{D}[f(x)]blackboard_E start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ] is taken to be the vector of predictions formed from the m𝑚mitalic_m trained emulators of the outputs at a point x𝑥xitalic_x, VarD[f(x)]subscriptVar𝐷delimited-[]𝑓𝑥\text{Var}_{D}[f(x)]Var start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT [ italic_f ( italic_x ) ] the corresponding covariance matrix, and 𝔼S[fv(x)]subscript𝔼𝑆delimited-[]subscript𝑓𝑣𝑥\mathbb{E}_{S}[f_{v}(x)]blackboard_E start_POSTSUBSCRIPT italic_S end_POSTSUBSCRIPT [ italic_f start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT ( italic_x ) ] the covariance matrix of the stochasticity, as predicted from the (trained) covariance emulators. This can be extremely useful when we have some understanding of the correlations between outputs, since this measure will constrain outputs according to their relations as well as their individual predictions.

The history matching approach proceeds in iterations. We refer to these iterations as ‘waves’: at each wave k𝑘kitalic_k, a set of emulators are constructed for a collection of outputs Oksubscript𝑂𝑘O_{k}italic_O start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT based on a representative sample of points and their simulator evaluations from wave k1𝑘1k-1italic_k - 1. These trained emulators are used to assess implausibility over the space, 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT, that remained at wave k1𝑘1k-1italic_k - 1, discarding those regions now deemed implausible to produce a representative sample of a smaller parameter space 𝒳k+1subscript𝒳𝑘1\mathcal{X}_{k+1}caligraphic_X start_POSTSUBSCRIPT italic_k + 1 end_POSTSUBSCRIPT. These points, and their simulator evaluations, are used to inform emulators at wave k+1𝑘1k+1italic_k + 1 and so on. The algorithm is shown in schematic form in Algorithm1.

Initial parameter region 𝒳0subscript𝒳0\mathcal{X}_{0}caligraphic_X start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT

k1𝑘1k\leftarrow 1italic_k ← 1

while𝒳k1subscript𝒳𝑘1\mathcal{X}_{k-1}\neq\emptysetcaligraphic_X start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT ≠ ∅ and no other stopping condition is satisfieddo

Generate an appropriate design {xi}i=1,,nsubscriptsubscript𝑥𝑖𝑖1𝑛\{x_{i}\}_{i=1,\dots,n}{ italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT } start_POSTSUBSCRIPT italic_i = 1 , … , italic_n end_POSTSUBSCRIPT of n𝑛nitalic_n points over 𝒳k1subscript𝒳𝑘1\mathcal{X}_{k-1}caligraphic_X start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT

Identify a collection of informative outputs, Oksubscript𝑂𝑘O_{k}italic_O start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT

Obtain simulator evaluations for outputs Oksubscript𝑂𝑘O_{k}italic_O start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT for each point xisubscript𝑥𝑖x_{i}italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT

Construct emulators defined only over 𝒳k1subscript𝒳𝑘1\mathcal{X}_{k-1}caligraphic_X start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT for the collection Oksubscript𝑂𝑘O_{k}italic_O start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT, trained using the simulator evaluations

Calculate implausibility across the entirety of 𝒳k1subscript𝒳𝑘1\mathcal{X}_{k-1}caligraphic_X start_POSTSUBSCRIPT italic_k - 1 end_POSTSUBSCRIPT, identifying the new non-implausible region 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT

kk+1𝑘𝑘1k\leftarrow k+1italic_k ← italic_k + 1

endwhile

If 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}\neq\emptysetcaligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT ≠ ∅, generate a large number of acceptable runs from 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT, sampled appropriately.

The stopping condition mentioned in Algorithm1 depends on the ultimate aim of our model analysis; a point mentioned in the context of HPVsim in Section4.2.2. We pause here to highlight one potential stopping condition which is unique to the history matching framework: the possibility of stopping because the current non-implausible space is empty, 𝒳k=subscript𝒳𝑘\mathcal{X}_{k}=\emptysetcaligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT = ∅ for some k𝑘kitalic_k. Unlike in methods such as optimisation and ABC, which return the ‘best’ available point or posterior distribution, history matching is not similarly constrained and we may rule out the entire space as implausible. Such a situation can be extremely informative, representing a fundamental divergence between our observations, model, and reality, and force us to consider our error structure and model behaviour relative to the problem we wish to address.

The question of how to sample from a geometrically complex non-implausible region is an area of research in its own right; here we have followed the procedure detailed in [33] which we summarise here.

  1. 1.

    Generate a large Latin hypercube and reject those points whose implausibility exceeds the cutoff;

  2. 2.

    Use the points from Step 1 to select pairs of non-implausible points and draw rays connecting them. Add points that lay on these rays and on the boundary of the non-implausible region to the collection of points already obtained;

  3. 3.

    Use the non-implausible set as the basis for a mixture distribution of uniform ellipsoids, performing importance sampling using this distribution;

  4. 4.

    Thin the final collection of non-implausible points using a maximin argument, to produce a set of space-filling parameter combinations.

This ensures, for most applications, that we obtain a collection of parameter sets that span the non-implausible region, and it proved to be adequate for point proposal in this application.

Just as we may choose outputs of interest, Oksubscript𝑂𝑘O_{k}italic_O start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT at each wave depending on the stability of the parameter space 𝒳ksubscript𝒳𝑘\mathcal{X}_{k}caligraphic_X start_POSTSUBSCRIPT italic_k end_POSTSUBSCRIPT, so too may we modify our methodology for emulation. At early waves, we may deem the output covariance structure to be too volatile to provide any meaningful insight due to extreme behaviour in parts of parameter space we have yet to rule out; nothing precludes us from simply focusing on the emulation of means (with a suitably conservative estimate of stochasticity included as model discrepancy) until we have reduced the space to a more stable region of interest.

There is one final consideration that must be made in light of our modelling problem, which we briefly mention here. In the formulation for stochastic models and emulation thereof, we implicitly assume that we may match to the ‘true’ means, (fi(x))subscript𝑓𝑖𝑥\mathcal{M}(f_{i}(x))caligraphic_M ( italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) ), of the system given sample data f¯i(x)subscript¯𝑓𝑖𝑥\bar{f}_{i}(x)over¯ start_ARG italic_f end_ARG start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( italic_x ) from a finite collection of realisations. The additional term in the denominator of the implausibility (5) and inverse matrix of (7) accounts for the potential disconnect between the two quantities; this is only relevant if we believe that the true mean is of interest. As touched upon earlier, one may instead consider matching to individual realisations: for this we would instead wish to more carefully consider the covariance structure between outputs to adequately constrain the non-implausible space. We reserve this avenue of exploration for further work.

Appendix C HPVsim Parameters and Outputs

C.1 Input Parameters

33333333 parameters were chosen to be varied, and the initial allowed ranges determined via expert consideration of the physical system. Where a parameter name is of the form name_xx, the xx corresponds to one of the four genotypes under consideration: 16161616, 18181818, hi5hi5\text{hi}5hi 5, or ohr.

NameDescriptionRange
betaTransmission probability[0.02,0.25]0.020.25[0.02,0.25][ 0.02 , 0.25 ]
de_xxMean (in years) of lognormal distribution of duration of infection prior to clearance, control, or transformation for genotype xx[3,10]310[3,10][ 3 , 10 ]
de_p2_xxVariance of lognormal distribution for episomal infection for genotype xx[5,15]515[5,15][ 5 , 15 ]
sr_xxGrowth rate parameter mapping episomal duration to severity for genotype xx[0.1,0.5]0.10.5[0.1,0.5][ 0.1 , 0.5 ]
dp_xxMean (in years) of folded normal for precin duration of HPV for genotype xx[0.25,4]0.254[0.25,4][ 0.25 , 4 ]
tp_xxPer-cell annual probability of transformation for genotype xx[1011,108]superscript1011superscript108[10^{-11},10^{-8}][ 10 start_POSTSUPERSCRIPT - 11 end_POSTSUPERSCRIPT , 10 start_POSTSUPERSCRIPT - 8 end_POSTSUPERSCRIPT ]
rb_xxRelative genotype transmissibility to the baseline figure for genotype xx[0.7,1.3]0.71.3[0.7,1.3][ 0.7 , 1.3 ]
debut_femaleMean age of female sexual debut[12,19]1219[12,19][ 12 , 19 ]
controlProbability that hpv infection will be cleared by the host[0,1]01[0,1][ 0 , 1 ]
reactiveProbability of reactivation of latent HPV infection[0,0.15]00.15[0,0.15][ 0 , 0.15 ]
sev_distMean of folded normal for individual level severity scale factors[0.1,2]0.12[0.1,2][ 0.1 , 2 ]
pmMean number of concurrent marital partners[0.001,0.15]0.0010.15[0.001,0.15][ 0.001 , 0.15 ]
pcMean number of concurrent casual partners[0.1,0.6]0.10.6[0.1,0.6][ 0.1 , 0.6 ]
dmMean duration of marital partnership[8,25]825[8,25][ 8 , 25 ]
dcMean duration of cancer[6,15]615[6,15][ 6 , 15 ]

C.2 Outputs and Uncertainties

22222222 outputs were identified: of these, 14141414 corresponded to age-stratified incidence of cancer in the year 2020202020202020, and the remaining 8888 were proportions of the different HPV genotypes presenting in patients with cervical cancer and in those with high-grade lesions (cin3), measured in 2015201520152015. Observation error is encoded via the intervals of the output values. Here, model discrepancy is entered as Var[ϵ]Vardelimited-[]italic-ϵ\sqrt{\text{Var}[\epsilon]}square-root start_ARG Var [ italic_ϵ ] end_ARG for each output.

Output nameTargetModel Disc.
cancer202015.0[1,38]138[1,38][ 1 , 38 ]0.8120.8120.8120.812
cancer202020.0[183,401]183401[183,401][ 183 , 401 ]9.8479.8479.8479.847
cancer202025.0[713,1032]7131032[713,1032][ 713 , 1032 ]28.15128.15128.15128.151
cancer202030.0[1023,1481]10231481[1023,1481][ 1023 , 1481 ]40.54640.54640.54640.546
cancer202035.0[1275,1846]12751846[1275,1846][ 1275 , 1846 ]50.30950.30950.30950.309
cancer202040.0[1372,1986]13721986[1372,1986][ 1372 , 1986 ]54.39354.39354.39354.393
cancer202045.0[1263,1829]12631829[1263,1829][ 1263 , 1829 ]50.45050.45050.45050.450
cancer202050.0[1090,1579]10901579[1090,1579][ 1090 , 1579 ]44.37644.37644.37644.376
cancer202055.0[934,1352]9341352[934,1352][ 934 , 1352 ]38.38838.38838.38838.388
cancer202060.0[674,1234]6741234[674,1234][ 674 , 1234 ]32.98232.98232.98232.982
cancer202065.0[443,1002]4431002[443,1002][ 443 , 1002 ]25.46125.46125.46125.461
cancer202070.0[278,761]278761[278,761][ 278 , 761 ]18.67918.67918.67918.679
cancer202075.0[155,505]155505[155,505][ 155 , 505 ]12.08512.08512.08512.085
cancer202080.0[62,239]62239[62,239][ 62 , 239 ]5.6595.6595.6595.659
cancer_16[0.45,0.611]0.450.611[0.45,0.611][ 0.45 , 0.611 ]0.0240.0240.0240.024
cancer_18[0.098,0.214]0.0980.214[0.098,0.214][ 0.098 , 0.214 ]0.0110.0110.0110.011
cancer_hi5[0.104,0.222]0.1040.222[0.104,0.222][ 0.104 , 0.222 ]0.0180.0180.0180.018
cancer_ohr[0.115,0.238]0.1150.238[0.115,0.238][ 0.115 , 0.238 ]0.0110.0110.0110.011
cin3_16[0.173,0.425]0.1730.425[0.173,0.425][ 0.173 , 0.425 ]0.0110.0110.0110.011
cin3_18[0.034,0.203]0.0340.203[0.034,0.203][ 0.034 , 0.203 ]0.0110.0110.0110.011
cin3_hi5[0.227,0.425]0.2270.425[0.227,0.425][ 0.227 , 0.425 ]0.0130.0130.0130.013
cin3_ohr[0.173,0.425]0.1730.425[0.173,0.425][ 0.173 , 0.425 ]0.0140.0140.0140.014
Investigating Complex HPV Dynamics Using Emulation and History Matching (2024)
Top Articles
Latest Posts
Article information

Author: Margart Wisoky

Last Updated:

Views: 6591

Rating: 4.8 / 5 (58 voted)

Reviews: 89% of readers found this page helpful

Author information

Name: Margart Wisoky

Birthday: 1993-05-13

Address: 2113 Abernathy Knoll, New Tamerafurt, CT 66893-2169

Phone: +25815234346805

Job: Central Developer

Hobby: Machining, Pottery, Rafting, Cosplaying, Jogging, Taekwondo, Scouting

Introduction: My name is Margart Wisoky, I am a gorgeous, shiny, successful, beautiful, adventurous, excited, pleasant person who loves writing and wants to share my knowledge and understanding with you.