Abstract
Mathematical models are getting increasingly detailed to better predict phenomena or gain more accurate insights into the dynamics of a system of interest, even when there are no validation or training data available. Here, we show through ANOVA and statistical theory that this practice promotes fuzzier estimates because it generally increases the model’s effective dimensions, i.e., the number of influential parameters and the weight of high-order interactions. By tracking the evolution of the effective dimensions and the output uncertainty at each model upgrade stage, modelers can better ponder whether the addition of detail truly matches the model’s purpose and the quality of the data fed into it.
Without validation data, more complex models may succumb to uncertainty.
INTRODUCTION
Many mathematical models are getting increasingly complex with the assumption that ever-comprehensive representations of the process under study will eventually bound uncertainties and produce more accurate insights (1–3). This trend has been fueled by computational advances that have permitted to increase the number of transistors in a chip and thus the speed of arithmetic operations without substantially increasing power requirements—the combination of what is known as Moore’s law (4) and Dennard’s scaling (5). Such breakthroughs have allowed the simple general circulation models of the 1960s to evolve into comprehensive atmosphere-ocean general circulation models (6, 7), or the “bucket”-type models of the 1970s to turn into global hydrological models that simulate water uses and the human impact on the global water cycle (8, 9). In mathematical epidemiology, the basic compartmental models of Kermack and McKendrick (10) have expanded into the Imperial College’s transmission model of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), based on more than 900 parameters (11).
In the environmental/climate sciences, hydrology, and epidemiology domains, the development of finer-grained models often proceeds without having, at the scale required, specific data available to train or validate the model (12). Models tend to be based on physical laws or principles specific to the field and hence may project their estimations, predictions, or dynamics into the unknown. When the lack of validation data renders the assessment of model bias (i.e., how far off are the modeled estimates from the data available) unfeasible, modelers cannot benefit from existing statistical instruments that help balance model complexity with error so as to align with science’s quest for parsimony, such as Akaike’s (13) or Schwarz’s (14) information criterion. In these cases, model expansion may proceed without standardized approaches to assess whether the level of model detail matches the quality of the knowledge available, whether the newly added processes truly lead to more accurate estimates, or how the output uncertainty builds up at each model upgrade stage. This often limits the social usefulness of process-based models as instruments to inform policies in the real world (15), where reliance on excessively complex and overconfident models may have deleterious social-environmental consequences (16).
Here, we show that modelers can gauge the connection between model complexity and uncertainty at all stages of model development by calculating the model’s “effective dimensions,” that is, the number of influential parameters and active higher-order effects (17–19). The concept of effective dimensions helps to better tune the level of model detail to the context and purpose of the application, ultimately improving the quality of models that do not (or cannot) fit any training or validation data. We also provide numerical evidence that the addition of model detail in process-based models tends to produce more (and not less) uncertain estimates because it increases the model’s effective dimensions, which generally boost the output variance. This fact, which may have gone unnoticed as yet because of the scarce uptake of uncertainty and sensitivity analyses in mathematical modeling (20), suggests that the quest toward ever-detailed mathematical models as a means to get sharper estimates or insights should be reassessed.
We first describe the mathematical foundations of our approach and the statistical theory behind the connection between model uncertainty, complexity, and the notion of effective dimensions. After illustrating this relation with a numerical simulation, we stress-test our theory with a meta-model that generates a very broad range of functional forms and faithfully reproduces the effect of complexification at higher effective dimensions. We then illustrate how the concept of effective dimensions can help modelers balance model complexity with uncertainty using several increasingly complex models of the energy, agriculture, and epidemiology domain. We conclude by discussing the implications of our results for the design of mathematical models and model-based policies.
RESULTS
Model complexity and uncertainty
We regard the number of parameters and the pattern of their connections as key contributors to the complexity of mathematical models (21, 22). Our approach thus relies on the notion of “aggregate complexity” (23) and focuses on system components and their interactions when no validation data are available. This can be the case of predicting the evolution of a new epidemic, exploring the impact of a new technology, predicting the potential scarcity of a natural resource, and so on, over a myriad of settings often encountered in impact assessment studies.
By setting the focus on the number of model parts and their connections, we can relate model complexity with uncertainty through statistical theory via the analysis of variance (ANOVA) decomposition framework and the notion of effective dimensions (17–19) (see Materials and Methods). In this framework, parameters are regarded as stochastic variables whose uncertainty is described by probability distributions reflecting their measurement error, natural variation, inherent randomness, or the subjective judgement of experts (24). Given a model of the form y = f(x), x = (x1, x2, …, xi, …, xk) ∈ ℝk, where y is a scalar output and x1, …, xk are k independent parameters, we can calculate the proportion of variance conveyed to y by each parameter (first-order effect, Si) and by the interaction between pairs of parameters (second-order effect, Si,j), triplets of parameters (third-order effect, Si,j,l), etc., up to the kth-order interaction. For a model with just three parameters, this is S1 + S2 + S3 + S1,2 + S1,3 + S2,3 + S1,2,3 = 1. This variance decomposition applies when f(x) is square-integrable over the dominion of existence, and is linked to Sobol’s (25) functional decomposition scheme, where f(x) is decomposed as the sum of 1,2, …, k-dimensional functions.
The estimation of up to the kth-order interaction may be impractical for high-dimensional or computationally demanding models. Under these circumstances, the calculation of the total-order effect (Ti) allows one to capture the proportion of variance conveyed to y by the first-order effect of xi jointly with its interactions up the kth order (26). For parameter x1 in a three-parameter model, this would be and similarly for x2 and x3. We can now introduce the concept of effective dimensions (17–19).
Effective dimension in the “superposition” sense
Let λ = {1,2, …, k}. For any subset u ⊆ λ, let ∣u∣ denote its cardinality (18). In the “superposition sense,” the effective dimension of a model f is the smallest integer ks such that
(1) |
where 0 < p < 1. The value of p is arbitrarily set, and here, we use p = 0.99 following Caflisch et al. (17). Using again a three-parameter model as an example, if (S1 + S2 + S3 + S1,2 + S2,3 + S1,3) ≥ p, then ∣u∣= ks = 2. The effective dimension ks is therefore the order of the highest effect that needs to be included in Eq. 1 to reach p. Models with a high effective dimension in the superposition sense have important high-order effects and hence can be considered as a sum of ks-dimensional functions (see Materials and Methods and Eq. 4) (19). Since ks reflects the weight of interactions between the model’s parameters, ks is also a proxy for the pattern of connections in the model. For computation simplicity, here we calculate ks up to the third-order indices and assume that ks ≥ 4 when (∑ Si + ∑ Si, j + ∑ Si,j,l) < p.
Effective dimension in the “truncation” sense
Let us now consider the vector of total-order indices T = {T1, T2, …, Tk}. In the truncation sense, we define the effective dimension of a model f to be the integer kt, such that
(2) |
where ∣C∣ is the cardinality of the subset C formed by the number of elements Ti in T such that Ti > q. Here, we define q = 0.05 as this is the threshold commonly used in sensitivity analysis to screen “influential” from “noninfluential” parameters (those that convey uncertainty to y from those that do not). Models with a high effective dimension in the truncation sense have a large number of influential parameters and thus a large kt value. Note that our definition of effective dimension in the truncation sense is a simplification of the notion introduced by Wang and Fang (18) and Caflisch et al. (17). Here, we give precise kt values for all the models examined.
Usually, k ≫ kt ≫ ks due to the general dominance of low-order effects in mathematical models and the preeminence of the Pareto principle (c. 80% of the effects are conveyed by c. 20% of the parameters) (27–29). Models live in the space set by kt and ks and not in that nominally defined by k, which may be artificially large if the model includes a non-negligible number of noninfluential parameters. The space defined by kt and ks cannot be simplified without modifying the model’s behavior, and thus, it is irreducibly complex (30). It follows that more complex models will generally display a higher effective dimension in kt and ks, an increase that promotes a larger output uncertainty (Fig. 1A). This is because the output variance of more complex models is increasingly driven by higher-order effects activated by the progressive addition of influential model parameters: Note in Fig. 1A how the larger the ks dimension is, the smaller the sum of first-order indices Si.
Fig. 1. The link between model complexity and uncertainty.
(A) Trend between the effective dimension in the truncation sense (kt), the effective dimension in the superposition sense (ks), the sum of the first-order effects of the models’ parameters Si, and the coefficient of variation (CV), which we use as a proxy for uncertainty. Each dot is a simulation (N = 210) that randomizes the coefficients of Sobol’s G function (68) (see Materials and Methods). The small rectangle frames some simulations in which models with different kt and ks dimensions have an almost identical output variance. (B) The meta-model approach. Evolution of uncertainty through different dimensions k as a function of the order of the highest interaction activated, for n ≤ k. The line shows the median value, whereas the ribbon displays the SD. N = 211. (C) Evolution of uncertainty after fixing the model dimensionality at k = 15 and activating up to the nth-order interaction.
The meta-model check
To check whether the relation between higher effective dimensions, the addition of model detail, and larger output uncertainty holds regardless of the model functional form, we stress-test our approach with a meta-model based on the Becker (29) metafunction. Our meta-model randomizes over 13 different univariate equations representing common responses observed in physical systems, from linear to multimodal (fig. S1) (31). It also randomly activates up to the nth-order effect, for n ≤ k, to observe how different combinations of k and ks values affect the output uncertainty. This approach allows the creation of more than two million models with different dimensionalities, functional forms, and degree of interactions (see Materials and Methods).
Also in this case, the output uncertainty tends to increase with the number of model parameters and the activation of up to the kth-order interaction (the blue trend in Fig. 1B). Note how a model with more parameters does not necessarily lead to a larger uncertainty if n ≠ k (the orange trend in Fig. 1B). This highlights the importance of high-order interactions in defining both the complexity of the model and its output uncertainty. The number of parameters can be fixed and the uncertainty increases all the same just by progressively adding new interactions between the parameters (Fig. 1C).
The width of the trend in Fig. 1 indicates that the addition of new parameters does not mechanically have to promote more uncertain estimates: Several models with different effective dimensions in both kt and ks may display the same output variance. By routinely assessing the model’s effective dimensions in every model upgrade stage, modelers can appraise the link between complexity and uncertainty during model development and ponder whether the addition of detail truly suits the model’s purpose. Optimal complexity would depend on whether the model aims at predicting variables, understanding dynamics, managing resources, or guiding policy-making. In the next section, we illustrate the practical implications of our approach using several increasingly complex models as examples.
The link between effective dimensions and uncertainty in action
Compartmental models
In many fields, model complexification involves linking models in a causal chain, where the output of a given link is used as the input of the next one. In climate change research, greenhouse gas emission scenarios are usually transferred into impact models whose output is used to assess local impacts and, lastly, to design adaptation responses (32, 33). In flood management, flood wave forecasts require runoff predictions, which previously demand rainfall forecasts (34). Basic epidemiological models also follow this layout, as individuals flow from one compartment to another depending on whether they are susceptible, infected by the virus, or recovered (35). Eventually, new compartments are added to improve the model’s descriptive capacity and its usefulness for policy-making.
We illustrate how adding complexity affects the effective dimensions of compartmental models with the Probabilistic System Assessment Code Intercomparison (PSACOIN) Level 0 model (36). It simulates the leaking of radionuclides from a given repository and its transport to a buffer, then to the geosphere, and lastly to the biosphere, where they are ingested by humans and animals after drinking water from a contaminated well or stream (figs. S2A and S3). Overall, the PSACOIN Level 0 model includes 10 uncertain parameters described with probability distributions and 11 constants (table S1). If the model’s effective dimensions are assessed in every model upgrade stage, the progressive rise in kt and ks unfolds: kt goes from kt = 1 in the waste level to kt = 2, kt = 5, and kt = 6 in the buffer, geosphere, and biosphere level, respectively (Fig. 2A). As for ks, it goes from ks = 2 in the buffer level to ks ≥ 4 in the geosphere and biosphere level, since the sum of up to third-order effects in these levels is not enough to reach p (Fig. 2B). Note how uncertainties get larger in each model complexification stage, with the largest increase taking place in the transition between the buffer and the geosphere model: This is the critical leap in the development of the PSACOIN Level 0 model in terms of output uncertainty and complexity.
Fig. 2. Relation between uncertainty (represented by the CV) and complexity in the PSACOIN Level 0 model.
(A) Evolution of the CV and kt across the four compartments of the PSACOIN Level 0 model. (B) Proportion of the model output variance explained by the sum of first (Si)–, second (Si,j)–, and third (Si,j,l)–order effects. The horizontal dashed line is at p = 0.99. Note that the waste level is missing because it only contains one uncertain parameter and thus kt = ks = 1.
Flagging the progression of kt and ks values and related uncertainty in each PSACOIN upgrade stage allows one to see that the regulation on the safety of nuclear waste disposal should not set limits on doses to humans in the biosphere [as is usually done in the United States; see the case of the Yucca Mountain nuclear waste repository in Nevada (16, 37, 38)]. This is because once the contaminant reaches the buffer-geosphere interface, its fate is extremely hard to predict, and the addition of model detail just boosts the uncertainty. Regulations would thus establish more realistic and defensible safety standards if they define a maximum level of radioactivity leaving the buffer onto the aquifer and the geosphere (expressed in becquerels per year) rather than a total dose to the biosphere (expressed in sieverts per year). The PSACOIN Level 0 case illustrates how the concept of effective dimensions may help regulators distinguish in compartmental models which compartment produces the most solid inferences to guide policy-making, and where is the threshold beyond which the addition of complexity no longer makes the model fit for purpose.
Structural uncertainties
During the upgrade of a model, modelers may be unsure as to which is the best way to mathematically represent the newly added process. This structural uncertainty, also known as process-based uncertainty (39), often needs to be appraised to discern how the selection of a given equation over another conditions the model’s behavior and its output. The link between model complexity and uncertainty can be examined to unfold the uncertainty buildup due to the accumulation of both parametric and structural uncertainties, and to check whether model complexification should proceed without the need to address this source of structural uncertainties.
We illustrate this exercise with a basic formula to estimate the total amount of water withdrawn to irrigate a crop. Such a formula requires the calculation of the evapotranspiration of a reference crop (usually grass or alfalfa), the calculation of the crop evapotranspiration (wheat in this case), and lastly the estimation of the water withdrawn for irrigation (40). However, there are c. 50 different equations available to calculate the reference evapotranspiration and no agreement as to which method might work best (41). Some of the most used are the Priestley-Taylor and the FAO-56 Penman-Monteith (42), known to provide results that can differ by a factor of 2 or more (40). We examine the impact that this source of uncertainty has on the model’s effective dimensions and uncertainty with a trigger (43), a parameter that randomly decides which evapotranspiration equation should be used in each model simulation (see the Supplementary Materials).
The uncertainty in the selection of the reference evapotranspiration equation slightly increases the model’s effective dimensions in all compartments: kt values rise from 2 to 3 in the crop evapotranspiration compartment and from 5 to 6 in the irrigation water withdrawal compartment (Fig. 3A). The raise in ks is also moderate: ks values increase from 1, 1, and 2 to 2, 2, and 3 in the reference evapotranspiration, crop evapotranspiration, and water withdrawal compartment, respectively (Fig. 3B). Such minor increase in complexity, however, leads to a much larger output uncertainty than the uncertainty yielded by a model that uses either the Priestley-Taylor or the Penman-Monteith equations (blue lines in Fig. 3A).
Fig. 3. Relation between uncertainty (both parametric and structural) and complexity using the modeling of irrigation water withdrawals as a case study.
(A) Evolution of uncertainty and kt in the three compartments needed in the model: reference evapotranspiration, crop evapotranspiration, and water withdrawal. (B) Sum of first-, second-, and third-order effects in each compartment as a function of the reference evapotranspiration equation (ET0) used (Penman-Monteith, Priestley-Taylor, or uncertain).
The evolution of the model complexity and output uncertainty in these irrigation models suggests that (i) opting for either the Penman-Monteith or the Priestley-Taylor equation leads to overlook an important source of structural uncertainty whose effect in the model lasts until the very last compartment, (ii) the selection of just one reference evapotranspiration equation is hard to justify on the basis of parsimony since the effective dimensions of the resulting models are similar to those of the “uncertain” model, and (iii) regardless of the model selected, the largest boost in output uncertainty takes place in the transition to the irrigation water withdrawal compartment, which also displays comparatively higher ks values. Policy-makers using irrigation models to guide action in the field should plan with regard to an uncertainty that may be irreducible given the importance of higher-order interactions.
Complexification of compartmental models
The addition of model detail often goes along with the establishment of new compartment flows, links, and information delays. The upgraded model is then no longer a causal chain but a system with feedback loops and several input-output relationships. This is common practice in epidemiology, where the descriptive capacity of the classic susceptible-infected-recovered (SIR) model is often enhanced with the addition of fine-grained features such as seasonality or age stratification (44).
Such complexification may also produce fuzzier estimates at each model upgrade stage due to the increase in the model’s effective dimensions. To illustrate this fact, we propagate uncertainties in three increasingly detailed SIR-based models, extracted from Saad-Roy et al. (45, 46): a SIR-based model with waning immunity and no vaccination [SIR(S)], a SIR upgraded with a vaccination compartment [SIR(S-V)], and an extended SIR with different vaccination strategies [SIR(S-E)] (figs. S2, B to D, and S4). These SIR models aimed at providing insights into the magnitude of future SARS-CoV-2 cases in 2020-2021 given different assumptions on the nature of the adaptive immune response and pharmaceutical interventions.
Figure 4A indicates that the simplest SIR, the SIR(S), generally provides the least uncertain estimates, followed closely by the SIR(S-V). The more detailed model, the SIR(S-E), yields the most uncertain estimates, especially for IP and IS at t > 80. The progressive increase in uncertainty as a result of the addition of model detail is mirrored by a gradual raise in the effective dimensions: Notice how ks = 2 in all compartments of the SIR(S), ks = 3 in all compartments of the SIR(S-V), and ks ≥ 4 for the compartments SP, IP, R, and IS of the SIR(S-E), since at least fourth-order effects are needed to reach p (Fig. 4, B and C).
Fig. 4. Relation between complexity and uncertainty in three SIR(S)-based models.
Models from Saad-Roy et al. (45, 46). (A) Evolution of the uncertainty across time in the five compartments shared by all the models: fully susceptible individuals (SP), fully immune individuals as a result of recovery from either primary or secondary infection (IP), individuals that have recovered (R), individuals whose immunity has waned (SS), and individuals with secondary infection that transmit at reduced rate (IS). (B) Mean kt values of each of the compartments for each of the SIR(S) models considered. The error bar shows the SD. (C) Proportion of the model output uncertainty at t = 200 explained by the sum of first-, second-, and third-order effects. The horizontal dashed line is at p = 0.99.
The link between the output variance and the concept of effective dimensions permits to ponder which SIR model offers more appropriate insights to guide action. Note how the SIR(S) and the SIR(S-V) display an almost identical output variance through all the time intervals considered (Fig. 4A), and yet the SIR(S-V) has higher kt and ks values (Fig. 4, B and C). The propagation of uncertainty also evidences that the increase in the effective dimensions in the upgrade from the SIR(S-V) to SIR(S-E) blurs the evolution of susceptible, infected, and recovered individuals; the relative effects of vaccination programs; or the best timing for interventions from t > 80 onward (Fig. 5). If the goal of the model is to gain insights into the effects that vaccination and nonpharmaceutical interventions may have on the spread of the virus, then the SIR(S-V) might be preferred over the more complex SIR(S-E) because the extra detail in the latter does not help clarify potential courses of action. However, if the goal of the SIR-based model is to inform on the potential evolution of infected and susceptible individuals, then the SIR(S) might be favored over the SIR(S-V) because the higher effective dimensions of the latter do not provide any significant additional insights into the possible dynamics of the epidemic to justify its higher detail (Fig. 5).
Fig. 5. Monte Carlo propagation of uncertainties in three SIR-based models.
Models from Saad-Roy et al. (45, 46). Each of the 128 lines in each state variable and model reflects a simulation conducted with parameters taking specific values according to their uncertainty range. V, vaccinated; V1, one-dose vaccinal immunity; V2, two-dose vaccinal immunity; IV, infection after vaccination; SS1, waned one-dose immunity; SS2, waned two-dose immunity; IS1, infection after waned one-dose immunity; IS2, infection after waned two-dose immunity. For the rest of the state variables, see the caption of Fig. 4.
DISCUSSION
Here, we analyze the relation between complexity and uncertainty in process-based mathematical models that do not (or cannot) rely on a training and/or validation dataset. Such models, ubiquitous and often prevalent at the science-policy interface, provide insights into complex issues that may escape verification: How many radionuclides can leak from a waste repository and get into humans via drinking water? How will a new pandemic spread? How much water for irrigation will be needed worldwide by 2050? Will the detonation of the first nuclear device set fire to the atmosphere? What is the cost of CO2 averaged over the next century? For these types of questions, we show that the addition of model detail promotes fuzzier estimates because it generally increases the model’s effective dimensions kt (the number of influential parameters) and ks (the order of the highest-order effect active in the model). This makes the model’s uncertain space more dominated by interactions, which promote larger uncertainties. The addition of complexity, rather than sharpen the accuracy of the estimation, may swamp it under indeterminacy.
Greater complexity can be useful if the scope of the mathematical model is to explore the implications of its underpinning assumptions and theories. A higher level of detail permits the modeled process to exhibit in full its range of variation once the hyperspace of its input parameters is thoroughly explored, an insight that may help detect hitherto unexpected behaviors. In other words, “model responsibly” rather than “simple is beautiful” is the stance of the present work. What is undesirable is the coupling of excess complexity with the illusory accuracy that results from scarce attention to the propagation of uncertainties onto the prediction of the model. This practice may miss on the unanticipated behaviors, foster tunnel vision, and underestimate serious socioenvironmental risks such as the effects of climate change (47), the impact of insecticides on bees (48) or the environmental impact of irrigated agriculture (40, 43). Among practitioners, this pathology is known as “garbage in, garbage out” (“where uncertainties in inputs must be suppressed lest outputs become indeterminate”) (49).
Examining the link between effective dimensions and output uncertainty at each model upgrade stage may improve the quality of process-based models. The lack of validation data at the spatial/temporal scales required means that these models often only have access to the measurement error part of the systemic bias and measurement error trade-off, known as the O’Neill (50) conjecture in ecology, as Zadeh’s (51) principle of incompatibility in systems analysis, or as the “bias-variance trade-off” in machine learning (52). The concept of effective dimensions, however, enhances the examination of measurement error by tracking the uncertainty buildup at each model upgrade stage and facilitating consideration of a broad set of quality criteria beyond goodness-of-fit statistics, such as model purpose, complexity, and transparency. It sets the ground for a reflective (and not mechanical) approach to model selection, along the lines of what Fisher and Neyman recommended as best practices in inferential statistics (53). Unlike information criterion, the calculation of the effective dimensions can also integrate epistemic uncertainties [uncertainty due to the imperfection of our knowledge (54)], which often expand when models get enlarged to include new processes. This exercise ensures that model complexity remains within bounds compatible with the quality of the evidence fed into the model itself, ultimately providing modelers with insights to help them select the “most appropriate” model.
Our approach is especially valuable for models that aim at playing a role in policy arenas. Pinpointing the model’s effective dimensions can contribute to take more robust actions because the space of possible options is no longer artificially constrained by spuriously accurate model outputs. Both modelers and their constituency gain by a process of transparent communication of the uncertainty, which results from proper monitoring of each model upgrade stage. This facilitates the identification of the “stopping point” beyond which the addition of extra detail no longer makes the model suitable, thus promoting a reciprocal domestication between models and society (15). Such a control is especially needed to prevent policy-oriented models from getting as complex as to cloak value-laden assumptions and output uncertainties (55). For instance, how many readers outside the circle of practitioners know that the 1.5 degree climate target is achieved in integrated assessment models thanks to the adoption of so-called negative emissions, a technology still to be fully developed (56)? Or that global hydrological models assume that irrigation optimizes crop production and water use (a premise at odds with the practices of traditional irrigators) (40)?
Since larger-scale models can command more epistemic authority (57), modelers may be tempted to add detail to reinforce their status as influential actors at the science-policy interface and gain recognition and reward (58). More complex models are also more ductile to political calibration, a sequential process of continuously refining the fit between modeling and policy requirements (56). In this situation, the model becomes a moving target as it is continuously rearranged to include new data and processes to match evolving policy needs, with no evidence as to how it works prospectively. This makes larger models more prone to fall into the “Texas sharpshooter” fallacy, where you shoot first and then draw the bull’s-eye where the bullet hits. The calculation of the model’s effective dimensions may contribute to prevent or minimize these biases by systematically unfolding the balance between model prowess, uncertainty, and policy goals.
Our work also stresses the importance of stripping mathematical models of superfluous parameters, processes, or linkages. This is related to Occam’s razor principle by which “entities should not be multiplied without necessity” (59). There are few reasons to keep models in their nominal dimension k when k ≫ kt, because such inequality points to unneeded redundancy. An example is the irrigation module of many global hydrological models (60, 61): Although computationally demanding and formed by several parameters and equations, their behavior is ultimately very similar to a linear regression of irrigated areas (62). This suggests the existence of a mismatch between their “real” complexity and their computational requirements. The concept of effective dimensions, by distilling the model from superfluous elements, refines also the notion of algorithmic complexity down to its “true” value (23).
The key dimension determining the model output uncertainty is ks, i.e., the order of the highest-order effect active in the model. This dimension cannot be pinpointed by intuition alone and is often computationally elusive if k is high. However, the influence of high-order effects can be pondered with the calculation of the total-order index Ti. This requires submitting the model to a global sensitivity analysis (63). If many inputs are uncertain and the model is time-consuming to run, this may prove impossible: A model that takes days to run for a single combination of its inputs cannot be run hundreds, let alone thousands of times, to properly characterize its uncertainty space (64). The trend toward increasingly complex process-based mathematical models discussed in Introduction may be grounded in this paradox: More detailed models may be thought of as more accurate simply because their very design complicates any attempt at proving otherwise.
MATERIALS AND METHODS
Variance-decomposition framework
Given a model of the form y = f(x), x = (x1, x2, …, xi, …, xk) ∈ ℝk, where y is a scalar output and x1, …, xk are k independent parameters whose uncertainty is described by probability distributions, the variance of the output V(y) can be decomposed as
(3) |
where Vi, Vi,j, … are, respectively, the conditional variances of xi, (xi, xj), … on y. Equation 3 is linked to Sobol’s (25) functional decomposition scheme
(4) |
given that
(5) |
where E(.) is the mean operator and x∼i denotes all parameters except the ith, and therefore
(6) |
Sobol’s (25) indices are then calculated as
(7) |
where Si is the first-order effect of xi, Si,j is the second-order effect of (xi, xj), etc.
The total-order index (Ti) captures the proportion of variance in y propagated by the first-order effect of xi jointly with its interactions up to the kth order (26). In other words, Ti includes all terms in Eq. 3 with the index i and is computed as
(8) |
Uncertainty and sensitivity analysis
To assess how the addition of model complexity increases the output uncertainty, we submit all models and functions studied here to global uncertainty and sensitivity analysis. For each model, we construct a (N, k) Q matrix, where k is the number of model parameters and N is the row dimension, using Sobol’s quasi-random numbers (65). We allocate the leftmost k columns to an A matrix and the rightmost k columns to a B matrix. In these matrices, each column is a parameter and each row is a sampling point. Any sampling point in A or B can be indicated as xvi, where v indexes the row (from 1 to N) and i indexes the column (from 1 to k). We also construct k () matrices, where all parameters come from the A (B) matrix except the ith, which comes from the B (A) matrix.
The model f runs row-wise in each matrix, and we use the model output y to compute the coefficient of variance and the sensitivity indices. We compute Si and Ti with the Azzini estimators (66)
(9) |
(10) |
We conduct all of the uncertainty and sensitivity analyses with the sensobol package in R (67). We provide a detailed description of the models and of the probability distributions used to characterize parametric uncertainties in the Supplementary Materials.
Sobol’s G function
To create Fig. 1A, we randomize in a Monte Carlo setting the coefficients of Sobol’s G function (68), an analytically tractable, well-known function among modelers. It reads as
(11) |
where xi ∼ 𝒰(0,1), ai ∈ ℝ+, i = 1,2, …, k. The typology of the function is driven by k and the value of the coefficients ai: If more than one parameter has a low a, then high-order interactions will be present. The more parameters with a low a, the stronger the interactions and the higher the ks dimension of the function. Here, we sample with replacement the coefficients from the set {0,1,4.5,9,99}, weighted as {0.4,0.3,0.2,0.05,0.05} to promote the creation of nonadditive functions. This approach allows one to explore the connection between k, kt, ks, and output uncertainty in a wide range of model settings and across models with very different behaviors.
The meta-model
Our meta-model is based on the Becker (29) metafunction. Let u = {u1, u2, …, uk} be a k-length vector formed by randomly sampling integer values from 1 to 13, in which each integer value is linked to 1 of the 13 univariate functions in fig. S1. We then apply the ith function in u to the ith model input: If k = 3 and u = {u3, u1, u12}, then , , and .
We also randomly activate up to the nth-order interaction, for n ≤ k, and create n matrices V, …, W of dimension k2 × 2, …, kn × n, where . Each row in these matrices represents an nth-order interaction between the model inputs. Overall, this model can be formalized as
(12) |
where the first term represents the first-order effects, the second term the second-order effects, and so on up to the nth-order effect.
Acknowledgments
We thank A. Lenardic (Rice University) and two anonymous reviewers for peer review of the manuscript.
Funding: This work has been funded by the European Commission (Marie Skłodowska-Curie Global Fellowship, grant number 792178 to A.P.) and by the C3.ai Inc. and Microsoft Corporation, the Army Research Office Grant W911NF-18-1-0325, the NSF (CCF1917819), and by a gift from Google LLC.
Author contributions: All authors designed the manuscript. A.P. and A.S. conceptualized the theoretical approach. A.S. contributed the formulation, the description, and the encoding of the PSACOIN Level 0 model. A.P. and S.L.P coded the other models. A.P. ran the simulations. A.P., P.B., S.A.L., S.L.P., and A.S. discussed the results. A.P. wrote the manuscript, with contributions from P.B., S.A.L., and A.S. A.P., P.B., S.A.L., and S.L.P. corrected and revised the final version.
Competing interests: The authors declare that they have no competing interests.
Data and materials availability: The code to reproduce our results is available in Zenodo (69) (https://doi.org/10.5281/zenodo.5658383). All data needed to evaluate the conclusions in the paper are present in the paper and/or the Supplementary Materials.
Supplementary Materials
This PDF file includes:
Figs. S1 to S4
The Models
Tables S1 to S5
References
REFERENCES AND NOTES
- 1.Wood E. F., Roundy J. K., Troy T. J., van Beek L. P. H., Bierkens M. F. P., Blyth E., de Roo A., Döll P., Ek M., Famiglietti J., Gochis D., van de Giesen N., Houser P., Jaffé P. R., Kollet S., Lehner B., Lettenmaier D. P., Peters-Lidard C., Sivapalan M., Sheffield J., Wade A., Whitehead P., Hyperresolution global land surface modeling: Meeting a grand challenge for monitoring earth’s terrestrial water. Water Resour. Res. 47, W05301 (2011). [Google Scholar]
- 2.Palmer T., Climate forecasting: Build high-resolution global climate models. Nature 515, 338–339 (2014). [DOI] [PubMed] [Google Scholar]
- 3.Bierkens M. F. P., Bell V. A., Burek P., Chaney N., Condon L. E., David C. H., de Roo A., Döll P., Drost N., Famiglietti J. S., Flörke M., Gochis D. J., Houser P., Hut R., Keune J., Kollet S., Maxwell R. M., Reager J. T., Samaniego L., Sudicky E., Sutanudjaja E. H., van de Giesen N., Winsemius H., Wood E. F., Hyper-resolution global hydrological modelling: What is next?: “Everywhere and locally relevant”. Hydrol. Process. 29, 310–320 (2015). [Google Scholar]
- 4.Moore G. E., Cramming more components onto integrated circuits. Electronics 38, 114 (1965). [Google Scholar]
- 5.Dennard R. H., Gaensslen F. H., Yu H.-N., Rideout V. L., Bassous E., LeBlanc A. R., Design of ion-implanted mosfet’s with very small physical dimensions. IEEE J. Solid State Circuits 9, 256–268 (1974). [Google Scholar]
- 6.Hausfather Z., Drake H. F., Abbott T., Schmidt G. A., Evaluating the performance of past climate model projections. Geophys. Res. Lett. 47, e2019GL085378 (2020). [Google Scholar]
- 7.Sarofim M. C., Smith J. B., Juliana A. S., Hartin C., Improving reduced complexity model assessment and usability. Nat. Clim. Chang. 11, 1–3 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Manabe S., Climate and the ocean circulation I. The atmospheric circulation and the hydrology of the Earth’s surface. Mon. Weather Rev. 97, 739–774 (1969). [Google Scholar]
- 9.Bierkens M. F. P., Global hydrology 2015: State, trends, and directions. Water Resour. Res. 51, 4923–4947 (2015). [Google Scholar]
- 10.Kermack W. O., McKendrick A. G., Contributions to the mathematical theory of epidemics. III.—Further studies of the problem of endemicity. Proc. R. Soc. Lond. 141, 94–122 (1933). [Google Scholar]
- 11.N. M. Ferguson, D. Laydon, G. Nedjati Gilani, N. Imai, K. Ainslie, M. Baguelin, S. Bhatia, A. Boonyasiri, Z. Cucunuba, G. Cuomo-Dannenburg, A. Dighe, I. Dorigatti, H. Fu, K. Gaythorpe, W. Green, A. Hamlet, W. Hinsley, L. C. Okell, S. Van Elsland, H. Thompson, R. Verity, E. Volz, H. Wang, Y. Wang, P. G. T. Walker, C. Walters, P. Winskill, C. Whittaker, C. A. Donnelly, S. Riley, A. C. Ghani, Report 9: Impact of non-pharmaceutical interventions (NPIs) to reduce COVID-19 mortality and healthcare demand, Tech. Rep. March (Imperial College COVID-19 Response Team, 2020).
- 12.Refsgaard J. C., van der Sluijs J., Brown J., van der Keur P., A framework for dealing with uncertainty due to model structure error. Adv. Water Resour. 29, 1586–1597 (2006). [Google Scholar]
- 13.H. Akaike, Akaike’s Information Criterion, in International Encyclopedia of Statistical Science, M. Lovric, Ed. (Springer Berlin Heidelberg, 2011), p. 25. [Google Scholar]
- 14.Schwarz G., Estimating the dimension of a model. Ann. Stat. 6, 461–464 (1978). [Google Scholar]
- 15.Saltelli A., Bammer G., Bruno I., Charters E., Di Fiore M., Didier E., Nelson Espeland W., Kay J., Lo Piano S., Mayo D., Pielke R. Jr., Portaluri T., Porter T. M., Puy A., Rafols I., Ravetz J. R., Reinert E., Sarewitz D., Stark P. B., Stirling A., van der Sluijs J., Vineis P., Five ways to ensure that models serve society: A manifesto. Nature 582, 482–484 (2020). [DOI] [PubMed] [Google Scholar]
- 16.O. H. Pilkey, L. Pilkey-Jarvis, Useless Arithmetic: Why Environmental Scientists Can’t Predict the Future (Columbia Univ. Press, 2009). [Google Scholar]
- 17.Caflisch R. E., Morokoff W., Owen A., Valuation of mortgage-backed securities using Brownian bridges to reduce effective dimension. J. Comput. Finance 1, 27–46 (1997). [Google Scholar]
- 18.Wang X., Fang K. T., The effective dimension and quasi-Monte Carlo integration. J. Complex. 19, 101–124 (2003). [Google Scholar]
- 19.Kucherenko S., Feil B., Shah N., Mauntz W., The identification of model effective dimensions using global sensitivity analysis. Reliab. Eng. Syst. Saf. 96, 440–449 (2011). [Google Scholar]
- 20.Saltelli A., Aleksankina K., Becker W., Fennell P., Ferretti F., Holst N., Li S., Wu Q., Why so many published sensitivity analyses are false: A systematic review of sensitivity analysis practices. Environ. Model. Software 114, 29–39 (2019). [Google Scholar]
- 21.Brooks R. J., Tobias A. M., Choosing the best model: Level of detail, complexity, and model performance. Math. Comput. Model. 24, 1–14 (1996). [Google Scholar]
- 22.Simon H. A., The architecture of complexity. Proc. Am. Philos. Soc. 106, 467–482 (1962). [Google Scholar]
- 23.Manson S. M., Simplifying complexity: A review of complexity theory. Geoforum 32, 405–414 (2001). [Google Scholar]
- 24.Regan H. M., Colyvan M., Burgman M. A., A taxonomy and treatment of uncertainty for ecology and conservation biology. Ecol. Appl. 12, 618–628 (2002). [Google Scholar]
- 25.Sobol’ I. M., Sensitivity estimates for nonlinear mathematical models. Math. Model. Comput. Exp. 4, 407–414 (1993). [Google Scholar]
- 26.Homma T., Saltelli A., Importance measures in global sensitivity analysis of nonlinear models. Reliab. Eng. Syst. Saf. 52, 1–17 (1996). [Google Scholar]
- 27.V. Pareto, Manuale di Economia Politica (Societa Editrice, 1906), vol 13. [Google Scholar]
- 28.Box G. E. P., Meyer R. D., An analysis for unreplicated fractional factorials. Dent. Tech. 28, 11–18 (1986). [Google Scholar]
- 29.Becker W., Metafunctions for benchmarking in sensitivity analysis. Reliab. Eng. Syst. Saf. 204, 107189 (2020). [Google Scholar]
- 30.B. Edmonds, “Syntactic measures of complexity,” thesis, University of Manchester (1999). [Google Scholar]
- 31.Puy A., Becker W., Lo Piano S., Saltelli A., A comprehensive comparison of total-order estimators for global sensitivity analysis. Int. J. Uncertain. Quantif. 12, 1–18 (2022). [Google Scholar]
- 32.Wilby R. L., Dessai S., Robust adaptation to climate change. Weather 65, 180–185 (2010). [Google Scholar]
- 33.Maslin M., Cascading uncertainty in climate change models and its implications for policy. Geogr. J. 179, 264–271 (2013). [Google Scholar]
- 34.Pappenberger F., Beven K. J., Hunter N. M., Bates P. D., Gouweleeuw B. T., Thielen J., de Roo A. P., Cascading model uncertainty from medium range weather forecasts (10 days) through a rainfall-runoff model to flood inundation predictions within the European Flood Forecasting System (EFFS). Hydrol. Earth Syst. Sci. 9, 381–393 (2005). [Google Scholar]
- 35.Magal P., Ruan S., Susceptible-infectious-recovered models revisited: From the individual level to the population level. Math. Biosci. 250, 26–40 (2014). [DOI] [PubMed] [Google Scholar]
- 36.B. W. Goodwin, A. E. C. L. Canada, A. Saltelli, PSACOIN Level 0 Intercomparison, Tech. Rep. (Nuclear Energy Agency, Organisation for Economic Co-operation and Development, Paris, 1987).
- 37.D. Metlay, From tin roof to torn wet blanket: Predicting and observing ground water movement at a proposed nuclear waste site, in Prediction: Science, Decision Making, and the Future of Nature, D. Sarewitz, R. A. Pielke Jr., R. Byerly Jr., Eds. (Island Press, 2000). [Google Scholar]
- 38.D. Sarewitz, “Of cold mice and isotopes or should we do less science?,” in Exploring Relations Between Academic Research, Higher Education and Science Policy. Summer School in Higher Education Research and Science Studies (Universität Bonn, 2018), pp. 1–8.
- 39.Brugnach M., Process level sensitivity analysis for complex ecological models. Ecol. Model. 187, 99–120 (2005). [Google Scholar]
- 40.Puy A., Sheikholeslami R., Gupta H. V., Hall J. W., Lankford B., Lo Piano S., Meier J., Pappenberger F., Porporato A., Vico G., Saltelli A., The delusive accuracy of global irrigation water withdrawal estimates. Nat. Commun. 13, 3183 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Lu J., Sun G., McNulty S. G., Amatya D. M., A comparison of six potential evapotranspiration methods for regional use in the southeastern United States. J. Am. Water Resour. Assoc. 41, 621–633 (2005). [Google Scholar]
- 42.McMahon T. A., Peel M. C., Lowe L., Srikanthan R., McVicar T. R., Estimating actual, potential, reference crop and pan evaporation using standard meteorological data: A pragmatic synthesis. Hydrol. Earth Syst. Sci. 17, 1331–1363 (2013). [Google Scholar]
- 43.Puy A., Lo Piano S., Saltelli A., Current models underestimate future irrigated areas. Geophys. Res. Lett. 47, e2020GL087360 (2020). [Google Scholar]
- 44.Pitzer V. E., Viboud C., Simonsen L., Steiner C., Panozzo C. A., Alonso W. J., Miller M. A., Glass R. I., Glasser J. W., Parashar U. D., Grenfell B. T., Demographic variability, vaccination, and the spatiotemporal dynamics of rotavirus epidemics. Science 325, 290–294 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Saad-Roy C. M., Wagner C. E., Baker R. E., Morris S. E., Farrar J., Graham A. L., Levin S. A., Mina M. J., Metcalf C. J. E., Grenfell B. T., Immune life history, vaccination, and the dynamics of SARS-CoV-2 over the next 5 years. Science 370, 811–818 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Saad-Roy C. M., Morris S. E., Metcalf C. J. E., Mina M. J., Baker R. E., Farrar J., Holmes E. C., Pybus O. G., Graham A. L., Levin S. A., Grenfell B. T., Wagner C. E., Epidemiological and evolutionary considerations of SARS-CoV-2 vaccine dosing regimes. Science 372, 363–370 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Stern P. C., Perkins J. H., Sparks R. E., Knox R. A., The challenge of climate-change neoskepticism. Science 353, 653–654 (2016). [DOI] [PubMed] [Google Scholar]
- 48.Maxim L., van der Sluijs J. P., Uncertainty: Cause or effect of stakeholders’ debates? Sci. Total Environ. 376, 1–17 (2007). [DOI] [PubMed] [Google Scholar]
- 49.S. Funtowicz, J. R. Ravetz, Uncertainty and Quality in Science for Policy (Kluwer Academic Publishers, 1990). [Google Scholar]
- 50.R. V. O’Neill, Error analysis of ecological models, in Radionuclides in Ecosystems. Proceedings of the Third National Symposium on Radioecology, May 10–12, 1971, Oak Ridge, Tennessee, D. J. Nelson, Ed. (Springfield, 1971). [Google Scholar]
- 51.Zadeh L. A., Outline of a new approach to the analysis of complex systems and decision processes. IEEE Trans. Syst. Man Cybern. SMC-3, 28–44 (1973). [Google Scholar]
- 52.Geman S., Bienenstock E., Doursat R., Neural networks and the bias/variance dilemma. Neural Comput. 4, 1–58 (1992). [Google Scholar]
- 53.Gigerenzer G., Marewski J. N., Surrogate science: The idol of a universal method for scientific inference. J. Manag. 41, 421–440 (2015). [Google Scholar]
- 54.Walker W. E., Marchau V. A., Swanson D., Addressing deep uncertainty using adaptive policies: Introduction to section 2. Technol. Forecast. Soc. Change 77, 917–923 (2010). [Google Scholar]
- 55.Robertson S., Transparency, trust, and integrated assessment models: An ethical consideration for the intergovernmental panel on climate change. Wiley Interdiscip. Rev. Clim. Chang. 12, e679 (2021). [Google Scholar]
- 56.van Beek L., Oomen J., Hajer M., Pelzer P., van Vuuren D., Navigating the political: An analysis of political calibration of integrated assessment modelling in light of the 1.5 °C goal. Environ. Sci. Policy 133, 193–202 (2022). [Google Scholar]
- 57.Saltelli A., Pereira A. G., van der Sluijs J. P., Funtowicz S. O., What do I make of your latinorum? Sensitivity auditing of mathematical modelling. Int. J. Innov. Pol. 9, 213–234 (2013). [Google Scholar]
- 58.E. S. Quade, Pitfalls in formulation and modeling, in Pitfalls of Analysis, G. Majone, E. S. Quade, Eds. (John Wiley & Sons, 1980), pp. 23–43. [Google Scholar]
- 59.B. Duignan, Occam’s razor, in Encyclopedia Britannica (2021), [accesssed 24 August 2022]; www.britannica.com/topic/Occams-razor.
- 60.Döll P., Siebert S., Global modeling of irrigation water requirements. Water Resour. Res. 38, 8–1–8–10 (2002). [Google Scholar]
- 61.Sutanudjaja E. H., van Beek L. P. H., Drost N., de Graaf I. E. M., de Jong K., Peßenteiner S., Straatsma M. W., Wada Y., Wanders N., Wisser D., Bierkens M. F. P., PCR-GLOBWB 2: A 5 arcmin global hydrological and water resources model. Geosci. Model Dev. 11, 2429–2453 (2018). [Google Scholar]
- 62.Puy A., Borgonovo E., Lo Piano S., Levin S. A., Saltelli A., Irrigated areas drive irrigation water withdrawals. Nat. Commun. 12, 4525 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.A. Saltelli, M. Ratto, T. Andres, F. Campolongo, J. Cariboni, D. Gatelli, M. Saisana, S. Tarantola, Global Sensitivity Analysis. The Primer (John Wiley & Sons Ltd, 2008). [Google Scholar]
- 64.Parker W. S., Ensemble modeling, uncertainty and robust predictions. Wiley Interdiscip. Rev. Clim. Chang. 4, 213–223 (2013). [Google Scholar]
- 65.Sobol’ I. M., On the distribution of points in a cube and the approximate evaluation of integrals. USSR Comput. Math. Math. Phys. 7, 86–112 (1967). [Google Scholar]
- 66.Azzini I., Mara T., Rosati R., Comparison of two sets of Monte Carlo estimators of Sobol’ indices. Environ. Model. Software 144, 105167 (2021). [Google Scholar]
- 67.Puy A., Lo Piano S., Saltelli A., Levin S. A., sensobol: An R package to compute variance-based sensitivity indices. J. Stat. Softw. 102, 1–37 (2022). [Google Scholar]
- 68.Sobol’ I. M., On quasi-Monte Carlo integrations. Math. Comput. Simul. 47, 103–112 (1998). [Google Scholar]
- 69.A. Puy, R code for “Models with higher effective dimensions tend to produce more uncertain estimates”. Zenodo (2021); 10.5281/zenodo.5658383. [DOI] [PMC free article] [PubMed]
- 70.Allen R. G., Pereira L. S., Raes D., Smith M., Crop evapotranspiration: Guidelines for computing crop water requirements. Irrig. Drain. 300, 300 (1998). [Google Scholar]
- 71.A. Puy, R. Sheikholeslami, H. V. Gupta, J. W. Hall, B. Lankford, S. Lo Piano, J. Meier, F. Pappenberger, A. Porporato, G. Vico, A. Saltelli, Supplementary Materials for The delusive accuracy of global irrigation water withdrawal estimates. Zenodo (2021); 10.5281/zenodo.5528844. [DOI] [PMC free article] [PubMed]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Figs. S1 to S4
The Models
Tables S1 to S5
References