Skip to main content
PNAS Nexus logoLink to PNAS Nexus
. 2024 Dec 10;3(12):pgae554. doi: 10.1093/pnasnexus/pgae554

Uncertainty quantification in coupled wildfire–atmosphere simulations at scale

Paul Schwerdtner 1, Frederick Law 2, Qing Wang 3, Cenk Gazen 4, Yi-Fan Chen 5, Matthias Ihme 6,7, Benjamin Peherstorfer 8,b,
Editor: Dennis Discher
PMCID: PMC11660924  PMID: 39712072

Abstract

Uncertainties in wildfire simulations pose a major challenge for making decisions about fire management, mitigation, and evacuations. However, ensemble calculations to quantify uncertainties are prohibitively expensive with high-fidelity models that are needed to capture today’s ever-more intense and severe wildfires. This work shows that surrogate models trained on related data enable scaling multifidelity uncertainty quantification to high-fidelity wildfire simulations of unprecedented scale with billions of degrees of freedom. The key insight is that correlation is all that matters while bias is irrelevant for speeding up uncertainty quantification when surrogate models are combined with high-fidelity models in multifidelity approaches. This allows the surrogate models to be trained on abundantly available or cheaply generated related data samples that can be strongly biased as long as they are correlated to predictions of high-fidelity simulations. Numerical results with scenarios of the Tubbs 2017 wildfire demonstrate that surrogate models trained on related data make multifidelity uncertainty quantification in large-scale wildfire simulations practical by reducing the training time by several orders of magnitude from 3 months to under 3 h and predicting the burned area at least twice as accurately compared with using high-fidelity simulations alone for a fixed computational budget. More generally, the results suggest that leveraging related data can greatly extend the scope of surrogate modeling, potentially benefiting other fields that require uncertainty quantification in computationally expensive high-fidelity simulations.

Keywords: uncertainty quantification, multifidelity methods, neural networks, surrogate modeling, wildfire simulations


Significance Statement.

Today’s wildfires increasingly spread into populated areas, putting millions of homes at risk and impacting air quality. Numerical wildfire simulations are key building blocks for risk assessment and fire management, but uncertainties from environmental measurements must be quantified to establish trust for making high-consequence decisions. This work shows that surrogate models trained on related data make tractable uncertainty quantification with high-fidelity wildfire simulations of unprecedented scale with billions of degrees of freedom, which offers opportunities for using more complex wildfire models that can better aid fire management and evacuation planning. More generally, leveraging related data for surrogate modeling can be applied across other fields that require uncertainty quantification in large-scale simulations, such as climate modeling, aerospace engineering, and plasma physics.

Introduction

Today’s wildfires grow faster, burn hotter, and spread more often into populated areas than ever before (1, 2). While numerical simulations are key for developing more efficient warning and prediction systems (3, 4), they are affected by large uncertainties that can pose a major challenge for making decisions about fire management, mitigation, and evacuations (5, 6). In particular, the environmental conditions such as fuel load and wind condition that are given as inputs to wildfire simulations are a major source of uncertainty due to measurement inaccuracies and sparse or incomplete data (7). It, therefore, is essential to equip numerical wildfire predictions with mean estimates, standard deviations, and sensitivities that account for input uncertainties so that decision-makers can decide how much they trust the predictions and act accordingly. However, uncertainty quantification in wildfire simulations is challenging (8). High-fidelity wildfire models that include fluid dynamics with combustion models to capture the fire–atmosphere interactions are computationally demanding (9–12), which means that even small ensemble sizes for uncertainty quantification can lead to infeasible compute runtimes (13). This has led to a use of gross simplifications in operational models that limit predictive capabilities (6, 14).

Surrogate models trained on related data

In this work, we scale uncertainty quantification to high-fidelity coupled fire–atmosphere simulations with billions of degrees of freedom to accurately estimate mean quantities of interest from ensembles of fire simulations. We propose an approach that trains surrogate models on related data, which are often abundantly available from previous, related physics simulations or can be cheaply generated via simplified models obtained by ignoring some of the physical phenomena, linearizing dynamics, or stopping iterative solvers early (15), see Fig. 1. We contrast related data to direct data that correspond to actual outputs or reanalysis quantities of the high-fidelity numerical simulations, which are used for training in traditional surrogate modeling but are typically prohibitively expensive to generated in many cases (16–19). We also contribute a mathematical analysis that provides a foundation for our approach and insight into the performance of the approach with respect to the quality of the related data. We demonstrate the approach on a fire scenario resembling the Tubbs 2017 wildfire, where related data are generated with a down-scaled numerical simulation in 3 h, whereas generating the same amount of direct data with the high-fidelity model would require 3 months, and thus is intractable.

Fig. 1.

Fig. 1.

Training surrogate models on related data for multifidelity uncertainty quantification: (i) Simplifications are made to the physics model such as formulating it over smaller domains, ignoring some phenomena, linearizing, and stopping iterative numerical solvers early. (ii) The corresponding simplified model is simulated many times to rapidly generate large volumes of outputs that form the related training data. (iii) A surrogate model is trained on the related data from the simplified model. (iv) For realizations of the uncertain inputs such as environmental conditions, a few output samples are computed with the expensive physics model and many output samples are obtained with the affordable surrogate model. (v) The samples from physics and surrogate model are combined into unbiased multifidelity estimators of expectations and variances of the quantities of interest.

Surrogate models trained on related data provide outputs that generally are not predictive about high-fidelity simulations because related data samples typically have a large bias with respect to outputs of the high-fidelity simulations. Thus, also the surrogate models will have a large bias. The key insight is that related data samples are often still correlated and that correlation is all that matters while bias is irrelevant for speeding up uncertainty quantification when surrogate models are combined with high-fidelity models in multifidelity approaches (15, 20). Multifidelity uncertainty quantification methods leverage surrogate models to accelerate the estimation of uncertainties while, in a rigorous manner, occasionally utilize expensive high-fidelity models to establish unbiased uncertainty estimators. Mathematically, we focus on the correlation that is captured by the Pearson moment correlation coefficient (21), because it is useful in multifidelity uncertainty quantification (15). This means that the outputs of our surrogate models and the wildfire simulations show similar responses to changes in inputs even though the bias can be high in the sense that the absolute values of the outputs can be vastly different. For example, increasing wind speed means faster fire spread, higher fuel density leads to more heat release; such trends are captured at least approximately by surrogate models trained on related data with data samples that are correlated to the high-fidelity simulation outputs. In contrast, traditional surrogate modeling that learns from direct data aims to accurately approximate the actual fire spread and the actual fire temperature rather than just how the fire spread and temperature change with respect to inputs.

Literature review

Uncertainty quantification has been extensively studied for wildfire simulations; however, due to computational costs, only empirical or heuristic fire models are used (6, 14). There is work on using multilevel methods (22) but we consider orders of magnitude larger wildfire models in terms of number of degrees of freedom. Furthermore, the previous works rely either on coarse-grid approximations that provide limited speedup or traditional surrogate models that aim to keep the bias low and so require large amounts of direct training data.

While machine learning and artificial intelligence have a major impact on surrogate modeling (17, 23–34), generating sufficient direct training data samples can require many computationally expensive physics-based simulations, which is often intractable. This is one reason why there is an increasing body of literature (35–41) on training surrogate models on multifidelity and related data; however, there the aim is to use related data to improve the training of traditional surrogate models that aim to keep the bias low, which is in contrast to the approach proposed here that accepts bias as long as surrogate-model outputs are correlated, because the surrogate models are used in conjunction with multifidelity methods (20). Analogously, surrogate models based on concepts of transfer learning and one-shot learning (42, 43) aim to rapidly fine-tune a pretrained model on a new task with just a few, direct (i.e. labeled) data points. None of these studies realizes the potential of surrogate models that are trained purely on related data and thus are only correlated while having a large bias. Another line of work aims to construct surrogate models explicitly for use in multifidelity methods (44, 45), which can require less training data than generic methods; however, the training data are still direct in the sense that they correspond to input–output pairs computed with high-fidelity, physics-based models rather than related and biased data as we use in this work. Yet another line of work aims to scale surrogate modeling to large-scale settings with little training data by first performing a sensitivity analysis and sub-selecting only a few input components over which surrogate models are trained (46, 47); however, such an approach requires conducting potentially expensive sensitivity analyses first.

Surrogate models from related data for multifidelity uncertainty quantification at scale

Physics models with stochastic inputs

We denote a physics model as a function f:XY that maps an input vector xX, which consists of environmental conditions and model parameters, onto an output vector y=f(x)Y, which consists of quantities of interest that are computed from the simulation output such as the burned area. Evaluating the function f means performing a numerical simulation that typically incurs high computational costs, which we denote as cf>0. Notice that we consider situations where the challenge lies with the high computational costs of evaluating f, rather than the high dimension of the inputs and outputs. In fact, we consider situations where there is only a low number of inputs and outputs, which is common in many science and engineering applications.

To account for incomplete knowledge and uncertain environmental conditions, we consider random input vectors X that follow a distribution that models input uncertainties. For example, measuring the wind speed is affected by measurement errors and noise due to data sparsity in spatial coverage and temporal resolution. In some cases, inputs have to be first obtained from upstream simulations or inverse problems, which are also affected by uncertainties that need to be carried forward. Consequently, instead of computing a deterministic output y, we are interested in estimating statistics of the outputs such as expected outputs E[f(X)] and variances. Notice that only the inputs are stochastic whereas the map f corresponding to the numerical simulation is deterministic. In the following, we omit X when denoting expected values E[f], variances, and correlation coefficients because all of them are taken with respect to the distribution of the inputs X.

A classical approach for estimating statistics is via ensembles of m independent and identically distributed (i.i.d.) samples x1,,xm of the random input vector X and the corresponding ensemble of m output samples y1=f(x1),,ym=f(xm). The expected output E[f] is then estimated via Monte Carlo estimation as

y¯m(f)=1mi=1mf(xi). (1)

Obtaining an accurate estimate in Eq. (1) is challenging: Recall that each evaluation of f entails a numerical simulation, which can be expensive and thus severely limits the ensemble size m that is tractable.

Surrogate models and multifidelity uncertainty quantification

Surrogate models provide only approximations of the outputs computed by the high-fidelity physics model f but often with orders of magnitude lower runtimes, see (16–19). We denote a surrogate model as gθ:XY, which depends on a parameter vector θRp such as the weights of a neural network. The costs of evaluating the surrogate model are denoted as cg and are lower cgcf than the costs cf of performing a physics-based simulation with f. Using the surrogate gθ instead of the high-fidelity physics model f in the Monte Carlo estimator in Eq. (1) leads to speedups, but it also means that the estimator is biased with respect to E[f] because the surrogate model provides an approximation of the outputs of f only. To leverage the surrogate model for achieving runtime speedups while avoiding the introduction of a bias, we use the surrogate model gθ in a multifidelity Monte Carlo estimator (15). The multifidelity Monte Carlo estimator of E[f] is based on a variance reduction technique called control variates (48) and is given by

y¯m1,m2=y¯m1(f)+α(y¯m2(g)y¯m1(g)), (2)

where y¯m1(f) is a Monte Carlo estimator as in Eq. (1) based on m1 samples from f, and y¯m1(g),y¯m2(g) are Monte Carlo estimators based on m1 and m2 samples, respectively, of the surrogate model gθ; details in the Supplementary material. The number of samples m1 and m2 as well as the coefficient α in the estimator in Eq. (2) can be chosen optimally to minimize the mean-squared error (MSE) of y¯m1,m2. We remark that similar multifidelity estimators based on control variates have been introduced for higher-order moments and sensitivity analyses (49), to which our approach would also be applicable.

Using a multifidelity estimator as given in Eq. (2) is key for our approach of training surrogate models on related data: First notice that y¯m1,m2 is unbiased in the sense that E[y¯m1,m2]=E[f], independent of the bias of the surrogate model

b(gθ)=|E[gθ]E[f]| (3)

compared with the physics model f. One can see the unbiasedness of the multifidelity estimator by noting that the expected value E[y¯m2(g)y¯m1(g)] of the difference term in Eq. (2) is zero and thus vanishes, which means that the expected value of the multifidelity estimator is E[y¯m1,m2]=E[y¯m1(f)]=E[f] and thus y¯m1,m2 is an unbiased estimator of E[f]. Thus, even if the surrogate model gθ has a large bias in the sense of Eq. (3), the multifidelity estimator remains unbiased. Second, we want to understand the MSE e(y¯m1,m2) of the estimator in Eq. (2). Because we already know that the multifidelity estimator is unbiased, the MSE of the estimator equals the variance of the estimator, e(y¯m1,m2)=V[y¯m1,m2]. We now need a few more quantities to understand and interpret the MSE of y¯m1,m2. Let σf=V[f] and σg=V[gθ] denote the standard deviations of f and gθ, respectively, with respect to the random input vector X. Let further

ρ(f,gθ)=Cov[f,gθ]σfσg, (4)

denote the Pearson-moment correlation coefficient between the outputs of the physics model f and the surrogate model gθ (21). The operator Cov denotes the covariance. Setting α=σfσg1ρ(f,gθ) in the estimator in Eq. (2), which is the optimal choice that minimizes the variance of the estimator (15), applying transformations based on properties of the variance and leveraging that the samples x1,,xm of the random input vector are independent, one obtains from Eq. (2) that the MSE of the estimator is

e(y¯m1,m2)=σf2m1(1m11m2)σf2ρ(f,gθ)2. (5)

A full derivation of the MSE shown in Eq. (5) as well as the optimal choice for the samples m1 and m2 can be found in (15) and the Supplementary material.

The key quantity for interpreting the MSE is the correlation coefficient ρ(f,gθ). The higher the squared correlation coefficient ρ(f,gθ), the lower the MSE e(y¯m1,m2). Notice that the term 1m11m20 is nonnegative because naturally more samples m1m2 are taken from the cheap surrogate model, see (15) for details about the optimal choice of m1 and m2. It is critical that the MSE given in Eq. (5) depends on the correlation coefficient given in Eq. (4) only; and not on the bias given in Eq. (3) of the surrogate model. Thus, for the surrogate model gθ to be effective in estimating the expected value with the multifidelity estimator given in Eq. (2), it is sufficient that the correlation between the physics model f and the surrogate model gθ is high.

Let us briefly remark on the coefficient α in the estimator in Eq. (2). The coefficient α is a weight of the difference term that includes the surrogate model in the multifidelity estimator. We used the optimal α=σfσg1ρ(f,gθ) in Eq. (2) to derive the MSE given in Eq. (5), where optimal means that it minimizes the variance (15). Intuitively, setting α=σfσg1ρ(f,gθ) means that it weights samples from the surrogate model proportionally to the correlation coefficient and inverse proportionally to the variance. Thus, broadly speaking, a higher correlated surrogate model with low variance leads to a higher weight because the surrogate model can be trusted more than when the correlation is low or the variance of the surrogate model is high.

We further remark that the MSE given in Eq. (5) treats the surrogate model gθ as deterministic and thus ignores potential variations in the surrogate model due to stochastic training and random initializations, which is common when surrogate models are based on neural networks. We discuss this point in more detail in the Supplementary material, where we show that the variations introduced by different random initializations of the training are small compared with the uncertainties introduced by the random input vector X in our numerical experiments.

Training on related data

We now exploit that the MSE given in Eq. (5) of the estimator given in Eq. (2) depends on the correlation coefficient ρ(f,gθ) between the surrogate model gθ and f, and not on the bias given in Eq. (3). Thus, high correlation between gθ and f is sufficient for achieving a low MSE, whereas providing good approximations of the outputs of f in the sense of the bias defined in Eq. (3), i.e. a relative/absolute error, is unnecessary.

Requiring a surrogate model to have a high correlation is often a weaker requirement than having a low bias: Only the trend of the surrogate model gθ and the physics model f have to be similar. In particular, it suffices to train gθ on related data D={(xi,h(xi))}i=1N generated from a data source h:XY that is related to the physics model in the sense that the correlation ρ(f,h) is high. There are often plenty of related data samples available or they can be cheaply generated from related data sources, even in data-scarce science and engineering applications. For example, in our fire simulation application, we cheaply generate related data by running fire simulations on smaller scales obtained by simply shrinking the spatial domains, which leads to fewer degrees of freedom and thus lower runtimes per data sample. While the outputs obtained with the fire simulations on the reduced domains are not accurately approximating the outputs obtained with the high-fidelity, large-scale simulations in the sense of the bias defined in Eq. (3), the outputs are correlated to the outputs of the high-fidelity simulations in the sense of the correlation coefficient given in Eq. (4), see below. Other strategies for generating related data points rely on simplified physics models that ignore some of the phenomena, build on linear approximations, or stop iterative solvers early (20).

Motivated by this, we propose to train the surrogate model gθ on related training data D so that the trained surrogate model gθ yields outputs that are correlated with outputs of f: We formally have the correlation ρ(f,h) between the physics model outputs and the data source h and the correlation ρ(h,gθ) between the data source h and the surrogate model gθ. The following proposition provides upper and lower bounds on the correlation coefficient ρ(f,gθ) between the physics model and the surrogate model. A proof of the proposition can be found in the Supplementary material.

Proposition 1.The correlation coefficient ρ(f,gθ) between outputs of the physics model f and the surrogate model gθ is bounded from above and below as

ρ(f,gθ)ρ(f,h)ρ(h,gθ)+(1ρ(f,h)2)(1ρ(h,gθ)2),ρ(f,gθ)ρ(f,h)ρ(h,gθ)(1ρ(f,h)2)(1ρ(h,gθ)2).

The upper bound is maximized when ρ(f,h) and ρ(h,gθ) are balanced, which indicates that if f and h are poorly correlated, then it is unnecessary to accurately train gθ to match well the correlated data sampled from h. In Fig. 2, we plot the upper bound of ρ(f,gθ) against ρ(f,h) and ρ(h,gθ), which visualizes that the upper bound is high when ρ(f,h) and ρ(h,gθ) are of comparable magnitude, i.e. when they are balanced. With regards to the lower bound, which is also shown in Fig. 2, note that the squared correlation coefficient ρ(f,gθ)2 enters in the MSE in Eq. (5) rather than directly ρ(f,gθ). The lower bound stated in Proposition 1 shows that the squared correlation coefficient ρ(f,gθ)2>0 is greater than zero if

Fig. 2.

Fig. 2.

The plots visualize the upper and lower bound (Proposition 1) of the correlation between the surrogate model gθ and the physics model f when the surrogate model gθ is trained on a related data source h that is correlated to f. The upper and lower bounds depend on the correlation between the data source h and the physics model f (“quality of the data”) as well as correlation between the data source h and the surrogate model gθ (“quality of training”).

ρ(f,h)2+ρ(h,gθ)2>1 (6)

holds, see the Supplementary material for a detailed derivation. This result implies that both the high-fidelity model f and the data source h as well as the data source h and the surrogate model gθ need to be sufficiently well correlated so that the correlation between the high-fidelity model and the surrogate model is not zero. We will show in the numerical experiments in Fig. 4 that this condition is met with a large margin in all of our experiments. Further visualizations of the bounds and the condition given in Eq. (6) can be found in the Supplementary material.

Fig. 4.

Fig. 4.

The estimated correlation coefficients show that the physics model is well correlated to the surrogate models trained on the related training data. The shaded area corresponds to the upper and lower bound of the correlation coefficient ρ(f,gθ), see Proposition 1. a) Input distribution given in Eq. (7). b) Input distribution given in Eq. (8).

Critically, as long as the squared correlation coefficient between gθ and f is high, the surrogate model gθ is useful in multifidelity estimation as with the multifidelity Monte Carlo estimator given in Eq. (2).

Surrogate models from related data for Tubbs 2017 wildfire scenarios

Coupled fire–atmosphere model with about eight billion degrees of freedom

The wildfire scenarios that we consider are motivated by the Tubbs wildfire that took place in Northern California in October 2017 (50, 51). We simulate the spread of a wildfire on a 20 km by 20 km area with a terrain that is representative of the mountain area between Calistoga and Santa Rosa, see Fig. 3. The fire is simulated up to heights of 4 km to accurately capture atmospheric effects. The domain is resolved with 20 m in horizontal and 4 m in vertical direction, which corresponds to 1,024×1,024×1,024 grid points and leads to about eight billion degrees of freedom because the physics model is formulated over density, velocity in the three spatial directions, potential temperature, oxygen mass fraction, solid temperature, and fuel density. The physics model is based on the Navier–Stokes equations in a low-Mach formulation (11), which eliminates the constraints on the time step size by the acoustic waves in the numerical simulation. A multiphase combustion model is used to represent the fire behavior, and it is fully coupled with the fluid dynamics of the atmosphere to capture the fire–atmosphere interactions (12). We remark that the model has been validated with other, controlled fire scenarios in (7). Details about the model can be found in the Supplementary material.

Fig. 3.

Fig. 3.

The high-fidelity physics model as well as the simplified model predict that increasing moisture density leads to smaller burned areas, which indicates that the simplified model captures the trends of the high-fidelity physics model and thus is sufficient for generating related training data in this example. a) Burned area in Tubbs domain for three input realizations with moisture densities 0.225, 0.136, and 0.029. The surface plots show the terrain height using a color gradient and the burned area in black. b) Comparison of burned areas over 20 different input realizations (environmental configurations) at different times after ignition between the large-scale simulation and the surrogate model trained on related data.

The input vector X to the simulation contains the wind speed, initial fuel density, and initial moisture content, which are key environmental conditions that influence the fire dynamics. The components of the input vector X are considered uncertain and distributed uniformly in the intervals [5.0,12.0], [0.2,3.0], and [0.03,0.12], respectively so that

XU([5,12]×[0.2,3]×[0.03,0.12]), (7)

where U denotes a uniform distribution. The output is the burned area over time after ignition. The burned area is computed via the fuel density field to identify areas where fuel density has decreased, which indicates burning; details about computing the burned area can be found in Supplementary material. For a single realization of the random input vector, the simulation (one evaluation of f) takes about 19.5 h on 128 Tensor Processing Unit (TPU) v5e cores of Google Cloud. Even when performing ten simulations in parallel, generating 1,000 direct data points for training surrogate models would take almost 3 months on 1,280 TPUs, which is intractable. In our experiments below, we have available up to 20 simulation results from the high-fidelity physics model, which is too little for training surrogate models. These 20 simulations are also insufficient for estimating moments with quadrature rules on grids, because even only five grid points in each dimension would already require >120 simulations in our case and thus six times the amount of simulations that we have available.

Wildfire simulations: Related data source

For training surrogate models, we generate related data by using a small-scale simulation, in which the domain is scaled (shrunken) down by a factor of 10 proportionally. Correspondingly, we require only 256×256×160 spatial grid points so that one simulation takes on average 28 min on eight TPU cores. Generating 1,000 related training data samples takes 2.92 h with 10 simulations in parallel and each on 128 TPU cores. Thus, the runtime of generating the training data are reduced from 3 months to <3 h. We train as surrogate model gθ a multilayer perceptron with three input nodes (fuel density, moisture density, and wind speed), three hidden layers of width five, and a linear output layer on the related data to obtain outputs that are correlated with the burned area predicted by the physics model; details of the training setup can be found in the Supplementary material.

To demonstrate the correlation between the high-fidelity physics and the simplified model, we plot in Fig. 3a the burned area with respect to a decreasing moisture density. The absolute value of the burned area given by the physics model and the simplified model differ by >30%, which indicates a large bias in the sense of Eq. (3). However, with both models we obtain that the burned area increases with decreasing moisture density, which indicates that the outputs of the physics-based and simplified model are correlated and the simplified model is able to reproduce a physically meaningful behavior. Figure 3b shows the correlation more directly by plotting the outputs of the surrogate model obtained from training data generated with the simplified model against the predictions of the physics model over time t when varying the input components. The estimated correlation coefficient given in Eq. (4) is shown in Fig. 4a, with the mean value of the correlation coefficient being 0.8335 over time t.

Wildfire simulations: Performance

We now use the surrogate model trained on related data in the multifidelity estimator given in Eq. (2) to estimate the expected burned area. We compare three cases. First, we consider the physics model alone in a regular Monte Carlo estimator given in Eq. (1), which is expensive but leads to an unbiased estimator. Second, we use the surrogate model alone in a regular Monte Carlo estimator, which leads to a biased estimator because the surrogate model has a bias with respect to the physics model. Third, we leverage the surrogate model in a multifidelity Monte Carlo estimator given in Eq. (2) together with the physics model, which provides an unbiased estimator. We stress that the multifidelity estimator of the burned area is unbiased even though the surrogate model is used and has been trained on related data rather than on direct data.

The estimated expected burned area is shown in Fig. 5a, together with the estimated root mean squared error (RMSE) shown as shaded area. Notice the bias obtained when the surrogate model is used alone, which is in agreement with the discussion above that a surrogate model trained on related data is not predictive in the sense that the surrogate model outputs can have a large bias in the sense of Eq. (3) and thus the surrogate model cannot replace the physics models. However, the surrogate model is still useful, namely when it is combined with the physics model within the multifidelity estimator, where it introduces no bias. As the results in Fig. 5a show, combining the surrogate model and the physics models leads to estimates that predict a burned area of about 2.4km2 at already about 10,000 TPU hours at time t=600 s, whereas the estimator that uses the physics model alone takes almost 50,000 TPU hours to converge to a comparable burned area. Similar observations hold at later times after ignition. Figure 5b shows the estimated RMSEs of the estimators as a bar plot. We stress that the estimated RMSEs shown in Fig. 5b and the estimated RMSEs shown as shaded area in Fig. 5a are affected by estimation errors and we thus use them as crude indicators only to see if they are in agreement with the other results, see Supplementary material for how we estimate the RMSEs. In agreement with the low variance of the curve corresponding to the estimator obtained with the surrogate model together with the physics model shown in Fig. 5a, the estimated RMSE indicates that using the surrogate model in the multifidelity estimator leads to a lower RMSE than using the samples from the physics model alone. The numbers of samples used from the physics model versus the correlated surrogate model at time t=2,400 s are shown in Fig. 5c.

Fig. 5.

Fig. 5.

Scenario with input distribution given in Eq. (7). a) Using the surrogate models trained on related data together with the physics models leads to unbiased multifidelity estimators of the expected burned area that exhibit less variance with increasing computational budget (TPU hours) than using the physics model alone. Including the surrogate model from related data achieves accurate estimates of the expected burned area at already around 10,000 TPU hours, whereas using the physics model alone requires up to almost 50,000 TPU hours to achieve a comparable expected burned area. For a comparison at additional times after ignition; see Fig. S3a. b) Estimates of the RMSEs are in agreement with the previous results and indicate that including the surrogate models trained on related data leads to almost 2× more accurate estimates of the expected burned area compared with using the physics model alone. c) The plot shows the number of samples used from the physics and the surrogate model when the two models are combined by the multifidelity estimator.

Wildfire simulations: Performance with reused surrogate model

It is common to train a surrogate model on an input distribution with a larger variance so that it can be reused on input distributions with a lower variance in scenarios where more information is available. To illustrate this, we consider a smaller range of wind speeds and fuel densities so that

XU([7,10.25]×[0.9,2.3]×[0.03,0.12]). (8)

We reuse the surrogate model from the previous scenario and demonstrate now that it still leads to an efficient multifidelity estimator even over this changed input distribution because it is still correlated with the physics model, see Fig. 4b. The predictions of the estimators are shown in Fig. 6. The lower variance of the input distribution leads to a higher correlation of the surrogate model to the physics model, which results in a multifidelity estimator with the surrogate model that shows less variation over the budget and thus settles more quickly on a value for the estimated expected burned area compared with using the physics model alone.

Fig. 6.

Fig. 6.

Scenario with input distribution given in Eq. (8): Already for low computational budgets of about 10,000 TPU hours, the surrogate model trained on related data together with the physics model lead to a multifidelity estimator of the expected burned area that shows little variance as the budget of TPU hours is increased. This indicates that including the surrogate model from related data provides more accurate estimates than using the physics model alone, because the surrogate model is used via a multifidelity estimator so that it introduces no bias. The estimated RMSEs are in agreement with these results and indicate an almost 3× lower RMSE than the estimator that uses the physics model alone. For a comparison at additional times after ignition, see Fig. S3b.

Discussion

In this study, we demonstrate that training surrogate models on related data and using them together with multifidelity estimators allows scaling uncertainty quantification to estimate expectations of quantities of interest from ensembles of large-scale wildfire simulations with billions of degrees of freedom. Our approach captures more accurately the burned area than using either surrogate models or physics models alone for the same computational costs. This study further demonstrates that viewing surrogate modeling through a broader lens than just aiming for accurate point-wise predictions of the outputs of physics models can greatly extend the scope of surrogate modeling. In particular, the results of this study show that it is sufficient for surrogate models to yield outputs that are statistically correlated with the outputs of physics models with respect to the Pearson moment correlation coefficient; it is unnecessary that the surrogate models are point-wise predictive about the quantities of interest in terms of having a low bias. This allows training surrogate models on related data rather than on direct data that directly describe the input–output relationship given by the physics model. Learning from related data broadens the scope of data-driven surrogate modeling to settings where direct data are scarce, as demonstrated by our application to wildfire simulations.

The focus of this study is on uncertainty quantification, for which surrogate models trained on related data are useful in multifidelity computations such as multifidelity Monte Carlo methods. However, we expect surrogate models that provide correlated outputs to be useful beyond our specific setting of uncertainty quantification such as performing sensitivity analyses with multifidelity estimators (49) and estimating objective functions in optimal design problems under uncertainty, inverse problems, and control. Overall, the findings of this study encourage the perspective on surrogate modeling that shifts the focus away from the classical aim of directly mimicking physics-based simulations towards being another information source in multifidelity computations (20).

Supplementary Material

pgae554_Supplementary_Data

Contributor Information

Paul Schwerdtner, Courant Institute of Mathematical Sciences, New York University, 251 Mercer Street, New York, NY 10012, USA.

Frederick Law, Courant Institute of Mathematical Sciences, New York University, 251 Mercer Street, New York, NY 10012, USA.

Qing Wang, Google Research, Mountain View, CA 94043, USA.

Cenk Gazen, Google Research, Mountain View, CA 94043, USA.

Yi-Fan Chen, Google Research, Mountain View, CA 94043, USA.

Matthias Ihme, Google Research, Mountain View, CA 94043, USA; Department of Mechanical Engineering, Stanford University, Stanford, CA 94305, USA.

Benjamin Peherstorfer, Courant Institute of Mathematical Sciences, New York University, 251 Mercer Street, New York, NY 10012, USA.

Supplementary Material

Supplementary materialis available at PNAS Nexus online.

Funding

This material is based upon work supported by the Google Research Collab on “Uncertainty quantification for wildfire simulations: Reaching operational scale with Google’s ML stack and multifidelity methods,” the Air Force Office of Scientific Research under Award Number FA9550-21-1-0222 (Dr. Fariba Fahroo) and the U.S. Department of Energy, Office of Science Energy Earthshot Initiative as part of the project “Learning reduced models under extreme data conditions for design and rapid decision-making in complex systems” under Award no. DE-SC0024721.

Author Contributions

Paul Schwerdtner (Methodology, Software, Validation, Formal Analysis, Investigation, Writing—original draft, Writing—review & editing, Visualization), Frederick Law (Conceptualization, Data curation, Writing—review & editing), Qing Wang (Methodology, Conceptualization, Software, Data curation, Writing—original draft, Writing—review & editing, Visualization) Cenk Gazen (Methodology, Conceptualization, Software, Data curation, Writing—review & editing), Yi-fan Chen (Methodology, Conceptualization, Software, Data curation, Writing—review & editing), Matthias Ihme (Methodology, Conceptualization, Software, Writing—review & editing), Benjamin Peherstorfer (Conceptualization, Methodology, Investigation, Writing—original draft, Writing—review & editing).

Data Availability

Original data created for the study are or will be available in a persistent repository upon publication. The type of data is code and is available via a Zenodo repository https://doi.org/10.5281/zenodo.11391078.

References

  • 1. Abatzoglou  JT, Williams  AP. 2016. Impact of anthropogenic climate change on wildfire across western US forests. Proc Natl Acad Sci U S A. 113(42):11770–11775. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Wang  D, et al.  2021. Economic footprint of california wildfires in 2018. Nat Sustain. 4(3):252–260. [Google Scholar]
  • 3. Bakhshaii  A, Johnson  EA. 2019. A review of a new generation of wildfire-atmosphere modeling. Can J For Res. 49(6):565–574. [Google Scholar]
  • 4. Liu  Y, et al.  2019. Fire behavior and smoke modeling: model improvement and measurement needs for next-generation smoke research and forecasting systems. Int J Wildland Fire. 28(8):570. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Coen  JL, Schroeder  W. 2013. Use of spatially refined satellite remote sensing fire detection data to initialize and evaluate coupled weather-wildfire growth model simulations. Geophys Res Lett. 40(20):5536–5541. [Google Scholar]
  • 6. Coen  JL, Schroeder  W, Conway  S, Tarnay  L. 2020. Computational modeling of extreme wildland fire events: a synthesis of scientific understanding with applications to forecasting, land management, and firefighter safety. J Comput Sci. 45:101152. [Google Scholar]
  • 7. Wang  Q, Ihme  M, Gazen  C, Chen  Y-F, Anderson  J. 2024. A high-fidelity ensemble simulation framework for interrogating wildland-fire behaviour and benchmarking machine learning models. Int J Wildland Fire. 33:WF24097. [Google Scholar]
  • 8. Willcox  K, Ghattas  O, Heimbach  P. 2021. The imperative of physics-based modeling and inverse theory in computational science. Nat Comput Sci. 1(3):166–168. [DOI] [PubMed] [Google Scholar]
  • 9. Coen  JL, et al.  2013. Coupled weather-wildland fire modeling with the weather research and forecasting model. J Appl Meteorol Climatol. 52(1):16–38. [Google Scholar]
  • 10. Linn  RR, Cunningham  P. 2005. Numerical simulations of grass fires using a coupled atmosphere-fire model: basic fire behavior and dependence on wind speed. J Geophys Res. 110(D13):D13107. [Google Scholar]
  • 11. Wang  Q, Ihme  M, Chen  Y-F, Anderson  J. 2022. A tensorflow simulation framework for scientific computing of fluid flows on tensor processing units. Comput Phys Commun. 274:108292. [Google Scholar]
  • 12. Wang  Q, et al.  2023. A high-resolution large-eddy simulation framework for wildland fire predictions using tensorflow. Int J Wildland Fire. 32(12):1711–1725. [Google Scholar]
  • 13. Westerling  AL. 2016. Increasing western US forest wildfire activity: sensitivity to changes in the timing of spring. Philos Trans R Soc Lond B Biol Sci. 371(1696):20150178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Cruz  MG. 2010. Monte Carlo-based ensemble method for prediction of grassland fire spread. Int J Wildland Fire. 19(4):521–530. [Google Scholar]
  • 15. Peherstorfer  B, Willcox  K, Gunzburger  M. 2016. Optimal model management for multifidelity Monte Carlo estimation. SIAM J Sci Comput. 38(5):A3163–A3194. [Google Scholar]
  • 16. Benner  P, Gugercin  S, Willcox  K. 2015. A survey of projection-based model reduction methods for parametric dynamical systems. SIAM Rev. 57(4):483–531. [Google Scholar]
  • 17. Kovachki  NB, et al.  2023. Neural operator: learning maps between function spaces with applications to PDEs. J Mach Learn. 24(89):1–97. [Google Scholar]
  • 18. Kramer  B, Peherstorfer  B, Willcox  K. 2024. Learning nonlinear reduced models from data with operator inference. Annu Rev Fluid Mech. 56(1):521–548. [Google Scholar]
  • 19. Rozza  G, Huynh  DBP, Patera  A. 2007. Reduced basis approximation and a posteriori error estimation for affinely parametrized elliptic coercive partial differential equations. Arch Comput Methods Eng. 15(3):1–47. [Google Scholar]
  • 20. Peherstorfer  B, Willcox  K, Gunzburger  M. 2018. Survey of multifidelity methods in uncertainty propagation, inference, and optimization. SIAM Rev. 60(3):550–591. [Google Scholar]
  • 21. Robert  C, Casella  G. 2004. Monte Carlo statistical methods. Springer. [Google Scholar]
  • 22. Valero  MM, Jofre  L, Torres  R. 2021. Multifidelity prediction in wildfire spread simulation: modeling, uncertainty quantification and sensitivity analysis. Environ Model Softw. 141:105050. [Google Scholar]
  • 23. Brunton  SL, Proctor  JL, Kutz  JN. 2016. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proc Natl Acad Sci U S A. 113(15):3932–3937. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. de Avila Belbute-Peres  F, Smith  K, Allen  K, Tenenbaum  J, Kolter  JZ. 2018. End-to-end differentiable physics for learning and control. In: Bengio  S, Wallach  H, Larochelle  H, Grauman  K, Cesa-Bianchi  N, Garnett  R, editors. Advances in neural information processing systems. Vol. 31. Curran Associates, Inc. [Google Scholar]
  • 25. Fresca  S, Dede’  L, Manzoni  A. 2021. A comprehensive deep learning-based approach to reduced order modeling of nonlinear time-dependent parametrized PDEs. J Sci Comput. 87(2):61. [Google Scholar]
  • 26. Ghattas  O, Willcox  K. 2021. Learning physics-based models from data: perspectives from inverse problems and model reduction. Acta Numer. 30:445–554. [Google Scholar]
  • 27. Kabacaoğlu  G, Biros  G. 2019. Machine learning acceleration of simulations of Stokesian suspensions. Phys Rev E. 99(6):063313. [DOI] [PubMed] [Google Scholar]
  • 28. Kim  Y, Choi  Y, Widemann  D, Zohdi  T. 2022. A fast and accurate physics-informed neural network reduced order model with shallow masked autoencoder. J Comput Phys. 451:110841. [Google Scholar]
  • 29. Kochkov  D, et al.  2021. Machine learning-accelerated computational fluid dynamics. Proc Natl Acad Sci U S A. 118(21):e2101784118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Kutz  JN, Brunton  SL, Brunton  BW, Proctor  JL. 2016. Dynamic mode decomposition: data-driven modeling of complex systems. SIAM. [Google Scholar]
  • 31. Qian  E, Kramer  B, Peherstorfer  B, Willcox  K. 2020. Lift & learn: physics-informed machine learning for large-scale nonlinear dynamical systems. Physica D. 406:132401. [Google Scholar]
  • 32. Swischuk  R, Mainini  L, Peherstorfer  B, Willcox  K. 2019. Projection-based model reduction: formulations for physics-based machine learning. Comput Fluids. 179:704–717. [Google Scholar]
  • 33. Vlachas  PR, Arampatzis  G, Uhler  C, Koumoutsakos  P. 2022. Multiscale simulations of complex systems by learning their effective dynamics. Nat Mach Intell. 4(4):359–366. [Google Scholar]
  • 34. Vlachas  PR, Byeon  W, Wan  ZY, Sapsis  TP, Koumoutsakos  P. 2018. Data-driven forecasting of high-dimensional chaotic systems with long short-term memory networks. Proc Math Phys Eng Sci. 474(2213):20170844. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Conti  P, Guo  M, Manzoni  A, Hesthaven  JS. 2023. Multi-fidelity surrogate modeling using long short-term memory networks. Comput Methods Appl Mech Engrg. 404:115811. [Google Scholar]
  • 36. Guo  M, Manzoni  A, Amendt  M, Conti  P, Hesthaven  JS. 2022. Multi-fidelity regression using artificial neural networks: efficient approximation of parameter-dependent output quantities. Comput Methods Appl Mech Engrg. 389:114378. [Google Scholar]
  • 37. Kennedy  MC, O’Hagan  A. 2000. Predicting the output from a complex computer code when fast approximations are available. Biometrika. 87(1):1–13. [Google Scholar]
  • 38. Parussini  L, Venturi  D, Perdikaris  P, Karniadakis  GE. 2017. Multi-fidelity Gaussian process regression for prediction of random fields. J Comput Phys. 336:36–50. [Google Scholar]
  • 39. Poloczek  M, Wang  J, Frazier  P. 2017. Multi-information source optimization. In: Guyon  I, Von Luxburg  U, Bengio  S, Wallach  H, Fergus  R, Vishwanathan  S, Garnett  R, editors. Advances in neural information processing systems. Vol. 30. Curran Associates, Inc. [Google Scholar]
  • 40. Raissi  M, Perdikaris  P, Karniadakis  GE. 2017. Inferring solutions of differential equations using noisy multi-fidelity data. J Comput Phys. 335:736–746. [Google Scholar]
  • 41. Song  DH, Tartakovsky  DM. 2022. Transfer learning on multifidelity data. J Mach Learn Modeling Comput. 3(1):31–47. [Google Scholar]
  • 42. Finn  C, Abbeel  P, Levine  S. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning - Volume 70, ICML’17, JMLR.org. p. 1126–1135.
  • 43. Thrun  S, Pratt  L. 1998. Learning to learn: introduction and overview. Boston (MA): Springer US. p. 3–17. [Google Scholar]
  • 44. Farcaş  I-G, Peherstorfer  B, Neckel  T, Jenko  F, Bungartz  H-J. 2023. Context-aware learning of hierarchies of low-fidelity models for multi-fidelity uncertainty quantification. Comput Methods Appl Mech Eng. 406:115908. [Google Scholar]
  • 45. Peherstorfer  B. 2019. Multifidelity Monte Carlo estimation with adaptive low-fidelity models. SIAM/ASA J Uncertain Quantif. 7(2):579–603. [Google Scholar]
  • 46. Farcaş  I-G, Merlo  G, Jenko  F. 2022. A general framework for quantifying uncertainty at scale. Commun Eng. 1(1):43. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Konrad  J, et al.  2022. Data-driven low-fidelity models for multi-fidelity Monte Carlo sampling in plasma micro-turbulence analysis. J Comput Phys. 451:110898. [Google Scholar]
  • 48. Nelson  BL. 1987. On control variate estimators. Comput Oper Res. 14(3):219–225. [Google Scholar]
  • 49. Qian  E, Peherstorfer  B, O’Malley  D, Vesselinov  VV, Willcox  K. 2018. Multifidelity Monte Carlo estimation of variance and sensitivity indices. SIAM/ASA J Uncertain Quantif. 6(2):683–706. [Google Scholar]
  • 50. Martinez  J, et al.  2017. Incident report. Technical Report 17CALNUO10045, California Department of Forestry and Fire Protection.
  • 51. Wang  Q, et al.  2022. Towards real-time predictions of large-scale wildfire scenarios using a fully coupled atmosphere-fire physical modelling framework. In: Viegas D, Ribeiro L, editors. Advances in forest fire research. p. 415–421. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

pgae554_Supplementary_Data

Data Availability Statement

Original data created for the study are or will be available in a persistent repository upon publication. The type of data is code and is available via a Zenodo repository https://doi.org/10.5281/zenodo.11391078.


Articles from PNAS Nexus are provided here courtesy of Oxford University Press

RESOURCES