Improved treatment of uncertainty in hydrologic modeling: Combining the strengths of global optimization and data assimilation
Abstract
[1] Hydrologic models use relatively simple mathematical equations to conceptualize and aggregate the complex, spatially distributed, and highly interrelated water, energy, and vegetation processes in a watershed. A consequence of process aggregation is that the model parameters often do not represent directly measurable entities and must therefore be estimated using measurements of the system inputs and outputs. During this process, known as model calibration, the parameters are adjusted so that the behavior of the model approximates, as closely and consistently as possible, the observed response of the hydrologic system over some historical period of time. In practice, however, because of errors in the model structure and the input (forcing) and output data, this has proven to be difficult, leading to considerable uncertainty in the model predictions. This paper surveys the limitations of current model calibration methodologies, which treat the uncertainty in the input-output relationship as being primarily attributable to uncertainty in the parameters and presents a simultaneous optimization and data assimilation (SODA) method, which improves the treatment of uncertainty in hydrologic modeling. The usefulness and applicability of SODA is demonstrated by means of a pilot study using data from the Leaf River watershed in Mississippi and a simple hydrologic model with typical conceptual components.
1. Introduction and Scope
[2] Hydrologic models often contain parameters that cannot be measured directly, but can only be meaningfully inferred by calibration to a historical record of input-output data. During model calibration, the parameters are adjusted in such a way that the behavior of the model approximates, as closely and consistently as possible, the observed response of the hydrologic system over some historical period of time. Because of the time consuming nature of manual trial-and-error model calibration, there has been a great deal of research into the development of automated (computer based) calibration methods [see, e.g., Duan et al., 1992; Gupta and Sorooshian, 1994; Gupta et al., 1998; Yapo et al., 1998; Boyle et al., 2000]. Automatic methods for model calibration seek to take advantage of the speed and power of computers, while being relatively objective and easier to implement than manual methods.
[3] While considerable progress has been made in the development and application of automated calibration methods, many of these approaches treat the underlying uncertainty in the input-output representation of the model as being primarily (and explicitly) due to uncertainty in the parameter estimates. A few papers do discuss the treatment of input, state, and model structural uncertainties for environmental models [e.g., Young and Beck, 1974; Beck and Young, 1975; Kitanidis and Bras, 1980a, 1980b; Beck, 1985, 1987; Hebson and Wood, 1985; Young, 1986], but such approaches have not become common practice for nonlinear watershed models. Clearly, uncertainties in the modeling procedure stem not only from uncertainties in the parameter estimates, but also from measurement errors associated with the system input (forcing) and output, and from model structural errors arising from the aggregation of spatially distributed real-world processes into a mathematical model. Not properly accounting for these errors during model calibration, can result in model simulations and associated prediction uncertainty bounds, which do not consistently represent and bracket the measured system behavior. This is evidenced by residuals which exhibit considerable variation in bias (nonstationarity), variance (heteroscedasticity), and correlation structures under different hydrologic conditions. Several contributions to the hydrologic literature have therefore brought into question the continued usefulness of the classical paradigm for estimating model parameters (see, e.g., Beven and Binley [1992], Gupta et al. [1998], Kavetski et al. [2003], and the papers referenced earlier in this paragraph), especially in the face of the emerging generation of spatially distributed multi-input-output hydrologic models for which multiple (and often conflicting) sources of information are available for model calibration.
[4] Some interesting methods for addressing these problems, particularly in the context of estimating reasonable confidence bands on model simulations for conceptual nonlinear watershed models, have begun to appear in the literature. These methods include the use of classical Bayesian [Kuczera and Parent, 1998; Thiemann et al., 2001; Vrugt et al., 2003a; Kavetski et al., 2003], pseudo-Bayesian [Beven and Binley, 1992; Freer et al., 1996], set theoretic [Keesman, 1990; Klepper et al., 1991; van Straten and Keesman, 1991; Vrugt et al., 2001], and multiple criteria [Gupta et al., 1998; Yapo et al., 1998; Boyle et al., 2000; Vrugt et al., 2003b] methods to represent model parameter and prediction uncertainty. Each of these methods, however, uses one or more aggregate statistics of model performance over a large range of hydrologic behaviors, an action that is now understood to result in considerable loss of important information that can be used to distinguish between competing parameter sets. To increase the discriminative power of model calibration strategies, interest has therefore recently begun to switch toward recursive model identification strategies [e.g., Thiemann et al., 2001; Young, 2001; Vrugt et al., 2002; Wagener et al., 2003], which sequentially move through a time series of discharge data and can provide estimates of parameter uncertainty. Such approaches can be used to localize the most informative measurements for parameter estimation [see Vrugt et al., 2002], and also, perhaps more importantly, to provide a method for checking for violations of the underlying assumption that parameters are constant [Wagener et al., 2003; Misirli, 2003; Vrugt et al., 2003c]. Despite progress made, such methods still assign the uncertainty in the input-output representation primarily to a combination of uncertainty in the parameter estimates plus a residual “model” error term [see Thiemann et al., 2001; Gupta et al., 2003], and therefore lack the conceptual rigor needed to distinguish between all the important sources of uncertainty.
[5] In a separate line of research, considerable progress has been made in the development and application of sequential data assimilation (SDA) techniques. Such methods provide a general framework for explicitly dealing with input, output and model structural uncertainty, and for optimal merging of uncertain model predictions with observations. In contrast to classical model calibration strategies, SDA methods continuously update the states in the model when new measurements become available to improve the model forecast and evaluate the forecast accuracy. The prototype of the SDA methods, the Kalman filter (KF), was developed in the 1960s for optimal control of systems governed by linear equations and (as referenced earlier) was introduced into hydrology during the early 1980s to improve real-time forecasting of river discharges and to recursively estimate model parameters. Note that, for nonlinear dynamics, the extended Kalman filter (EKF) can in principle be used, which linearizes the error covariance equation using a tangent linear operator. However, because third- and higher-order moments in the closure scheme are discarded, this linearization is notoriously unstable if the nonlinearities are strong [Evensen, 1992; Miller et al., 1994].
[6] Although the KF and EKF offer a very general framework for accounting for all sources of uncertainty, the filter typically assumes that the optimal values of the model parameters are known prior to the modeling procedure. This is particularly difficult in hydrologic modeling, where parameters often represent conceptual properties whose values can generally not be independently measured or assessed from lookup tables. This uncertainty in the correct choice of the parameter values, results in uncertainty in the estimated state values and therefore the model output prediction. To extend the usefulness and applicability of SDA methods to hydrologic modeling, it is important to explicate a formal methodology, which not only recursively updates model states but simultaneously also estimates model parameters. The major difference between such a framework and the aforementioned classical model calibration strategies is that this approach no longer assigns the primary uncertainty in the input-output representation to uncertainty in the parameters, but explicitly accounts for input, output and model structural errors during model calibration.
[7] This paper presents a simultaneous parameter optimization and data assimilation (SODA) method, which combines the strengths of the parameter search efficiency and explorative capabilities of the Shuffled Complex Evolution Metropolis (SCEM-UA) algorithm [Vrugt et al., 2003a], with the power and computational efficiency of the ensemble Kalman filter [Evensen, 1994] to provide a better treatment of the input, output, parameter and model structural uncertainties in hydrologic modeling. The implementation of SODA requires a change in philosophy because it means that the model calibration procedure can no longer be considered as finding a single set of parameter values which generate the best possible long-term model forecasts. Instead, the optimal set of parameters identified with SODA only has meaning in combination with a SDA method, and consequently, will (per definition) not generate the best possible forecasts when no state adjustments are allowed, which is typically the case when evaluating the predictive capabilities of the model during an independent evaluation period. We hypothesize that the implementation of SODA will result in (1) meaningful prediction uncertainty bounds on the model simulations, (2) parameter estimates which better represent system properties that are less corrupted by modeling errors, thereby increasing the prospects of finding useful regionalization relationships, and (3) a time series of recursive state and output adjustments, whose interpretation will generate inspiration to improve our model concepts and as such our understanding of the functioning of hydrologic systems.
[8] This paper is organized as follows. Section 2 surveys the limitations of current model calibration strategies, which do not explicitly treat input, output and model structural uncertainty, but assign the uncertainty in the input-output representation primarily to uncertainty in the parameter estimates. In section 3 and 4 we discuss the rationale and architecture of SDA and global optimization methods, and demonstrate how the strengths of these methods can be combined to derive the hybrid SODA framework. In section 5 we illustrate the power and applicability of SODA by means of a standard mathematical study using the three-parameter highly nonlinear Lorenz model [Lorenz, 1963], and a classical rainfall-runoff modeling example. In this section we also introduce a nonparametric variance estimator, which is especially designed to estimate the size of the measurement error of the output data. Finally, in section 6 we summarize the results. Note that although the SODA strategy results in considerable increase in computational burden, this increase is quite easily handled using currently available computing capabilities. The simple watershed model computation performed in this paper required approximately 4 hours on a Pentium 4 desktop personal computer. Application to more complex hydrologic models may, of course, possibly require taking advantage of parallel computing facilities.
2. Basic Inverse Formulation
[11] The hydrologic model (HYMOD) conceptual watershed model was calibrated using the Shuffled Complex Evolution (SCE-UA) global optimization algorithm developed by Duan et al. [1992] using the SLS criterion and 1 year of calibration data. Notice, that the fit to the observed data can be considered quite good, but that the residuals exhibit considerable variation in bias (nonstationarity), variance (heteroscedasticity) and correlation structure under different hydrologic conditions. Indeed, the common approach of dealing with model structural and data errors, as being “small” or somehow “absorbed” into the output error residual needs serious reconsideration. Moreover, given the presence of these errors, overconditioning of the model to a single parameter set is unreasonable, and cannot be justified.
[12] One set of responses set forth to directly address the problem of overconditioning is to abandon the Frequentists approach of believing that the model parameters in equation (1) are fixed but unknown, and to adopt a Bayesian viewpoint which allows the identification of a plausible set of values for the parameters of the model given the available data. The Bayesian approach treats the model parameters in equation (1) as probabilistic variables having a joint posterior probability density function (pdf), which summarizes our probabilistic belief about the parameters in the light of the observed data . Examples of Bayesian approaches to hydrology include the Generalized Likelihood Uncertainty Estimation (GLUE) framework of Beven and Binley [1992] and the Bayesian Recursive Estimation (BaRE) approach of Thiemann et al. [2001] for representing model parameter and prediction uncertainty within the context of Monte Carlo analysis, and the Shuffled Complex Evolution Metropolis (SCEM-UA) global optimization algorithm of Vrugt et al. [2003a] for simultaneously estimating the traditional “best” parameter set along with a sample set of parameter values describing the probabilistic representation of the remaining parameter uncertainty (which is then used to generate probabilistic forecasts). We note that the SCEM-UA and BaRE methods operate within the context of the classical Bayesian framework (which makes implicit assumptions about the correctness or adequacy of the model of the data generating process), while GLUE requires the user to make other kinds of subjective decisions including the selection of a cutoff threshold that separates behavioral from nonbehavioral parameter sets.
[13] An alternative set of responses set forth to identify model structural inadequacies, is to use Recursive Parameter Estimation algorithms, which provide better extraction of the information in the data, because the temporal aggregation associated with traditional batch processing is reduced. Examples of recursive algorithms which can be applied to complex nonlinear conceptual hydrologic models, include the PIMLI and recursive SCEM-UA algorithms [Vrugt et al., 2002, 2003c], the DYNIA approach [Wagener et al., 2003], and the BaRE algorithm [Thiemann et al., 2001]. While these recursive methods have considerable advantages, their usefulness to help diagnose and quantify model structural errors has yet to be satisfactorily demonstrated. The novel instrumental variable techniques developed by Young [2001, 2002a, 2002b] does address this problem for the case of transfer function model structures by diagnosing time-varying parameters and showing how this information can be used to update the underlying model structure. However, this approach effectively neglects input and parameter uncertainty.
[14] A third set of responses set forth to more directly confront the problem of model structural errors is to pose the model calibration problem into a multiobjective framework, as was advocated by Gupta et al. [1998]. A major consequence of model structural imperfection is that the model is incapable of reproducing all aspects and portions of the output data with a single parameter set. By employing a number of complementary criteria in the optimization procedure, and analyzing the tradeoffs in the fitting of these criteria, the hydrologist is able to better understand the limitations of current hydrologic model structures, and gain insights into possible model improvements. The resulting Pareto solution set of parameters defines a kind of parameter uncertainty attributable to model structural errors. Yapo et al. [1998] and later Vrugt et al. [2003b] presented the MOCOM and MOSCEM-UA algorithms, respectively, capable of solving, in a single optimization run, the multiobjective problem posed by this approach. Unlike the previous methods, the multiobjective approach explicitly acknowledges the existence of model structural errors during the parameter estimation procedure. However, because of subjectivity in the selection of the measures [Gupta et al., 1998] and data subsets [see Boyle et al., 2000] to be included in the multiobjective vector, considerable fuzziness in the specification of the Pareto optimal parameter space and thus related model structural errors exists. Moreover, multiobjective calibration is typically response focused, and does not explicitly consider the influence of input errors [see Kavetski et al., 2003]. Like the previous methods all sources of uncertainty are treated primarily as parameter uncertainty, giving rise to the concept of Pareto optimality.
[15] Finally, the last set of responses set forth to more directly confront the problem of input errors is to define realistic stochastic input error models. One novel approach is the recently proposed Bayesian Total Error Analysis (BATEA) of Kavetski et al. [2003], which explicitly takes into account input errors in the development of the likelihood function. However, it is not yet clear how this likelihood function can be easily extended to include model structural and output errors as well.
[16] In summary, current model calibration strategies are poorly suited to the task of explicitly treating the various important sources of uncertainty associated with the application of hydrologic models. In the next two sections of this paper, we will discuss the challenging question of how to develop a model calibration strategy that explicitly accounts for input, output and model structural errors.
3. Sequential Data Assimilation
[18] Even though the KF or EKF implement a more sensible modeling strategy than traditional deterministic simulation methods by explicitly specifying the uncertainty in the system states that arises from imperfect process representation and from input and output data uncertainty, their widespread application appears to have been limited, both due to the strongly nonlinear effects of threshold type structures commonly found in hydrologic models, and due to the extra programming and heavy computational requirements associated with the storage and forward integration of the error covariance matrix. To resolve these two major problems, Evensen [1994] proposed the ensemble Kalman filter (EnKF), which uses a Monte Carlo (MC) method to generate an ensemble of model trajectories from which the time evolution of the probability density of the model states, and related error covariances are estimated. The EnKF avoids many of the problems associated with the traditional EKF method, e.g., there is no closure problem as is introduced in the EKF by neglecting contributions from higher-order statistical moments in the error covariance evolution. Moreover, the conceptual simplicity, relative ease of implementation and computational efficiency of the EnKF make the method an attractive option for data assimilation in the meteorologic, oceanographic and hydrologic sciences [e.g., Evensen and van Leeuwen, 1996; Houtekamer and Mitchell, 1998; Lermusiaux, 1999; Madsen and Cañizares, 1999; Keppenne, 2000; Reichle et al., 2002].
3.1. Ensemble Kalman Filter
[19] The EnKF uses an ensemble of model trajectories to solve for -. The description of the basic algorithm of the EnKF below follows Evensen [1994] and is further illustrated in Figure 2.
[26] 7. Check stop criterion. If t is less than the total number of time steps, n, return to step 2, otherwise stop.
[27] In summary, the EnKF uses a large ensemble of model trajectories to approximate the probability density of the model states at each time step. The mean of this ensemble represents the “best” state estimate, whereas the variance provides a measure of the spread of the ensemble members (uncertainty). For more information about the derivation of the EnKF and its practical implementation, please refer to Evensen [1994].
3.2. Simulation of the Stochastic Model Error Term
[28] To successfully implement the EnKF it is particularly important to specify accurate values of the output measurement error and stochastic forcing term used to simulate the time evolution of dynamical model noise. Hence it is not difficult to see that the size of these terms directly controls the spread among the ensemble members, and as such the mean ensemble model forecasts and associated prediction uncertainty bounds. Fortunately, in most situations a reasonable a priori estimate can be made of the measurement error, σo, of the observations. A more delicate issue however, is how to retrieve a reasonable estimate of the stochastic forcing term. This term reflects many sources of uncertainty including model discretization errors and neglected dynamical processes.
4. Combining Global Optimization and Sequential Data Assimilation
[31] Even though the EnKF provides a general framework to account for input, output, and model structural uncertainty, a defining characteristic of the filter is its focus on time recursive assimilation of the output measurements to update estimates of the state variables, while the parameters are fixed at predefined values, thereby explicitly ignoring the effects of parameter uncertainty and interaction. In contrast, parameter estimation using a global optimization method does not use the output data to recursively update the state variables, but uses batch processing of the data to search for the parameter estimates that minimize the overall (statistical) variance of the model residuals, thereby improving the prediction accuracy in an overall statistical sense. This section discusses how the strengths of a global optimization method can be merged with EnKF state estimation to derive a hybrid framework, which recursively tracks the model states, while simultaneously estimating the values of the model parameters (and their associated uncertainty). This hybrid framework, simultaneous optimization and data assimilation (SODA), consists of an inner EnKF loop for recursive state estimation conditioned on an assumed parameter set, and an outer global optimization loop for batch estimation of the high probability region of the posterior density of the parameters.
[33] To generate samples from equation (25), and to summarize the posterior parameter pdf using statistical moments and histograms, we use an implementation of the Shuffled Complex Evolution Metropolis (SCEM-UA) algorithm. The SCEM-UA algorithm is a general purpose global optimization algorithm that provides an efficient estimate of the most likely parameter set and its underlying posterior probability distribution within a single optimization run [see Vrugt et al., 2003a]. The algorithm is a Markov Chain Monte Carlo (MCMC) sampler, which generates a sequence of parameter sets {θ(1),θ(2),…,θ(k+1)} that convergences to the stationary posterior distribution for a large enough number of simulations k. The SCEM-UA algorithm is related to the successful SCE-UA global optimization method, but uses the Metropolis-Hastings (MH) search strategy [Metropolis et al., 1953; Hastings, 1970] instead of the downhill simplex method for population evolution, and is therefore able to simultaneously infer both the most likely parameter set and its underlying posterior probability distribution within a single optimization run. A detailed description and explanation of the method appears in Vrugt et al. [2003a] and so will not be repeated here.
[34] In brief, the SCEM-UA method involves the initial selection of a population of points distributed randomly throughout the p-dimensional feasible parameter space. In the absence of prior information about the location of the maximum likelihood value a uniform sampling distribution is used. For each point the complete system of equations of the EnKF, as outlined in steps 1–7 in section 3.1, is executed and the time series of mean ensemble forecast errors returned to the SCEM-UA algorithm. This time series is subsequently inserted in equation (25) to calculate the posterior density for each point. The population of parameter sets is subsequently partitioned into a number of complexes, and in each complex a parallel sequence is launched from the point that exhibits the highest posterior density. A new candidate point in each sequence is generated using a multivariate normal distribution either centered on the current draw of the sequence or the mean of the points in the complex augmented with the covariance structure induced between the points in the complex. The Metropolis annealing [Metropolis et al., 1953] criterion is used to test whether the candidate point should be added to the current sequence. Finally, the new candidate point is shuffled into the original population of complexes. The evolution and shuffling procedures are repeated until the Gelman-Rubin convergence diagnostic for each of the parameters demonstrates convergence to a stationary posterior target distribution [Gelman and Rubin, 1992]. Experiments conducted using standard mathematical test problems have shown that the SCEM-UA-derived posterior distribution closely approximates the target distribution [Vrugt et al., 2003a].
[35] The overriding characteristic of the SODA methodology is the improved treatment of input, output, parameter and model structural uncertainties, during model calibration. Theoretically, if the model structure is correct, and the input and output data are observed without error, no state adjustments are needed during model calibration, and the set of parameters that maximizes the posterior density in equation (25), will also be associated with, on average, the smallest output residuals, when performing traditional batch calibration without state adjustments (as done in section 2). However, because of errors in the model structure and input (forcing) and output data this will generally not be the case. The major objective of this paper is to demonstrate how the hybrid SODA strategy can be used to simultaneously estimate model parameters and state variables, resulting in reliable model prediction uncertainty bounds and a time series of state and output innovations whose interpretation will generate inspiration to improve our model concepts and as such our understanding of the functioning of hydrologic systems.
5. Case Studies
[36] We illustrate the power and applicability of SODA by means of two different case studies. The first is a classical mathematical study using the three-parameter highly nonlinear Lorenz model [Lorenz, 1963], to demonstrate the ability of SODA to simultaneously estimate state variables and parameter values when confronted with highly nonlinear model dynamics. The second case study explores the usefulness of SODA by application to hydrologic modeling using the simple HYMOD conceptual watershed model and historical streamflow data from the Leaf River watershed in Mississippi. In this study we are especially concerned with the estimation of model structural errors and present a novel nonparametric method, which is especially designed to estimate the measurement error of streamflow data.
5.1. Case Study 1: A Simple Demonstration Using the Lorenz Equations
[38] The error covariance of the stochastic forcing is defined to be diagonal with variances equal to 2.00, 12.13, and 12.31 for the three equations in (26) respectively, and is included in the model through the stochastic forcing term in equation (22). The error variances of the initial conditions and observations were taken to be the same values as those used for generating the synthetic observations. The values for the parameters, initial conditions, and error variances used in the setup of this case study are identical to those used in similar studies reported in the literature [see, e.g., Evensen, 1994; Miller et al., 1994; Evensen, 1997; Evensen and van Leeuwen, 2000].
[39] The SODA algorithm was used to simultaneously estimate the state variables and model parameters using the synthetically generated observations and the density criterion in equation (25). A search population of 50 points in combination with five parallel sequences was selected based on recommendations in our previous work. The feasible parameter space was taken to be a uniform hypercube between [0–30], [0–50], and [0–10] for the parameters γ, ρ, and β, respectively. Furthermore, an ensemble of 100 members was used to compute the error statistics with the EnKF. Experiments with larger ensemble sizes (not shown here) gave only marginal improvements in the approximation of the probability density of the model states. The estimated ensemble prediction uncertainty ranges associated with the computed x values in equation (26) and corresponding to the most likely SODA identified parameter set is shown in Figure 3a. The solid circles correspond to the observed x values, whereas the dashed black line represents the evolution of the mean ensemble prediction. Note that the prediction uncertainty ranges generally bracket the observations very well, indicating that the EnKF does a good job in tracking the state transitions. Further, the width of the ensemble prediction uncertainty bounds is consistent with the specified error variance of the model and observations.
[40] The transitions of the parameters γ, ρ, and β in each of the five sequences (Markov chains) during the evolution of SODA to the stationary posterior distribution is illustrated in the Figures 3b–3d respectively.
[41] For clarity, the three different parallel sequences are coded with different symbols. The asterisks at the right-hand side indicate the “true” values of the parameters. The 1-D scatterplots of the sampled parameter space demonstrate that at early stages during the evolution, the individual sequences tend to occupy different regions of the parameter space. At a later stage during the evolution, however, all of the individual sequences have been able to fully explore the feasible parameter space, and indeed they assign the highest probability to the “true” parameter values used for generating the synthetic observations. All in all, we can conclude that SODA is able to simultaneously estimate state variables and model parameters when confronted with nonlinear model dynamics, so that we can proceed with the next case study.
5.2. Case Study 2: The HYMOD Conceptual Watershed Model
[42] We illustrate the power and applicability of the SODA algorithm to hydrologic modeling by application to the HYMOD conceptual watershed model using historical data from the Leaf River watershed (1950 km2) located north of Collins, Mississippi. The data, obtained from the Hydrologic Research Laboratory (HL), consists of mean areal precipitation (mm/day), potential evapotranspiration (mm/day), and streamflow (m3/s). Forty consecutive years of data (1948–1988) are available for this watershed, representing a wide variety of hydrologic conditions. The HYMOD model, consisting of a relatively simple rainfall excess model, described in detail by Moore [1985], connected with two series of linear reservoirs (three identical quick and a single reservoir for the slow response). A schematic overview of HYMOD appears in Figure 4.
[43] The model has five parameters: the maximum storage capacity in the catchment, Cmax (L), the degree of spatial variability of soil moisture capacity within the catchment, bexp, the factor distributing the flow between the two series of reservoirs, Alpha, and the residence time of the linear quick and slow flow reservoirs, Rq (days) and Rs (days), respectively. The upper and lower bounds that define the prior uncertainty ranges of these parameters appear in Table 1.
Parameter | Description | Minimum | Maximum | Unit |
---|---|---|---|---|
Cmax | maximum storage watershed | 200.00 | 500.00 | mm |
bexp | spatial variability soil moisture capacity | 0.10 | 2.00 | |
Alpha | distribution factor between two reservoirs | 0.50 | 0.99 | |
Rs | residence time slow flow reservoir | 0.00 | 0.10 | days |
Rq | residence time quick flow reservoir | 0.30 | 0.70 | days |
[44] Because the HYMOD model and the Leaf River data have been discussed extensively in previous work [e.g., Yapo et al., 1998; Boyle et al., 2000; Vrugt et al., 2002; Misirli, 2003], we will not describe the details of these here. To implement the SODA framework outlined above, it is further necessary to specify an appropriate value for the parameter ϕ in equation (21). As our knowledge about a correct value for this first-order decorrelation coefficient is quite poor, we added this parameter to the optimization problem, using a feasible range between [0] and [1]. On the basis of our experience with the model we used 3 years of streamflow data (28 July 1952–1955) for calibration purposes, to illustrate that the SODA method is relatively simple to implement and helps to distinguish between the various sources of uncertainty in hydrologic modeling. An additional data period of approximately 5 years (29 July 1955–30 September 1960) was subsequently used for evaluation purposes. To generate the one-step-ahead streamflow forecasts with the HYMOD model during the calibration and evaluation period we assumed full knowledge of the rainfall time series by using measured rainfall amounts of the next day.
5.2.1. Derivation of the Measurement Error of Streamflow Data
[45] To solve the model calibration problem using SODA, we must do two things. First, a realistic estimate of the error deviation of the output (streamflow) data must be retrieved. This issue will be addressed in this section. Second, we must find a way to obtain a reasonable estimate of input and model structural uncertainty. We will tackle this problem in the next section.
[48] To demonstrate the validity of the proposed approach for estimating the error deviation of streamflow data, we synthetically generated two different test problems. Synthetic daily streamflow data, {y1,…,yn} were first generated for the period 28 June 1952–30 September 1960, by driving the HYMOD model with observed mean areal rainfall of the Leaf River watershed and values for the parameters identical to those used in Figure 1. This output time series was subsequently corrupted by adding one of the two following error models to the data to yield two different time series of “observed” streamflow data; (1) a homoscedastic error model: ɛt ∼ N(0,10); and (2) a heteroscedastic error model: ɛt∼ N(yt,0.1yt). Both these time series were subsequently used in conjunction with equation (29) to estimate the error deviation of the streamflow measurements. The results of this analysis are summarized in Figure 5, which presents scatterplots of “observed” streamflow data versus estimated error deviations for the (Figure 5a) homoscedastic and (Figure 5b) heteroscedastic error case.
[49] To further facilitate graphic interpretation of the results, the dark black line in each of the figures represents the optimal fit of a spline function through the data, whereas the dashed line corresponds to the original error model used to corrupt the streamflow data. Notice, that the agreement between the predicted and actual error model can be considered quite excellent. This is especially true for the homoscedastic error case. Although not further reported here, we performed a variety of synthetic experiments with different error distributions, and found that the nonparametric estimator in equation (29) provides a close approximation to the actual error model, when the time series of streamflow data is sufficiently long (>2 years of daily data).
[50] To unravel the relationship between flow level and measurement error for the Leaf River watershed, Figure 6, presents a similar scatterplot as Figure 5 for this real world data set. Again, the black line denotes the optimal fit of a spline function through the scattered data using nonlinear regression.
[51] The pattern of the scatterplot and slope of the regression function clearly indicate the presence of heteroscedastic errors in the streamflow data. In general, the size of the streamflow error almost log-linearly increases from 0.01 to about 80 m3/s in the flow range of 2 to 1000 m3/s respectively. These results provide strong support for the claim by Sorooshian and Dracup [1980], that streamflow data exhibit nonhomogeneous (heteroscedastic) errors. To optimally exploit the information contained in the scatterplot of Figure 6, it would seem most productive to first cluster the data into various flow levels, and then to approximate the corresponding error probabilities with Gaussian mixture models. However, because the purpose of this case study is illustrative, we decided not to follow such an approach, but rather to implement the optimal fit nonlinear regression function (see Figure 6) into SODA, thereby relating the error deviation of the streamflow data to the flow level.
5.2.2. Derivation of Input Error Model and Error Deviation of the Stochastic Forcing
[52] In the absence of a compelling basis for the assignment of a reasonable input error model to stochastically perturb the elements of , we decided to merge input and model structural errors into a single forcing term. The approach that we propose here to estimating the total error deviation of this term, σm in equation (24), is to conduct a classical deterministic calibration of the model, and to use the resulting model residuals (total error) for the subsequent error analysis and decomposition into component terms as described below.
[53] If we proceed with classical model calibration the output is a set of residuals which constitute a combination of model structural, input and output errors (see Figure 1). The difference between this residual and the associated measurement error of the observation represents an estimate of the combined effect of model structural and input error. To illustrate this, consider Figure 7, which presents a scatterplot of the error deviation of the residuals of the previously calibrated HYMOD model (see section 2) against flow level, created using 10 years of streamflow data for the Leaf River watershed.
[54] The dashed black line denotes the estimated measurement error model (see Figure 6), whereas the solid line represents the optimal fit of a spline function through the scattered data. Accordingly, the size of the model structural and input error can be estimated from the difference between the solid and dashed black lines. A spline function was fitted through this difference function and subsequently implemented in SODA to relate the error standard deviation of the stochastic forcing to flow level.
5.3. Use of SODA for Hydrologic Model Calibration: Model Precision and Accuracy
[55] The SODA algorithm was used with the particular settings discussed in the previous two sections to simultaneously estimate the HYMOD state variables and model parameters using the daily streamflow data of the Leaf River basin and the density criterion specified in equation (25). A search population of 250 points in combination with five parallel sequences and 100 ensemble members was selected. The procedure used about 1500 parameter evaluations to converge to a reasonable estimate of the optimal model parameters, including their underlying posterior distribution. The results of this analysis are summarized in Table 2 and 8, 9– 10 and are discussed below.
Statistics | Calibration (WY 1952–1955) | Validation (WY 1955–1960) | ||||
---|---|---|---|---|---|---|
RMSE | CORR | BIAS, % | RMSE | CORR | BIAS, % | |
SCEM-UA | 17.25 | 0.903 | −5.04 | 20.48 | 0.88 | −4.65 |
SODA | 13.14 | 0.96 | 0.65 | 14.32 | 0.95 | 0.82 |
[56] The usefulness of the implementation of SODA can be demonstrated in a number of different ways. One of the most straightforward is to compare the hydrograph prediction uncertainty ranges of the HYMOD model using SODA, with those obtained using a classical Bayesian SCEM-UA calibration, which ignores input uncertainty and lumps output and model error into a single white noise term. The size of the prediction uncertainty ranges is a measure of the precision of the model. Figure 8 presents the results of this analysis.
[57] The hydrograph prediction uncertainty ranges (lightly shaded region) of the HYMOD model corresponding to the probability distribution of SODA identified parameter sets, each using an ensemble size of 100 members for state estimation, is presented in Figure 8a. The observed streamflows are indicated with solid circles, the parameter uncertainty is indicated by the darkly shaded region, the total uncertainty (parameter, structure plus input, and output) is indicated by the lightly shaded region, and the prediction corresponding to SODA identified “best” parameter set is indicated with the dashed black line. Figure 8b presents the prediction uncertainty ranges for the HYMOD simulated streamflows corresponding to only the SCEM-UA-derived posterior parameter estimates. From visual inspection of the two plots we see (as expected) that the input-output behavior of the HYMOD model when properly accounting for input, output and model structural errors, is more consistent with the observations. Hence in the case of SODA, the prediction uncertainty ranges are reasonably small and bracket the observations. On the contrary, when applying a Bayesian approach (Figure 8b), without state adjustments, the model seems to be unable to match large portions of the hydrograph; this is indicated by sections where the darkly shaded region does not bracket the observed streamflow data. Although one might argue that this is a problem of overconditioning, and that better results can be obtained by assigning more relaxed cutoff thresholds to determine what can be considered an acceptable parameter set or not (as is done in GLUE), that in fact leads to considerably larger hydrograph uncertainty bounds than those presented for SODA, particularly during the storm events. Moreover, as argued in the first two sections of this paper, it is not realistic to attribute all uncertainty in the modeling exercise to uncertainty in the parameter estimates.
[58] Another diagnostic measure, which can be used to contrast the results of SODA with those obtained using classical Bayesian model calibration, is to compare the autocorrelation functions of the SCEM-UA (see Figure 1c) and SODA-derived time series of forecast errors. Autocorrelation is a measure of the accuracy of the model predictions. Figure 9 presents the results of this analysis.
[59] To benchmark against a time series of white forecast errors, the dashed lines in these plots represent the theoretical 99% confidence intervals of a white error time series [Box and Jenkins, 1976]. When performing a traditional SCEM-UA calibration, there is significant autocorrelation between the residuals at the first few lags, confirming our earlier findings reported in Figure 1. Notice, however, that there is considerably less autocorrelation between the forecast residuals when using the SODA framework, suggesting that recursive state adjustments remove a large part of the bias in the model predictions.
[60] This is further demonstrated in Table 2, which presents summary statistics of the 1-day-ahead streamflow prediction performance of the SODA and SCEM-UA methods for the 3 year calibration (28 July 1952–1955) and 5 year evaluation (29 July 1955–30 September 1960) period.
[61] The results in this table indicate that SODA consistently provides better values of the Root Mean Square Error (RMSE), correlation (CORR), and bias (BIAS) statistics than conventional batch calibration for both time periods. Although not further demonstrated here, the performance of SODA with a simple five-parameter conceptual model compares favorably well with other modeling approaches, including the 13-parameter Sacramento Soil Moisture Accounting model of Burnash et al. [1973], and the multiparameter Self-Organizing Linear Output (SOLO) map, recently developed by Hsu et al. [2002]. Thus explicitly accounting for uncertainty in the input, output and model structure, during model calibration will lead to both an improved assessment of predictive uncertainty, and a simultaneous improvement in the model forecasts.
5.4. Use of SODA for Hydrologic Model Calibration: Parameter Uncertainty
[62] Figure 10 presents posterior marginal probability density distributions for each of the HYMOD model parameters estimated for the Leaf River watershed using the SCEM-UA (Figures 10a–10e) and SODA (Figures 10f–10j) algorithms.
[63] The results presented in this figure highlight two interesting observations. First, notice that for some of the model parameters the SCEM-UA and SODA algorithms have assigned the highest probability to different locations in the parameter space. If the model structure would be correct and the input and output data were observed without error, no state adjustments would be needed during model calibration and the SCEM-UA and SODA method would identify the same location for the mode of the posterior probability distribution in the full parameter space. Not properly accounting for model structural and input errors during model calibration leads to corrupted parameter estimates, which are compensating for these errors.
[64] On the basis of these propositions, we therefore suggest that the model parameters identified with SODA are less corrupted by model structural and input errors and, therefore, better represent the underlying properties of the physical system. Our view has also been supported by additional numerical experiments with synthetically generated streamflow data, and a corrupted input (rainfall) data during the inverse identification (see, e.g., Kavetski et al. [2003] for a similar approach), which demonstrated that model parameters identified with SODA seem to be less sensitive to errors in the input and converge in close proximity of the true parameter values used to generate the synthetic streamflow data. To increase the prospects of finding useful regionalization relationships, it would therefore seem most productive to use the SODA identified mode of the parameter space.
[65] A second significant and interesting observation is that the HYMOD model parameters become less well defined when allowing for recursive state adjustments during model calibration (compare histograms of identical counterparts in Figure 10). For most of the parameters, the dispersion around the mode of the distribution has significantly increased. In the systems theoretic sense, this can be explained as part of the “information” from the streamflow measurements being used in SODA to recursively estimate the state variables.
5.5. Use of SODA for Hydrologic Model Calibration: Toward Model Structural Improvements
[66] An interesting byproduct of SODA, which deserves further investigation, is the computed time series of state and output innovations. Any inability of the model to represent the input-output behavior of the underlying hydrologic system, will result in state updates in the model, when new measurements of the system are assimilated and processed. We are therefore left with the intuitively reasonable hypothesis that the computed time series of state updates should contain valuable information about model structural errors. In the last section of this paper, we verify the validity of this hypothesis by closer examination of the time series of output innovations.
[67] To facilitate this process, we begin by partitioning the hydrograph into a driven, nondriven quick, and a nondriven slow component, in a method similar to Boyle et al. [2000]. Another strategy to exploit the information contained in the innovations for the case of transfer function models, is presented in Young [2001, 2002a, 2002b]. For each of these portions of the hydrograph, we plot the SODA-computed mean ensemble output innovation against the mean ensemble streamflow prediction (prior to the update) for the calibration period. The results of this analysis are presented in Figure 11.
[68] For the driven and nondriven quick part of the hydrograph, these plots do not reveal any simple relationship between the ensemble mean predicted flow level (x axis) and associated output innovations (y axis). Nonlinear time series analysis (not presented here) using artificial neural networks (ANN) indicates that some structure is, indeed, present in the time series of output innovations, but the process of how to translate these findings into simple mathematical equations which can be used to improve the conceptual model and as such our understanding of hydrologic processes, remains the topic of future work. A more easily interpretable result is obtained for the nondriven slow part of the hydrograph (Figure 11c). Here, a clear systematic linear relationship between flow level and associated output innovation is apparent. This relationship can easily be exploited to improve the performance of the HYMOD model during low flows. Indeed, when incorporating a linear relationship between flow level and output innovation into HYMOD, the predictive capabilities of the model during nondriven slow flow increases from an average error of 0.20 to 0.09 m3/s. In contrast, the classical model calibration methodologies, which assign all the uncertainty in the input-output representation to uncertainty in the model parameters, provide little or no such guidance.
6. Summary
[69] The objective of model identification is to obtain a model where the input-state-output behavior is consistent with the measurements, and where the model predictions are accurate and precise. In practice, however, because of errors in the model structure, input (forcing), and output data, this has proven to be quite difficult, leading to considerable uncertainty in the model predictions. Classical model calibration strategies typically ignore input uncertainty, lump model structural and output errors into a single white noise term, and assign the uncertainty in the input-output representation of the model primarily to the parameters, therefore lacking the conceptual rigor required to properly account for all the important sources of uncertainty.
[70] In this paper we present a combined global optimization and data assimilation method, which improves the treatment of input, output, parameter, and model structural uncertainty in hydrologic model calibration. The method, entitled simultaneous optimization and data assimilation (SODA), merges the strengths of the parameter search efficiency and explorative capabilities of the Shuffled Complex Evolution Metropolis algorithm and the power and computational efficiency of the ensemble Kalman filter to simultaneously estimate parameters and state variables in hydrologic models.
[71] The usefulness and applicability of SODA has been demonstrated for two preliminary case studies. The first case study considered the highly nonlinear three-parameter Lorenz model, and demonstrated that SODA is indeed successfully able to simultaneously estimate state variables and model parameters when confronted with highly nonlinear model dynamics. The second case study explored the usefulness of SODA by application to hydrologic modeling using the HYMOD conceptual watershed model and historical streamflow data from the Leaf River watershed in Mississippi. The ability of SODA to deal with input, output, parameter and model structural errors, results in improved estimates of the parameter and model prediction uncertainty ranges. With this basis, a detailed investigation of the state and output innovation time series, can be used to investigate improvements to our model concepts and as such our understanding of the functioning of hydrologic systems.
[72] Research aimed at further improvements of SODA is ongoing. This includes (1) the derivation and implementation of realistic input error models to further distinguish between input and model structural errors, (2) the extension to multicriteria problems; SODA does not exclude the use of different parameter sets to match different portions of the hydrograph, (3) the use of recursive parameter estimation methods to further minimize the short-term bias using the state augmentation technique, and (4) the development of improved filtering methods which further relax the Gaussian assumptions of the error distributions in classical KF implementations. Potential improvements related to these issues will be reported in future papers. Also, one of the reviewers of this paper suggested that the use of probabilistic approaches that define uncertainty boundaries, such as SODA, might motivate and support the use of structurally simpler hydrologic models. This is an interesting conjecture that is beyond the scope of the current work and is definitely worthy of further investigation. Software used in this and related work can be found at http://www.science.uva.nl/ibed/cbpg/products.
Acknowledgments
[73] The Earth Life Sciences and Research Council (ALW) partly supported the investigations of the authors with financial aid from the Netherlands Organization for Scientific Research (NWO). The third author received support from the Center for Sustainability of Semi-arid Hydrology and Riparian Areas (SAHRA) under the STC Program of the National Science Foundation, agreement EAR-9876800. The constructive review comments of the Associate Editor, Peter Young, and Jean-Philippe Drecourt significantly improved the current version of this paper.