The model–data fusion pitfall: assuming certainty in an uncertain world

Oecologia (2011) 167:587–597
DOI 10.1007/s00442-011-2106-x
CONCEPTS, REVIEWS AND SYNTHESES
The model–data fusion pitfall: assuming certainty
in an uncertain world
Trevor F. Keenan • Mariah S. Carbone •
Markus Reichstein • Andrew D. Richardson
Received: 11 December 2010 / Accepted: 5 August 2011 / Published online: 8 September 2011
Ó Springer-Verlag 2011
Abstract Model–data fusion is a powerful framework by
which to combine models with various data streams
(including observations at different spatial or temporal
scales), and account for associated uncertainties. The
approach can be used to constrain estimates of model
states, rate constants, and driver sensitivities. The number
of applications of model–data fusion in environmental
biology and ecology has been rising steadily, offering
insights into both model and data strengths and limitations.
For reliable model–data fusion-based results, however, the
approach taken must fully account for both model and data
uncertainties in a statistically rigorous and transparent
manner. Here we review and outline the cornerstones of a
rigorous model–data fusion approach, highlighting the
importance of properly accounting for uncertainty. We
conclude by suggesting a code of best practices, which
should serve to guide future efforts.
Communicated by Russell Monson.
T. F. Keenan (&) A. D. Richardson
Department of Organismic and Evolutionary Biology,
Harvard University, 26 Oxford Street, Cambridge,
MA 02138, USA
e-mail: [email protected]
M. S. Carbone
National Center for Ecological Analysis and Synthesis,
Santa Barbara, CA 93101, USA
M. Reichstein
Biogeochemical Model–Data Integration Group,
Max-Planck-Institute for Biogeochemistry,
07745 Jena, Germany
Keywords Model–data fusion Data assimilation Parameter estimation Inverse analysis Carbon cycle
model
Introduction
‘‘Solum ut inter ista certum sit nihil esse certi’’
‘‘In these matters the only certainty is that nothing is
certain.’’
—Pliny the Elder (23 AD–79 AD)
Many fields within environmental biology and ecology
are rapidly becoming ‘‘data rich’’ (Luo et al. 2008). Novel
technologies (e.g., new instruments and automated measurement devices, wireless networking and communication
protocols, as well as data processing software and storage
hardware) are making the collection of vast amounts of data
possible. For example, continuous measurements of surface–atmosphere CO2 and H2O fluxes via eddy covariance
(Baldocchi 2008) or soil–atmosphere CO2 fluxes via soil
respiration chambers (Carbone and Vargas 2008; Vargas
et al. 2011) provide continuous, high-frequency data on
ecosystem carbon cycling at time scales from minutes to
years. Continental-scale monitoring through ‘‘ecological
observatories’’, such as NEON (http://www.neoninc.org),
ICOS (http://www.icos-infrastructure.eu), FLUXNET (http://
www.fluxnet.ornl.gov; http://www.fluxdata.org), and LTER
(http://www.lternet.edu), offer exciting new possibilities for
research, but will require many scientists to update the
statistical and analytical tools they use in order to meet new
challenges for analysis and synthesis.
One possible approach to making sense of the mountain
of data now available to ecologists is to embrace model–
data fusion (MDF) techniques. With MDF, also referred to
123
588
Fig. 1 The model–data fusion process: a conceptual diagram showing the main steps involved, and the iterative and interactive nature of
the approach. The figure is adapted from Williams et al. (2009)
as ‘‘data-model synthesis’’ or ‘‘data assimilation’’ (in a
broad sense), the process model (chosen according to the
research questions being pursued) is used to provide a
rigorous, analytical framework for data interpretation,
synthesis, and generalization or extrapolation (see Zobitz
et al. 2011 for a detailed primer). MDF results are conditional on the information content of observational data and
on what is known or assumed about uncertainties (Raupach
et al. 2005). Here, we refer specifically to three sources of
uncertainty, namely, those associated with (1) observational uncertainties (incomplete and noisy observational
data, systematic biases, etc.), (2) model structure uncertainties (e.g., in the specification of model processes and
internal relations), and (3) uncertainties in model parameters and states.
Briefly, in MDF, data of diverse types (including, in a
Bayesian context, ‘‘prior knowledge’’) can be used to
constrain model parameters or states, evaluate competing
model structures or process representations, and scale
upwards in space or time, while simultaneously accounting
for—and propagating—different sources of uncertainty.
MDF can be viewed as an iterative comparison of a model
with observations (Fig. 1). At each iteration model
parameters are changed, with the objective being to find the
parameter combinations that give equivalently best fits to
the available data. Such an approach can contribute to
model development and improvement, along with the
synthesis of various data forms (Williams et al. 2009).
MDF analyses can yield mechanistic insight into the processes underlying observed biological responses to environmental forcing and can be used to obtain best estimates
and the associated uncertainties of model parameters (e.g.,
physiological rate constants) and states––what Bayesians
refer to as the ‘‘posterior distributions’’. Recent studies,
(e.g., Fox et al. 2009; Medvigy et al. 2009; Wu et al. 2009;
Alton and Bodin 2010; Carvalhais et al. 2010; Desai 2010;
Richardson et al. 2010; Weng and Luo 2011; Zhang et al.
2010; Zhou et al. 2010), represent the MDF state-of-the-art
123
Oecologia (2011) 167:587–597
in the field of terrestrial ecosystem carbon cycle modeling
[for recent reviews see Wang et al. (2009) and Williams
et al. (2009)].
Uncertainty pervades all aspects of scientific research
(Taylor 1997), both for models and data. The strength of
MDF lies in, and is dependent on, the incorporation of both
model and data uncertainty within the analytical framework (Raupach et al. 2005; Williams et al. 2009; Cressie
et al. 2009). In fact, as suggested by Raupach et al. (2005),
results reached without a full and accurate consideration of
uncertainty might as well not have been reached at all!
Nevertheless, explicit assessments of uncertainty meet with
several conceptual and practical difficulties, ranging from
the accurate quantification of model and data error to
commonly held views that reporting uncertainty devalues
results (Pappenberger and Beven 2006).
Few MDF studies fully characterize model and data
uncertainty. In this paper we review aspects of uncertainty
and implications for MDF results. Using a forest carbon
cycle model in a synthetic experiment, we show that MDF
results are highly conditional on how the associated
uncertainties are treated. Based on this, we make the case
for an open and honest assessment of uncertainty (Brown
2010) in MDF applications and suggest a ‘‘code of best
practices’’ (Table 1). Although we focus on applications to
the terrestrial carbon cycle, the problems raised are relevant to a wide range of MDF applications in environmental
biology and ecology.
Synthetic experiment
A synthetic experiment is an experiment that uses artificially generated data. In such an experiment the ‘‘true’’
values of observations and parameters are known, giving a
controlled environment in which to test the impact of different assumptions (Code of best practices #4, Table 1).
Throughout the text, we make use of results from five
different synthetic MDF experiments using Markov chain
Monte-Carlo techniques (Hastings 1970) with a simple
forest carbon cycle model. The experiments (described
below) are designed to test the ability of the MDF framework to retrieve the parameters that generated the synthetic
data and to assess the importance of different assumptions
regarding data uncertainty.
We use the data assimilation-linked ecosystem carbon
model (DALEC, Williams et al. 2005) at the Howland
forest. The Howland Forest AmeriFlux site is located in
central Maine, USA, at the southern ecotone of the North
American boreal spruce–fir zone and consists largely of red
spruce (Picea rubens Sarg.) and eastern hemlock [Tsuga
canadensis (L.) Carr.]. The model integrates the key
components of forest productivity and carbon allocation
Oecologia (2011) 167:587–597
589
Table 1 Suggested components for a model–data fusion ‘‘code of best practices’’
‘‘Code of best practices’’ for model–data fusion analyses
1. Data uncertainties should be evaluated in an open and realistic manner. Data error should be fully characterized and fed forward into
the model–data fusion (MDF) framework. Enough detail should be provided so that the assumptions can be evaluated and the method
duplicated
2. Alternative model structures should be assessed, particularly for model components associated with unobserved variables (e.g., Zobitz
et al. 2008)
3. Multiple data constraints should be implemented whenever possible, to minimize over-fitting to any single data stream and to constrain
processes operating at different time scales (e.g., Xu et al. 2006; Barrett et al. 2005; Richardson et al. 2010)
4. The MDF framework should be tested against synthetic data with error properties similar to those of the real data in order to assess the
capacity for retrieving accurate estimates of model parameters and states
5. Attempts should be made to validate the optimized model against independent data (e.g., Moore et al. 2008; Zobitz et al. 2008;
Richardson et al. 2010)
(a) It should not be claimed that just because the optimized model can reproduce the data used as constraints that the model is good or
correct (inference only with extreme caution)
(b) It should be acknowledged that, without validation, predictions of an optimized model only represent untested hypotheses
6. Confidence intervals (posterior distributions) on model parameters, states, and predictions should be estimated in a transparent
manner. Alternative approaches to estimating these uncertainties should be considered
Table 2 Synthetic data
experiments
Experiment
Synthetic data description
Parameters
optimized
MDF assumptions
1.
Observed data gaps added and no noise
12 parameters
No error in data
2.
Observed data gaps and 10% noise
12 parameters
Fixed 10% error in data
3.
Observed data gaps and observed noise
12 parameters
Observed error in data
4.
Observed data gaps and observed noise
6 pools
Observed error in data
5.
Observed data gaps and observed noise
12 parameters
Understate data error by 25%
and respiration. Model states and parameters were optimized as described in Richardson et al. (2010; see their
Fig. 2, Table 1) using 4 years of observational data
(1997–2000). The model was optimized against five different data streams in parallel [half-daily net ecosystem
exchange of CO2 (NEE) from tower eddy-covariance
measurements, leaf area index, wood carbon, leaf litterfall,
and soil respiration] (see Richardson et al. 2010 for
details). The optimized model output for 1997 was then
used as the basis for our synthetic data.
Three synthetic data sets were created for five experiments as outlined in Table 2; the same synthetic data were
used in experiments 3 through 5. Gaps added to the model
output for each experiment were distributed according to
the observed gap frequency and inherent sampling frequency of the original data constraints. Experiment 1 used
output from the constrained model with added gaps but no
added noise to test the ability of the MDF framework to
retrieve the parameters that had generated the data. No data
error was assumed, and the cost function (the estimated
model–data mismatch, described below) was constructed
as a non-weighted least squares estimator. For the
remaining experiments, stochastic measurement error was
simulated as random noise added to each data point, drawn
from a normal distribution with a mean of zero. For
Experiment 2, the added data noise was heteroscedastic
with a standard deviation fixed at 10%. For experiments 3,
4, and 5, noise added to the synthetic data approximated the
observed error distribution (as described by Richardson
et al. 2010), including heteroscedasticity. In Experiment 5
we underestimated data uncertainty in the cost function to
illustrate the effect of assuming the data error is less than it
really is (Table 2).
For each experiment, the cost function (a measure of
data-model mismatch which guides the optimization process and parameter exploration stage) was constructed as
the mean (Eq. 2) of individual cost functions (Eq. 1) for
each data stream. Individual data stream cost functions
were constructed as the total uncertainty-weighted squared
data–model mismatch, divided by the number of observations for each data stream (NI):
!
NI X
yI ðtÞ pI ðtÞ 2
ð1Þ
jI ¼
=NI
dI ðtÞ
t¼1
where yI(t) is a data constraint at time t for data stream I,
pI(t) is the corresponding model predicted value, and dI(t)
is the measurement specific uncertainty as specified in
123
590
Oecologia (2011) 167:587–597
Richardson et al. (2010). For the aggregate multi-objective
cost function we use the mean of the individual cost
functions, which can be written as:
!
M
X
J¼
jI =M
ð2Þ
i¼1
where M is the number of data streams used. The cost
function approach applied here gives equal weight to each
data stream, regardless of the number of observations in each.
Three parameter optimizations were performed for each
experiment (and compared to test for convergence). In
total, 200,000 Monte Carlo iterations, with an ensemble of
10,000 sets of parameters accepted on the basis of a v2 test,
were used to characterize parameter distributions and 90%
confidence intervals. The v2 test was performed on each jI
individually, after normalizing each jI by the minimum
error obtained, jopt (see variance normalization: Franks
et al. 1999; Richardson et al. 2010). The retrieved parameter values and their associated confidence intervals were
then compared with the values that were used to generate
the synthetic data.
Data uncertainty
All measurements, whether from a meticulous lab experiment or a remote field campaign and whether recorded by a
human observer or a sophisticated instrument, are subject
to uncertainties (Taylor 1997). In fact, in many cases
(particularly within the environmental sciences), the
uncertainties can be on the same order of magnitude as the
data itself. As an example, tower and chamber measurements of CO2 fluxes have substantial uncertainties associated with them, and these frequently violate commonly
held assumptions of normality and constant variance
(Richardson et al. 2008; Savage et al. 2008). Random
uncertainties on individual, 30-min measurements of the
net ecosystem exchange of CO2 are commonly approximately ±50% (at 95% confidence interval) of the measured
flux, or larger (Richardson et al. 2008). Coupled with
inherent systematic biases, uncertainties on measured
fluxes are also substantial at longer time scales, such as at
the monthly or annual time step.
An additional source of data uncertainty often comes
from the difference between what the measurements represent and the real quantities of interest (Williams et al.
2009; Richardson et al. 2010). For example, flux tower
instrumentation measures the net flux of CO2, which is not
in itself a process that is modeled directly. Various
decomposition techniques are available to estimate the
associated component fluxes of gross photosynthesis (GPP)
and ecosystem respiration (RE) (Desai et al. 2008), but the
measured CO2 exchange does not directly correspond to
123
the processes that we are usually attempting to model.
While it is conceivable that partitioned fluxes could be used
as model constraints (e.g., Williams et al. 2005; Zhou et al.
2010), the partitioning itself is an additional source of
uncertainty.
Here we use our synthetic experiment to show that both
the amount of error present in data and how it is considered
in the MDF process affect the confidence intervals of
extracted parameters and states. Assigning a 10% error
(Experiment 2) increased the confidence intervals on estimated fluxes and decreased the number of constrainable
parameters (Fig. 2b) compared to considering data with no
error (Experiment 1), where data gaps contributed nominal
uncertainty (Fig. 2a). For synthetic data with the observed
scale-dependent error (Experiment 3), the confidence
intervals were much larger, and the potential for equifinality
(i.e., that different combinations of parameters and unobserved variables can give equally acceptable model behavior) (Beven 2006) in modeled processes increased through a
decrease in parameter constraints (Fig. 2c). This finding is
supported by other MDF studies (Braswell et al. 2005;
Zhang et al. 2010), although some other studies have also
suggested that magnitudes of measurement errors were not
found to considerably influence estimated fluxes (Williams
et al. 2005) or parameter identifiability (Luo et al. 2009).
The MDF framework offers a unique opportunity to
account for data uncertainties (Raupach et al. 2005). That
said, misspecification of the error model can yield welldefined but incorrect estimates of parameter distributions
(Beven et al. 2008). Data uncertainty is not considered in
many MDF studies (e.g., Carvalhais et al. 2010); in particular, data uncertainty is often not acknowledged in
studies using partitioned eddy–covariance flux data (e.g.,
Zhou et al. 2010). Many studies simply treat data error as a
data-constant nominal amount (e.g., Zhang et al. 2005,
2010; Wang et al. 2006; Xie et al. 2009; Zhou et al. 2010),
although the error structure of many data sets is well
characterized as heteroscedastic (e.g., flux data—Richardson et al. 2006, 2008). Another common approach is to
estimate data error using the MDF framework itself (e.g.,
Braswell et al. 2005; Sacks et al. 2007). It is often the case
that estimates of data error are not available. A conservative approach would be to take the largest measurement
error estimate (e.g., Desai et al. 2008; Wang et al. 2009;
Richardson et al. 2010). In such cases, one approach would
be to assume a range of error values and then test the
sensitivity of the results. Regrettably, the majority of
studies simply do not specify how data errors were treated
(Code of best practices #1, Table 1).
While a full treatment of data errors increases the
challenge of conducting MDF analyses, the failure to fully
propagate uncertainties may give a false sense of security
and over-confidence in the model output. This can hamper
Oecologia (2011) 167:587–597
591
Fig. 2 Accepted parameter ranges (left column) and subsequent
estimated annual sums (right column, gC m-2 year-1) for net
ecosystem productivity (iNEP), gross primary production (iGPP),
autotrophic respiration (iRa), and heterotrophic respiration (iRh),
when considering no noise (a Experiment 1), 10% fixed rate noise
(b Experiment 2), and observed data noise (c Experiment 3). Whiskers
90% confidence interval, based on a v2 test against measured data
after variance normalization, filled circles original parameter values
used to create the data set. Results presented represent optimization
against 1 year of synthetic data at Howland forest. Synthetic data
were thinned to represent data gap frequency and distribution at
Howland. Parameters are normalized to their initial ranges (see
Richardson et al. 2010 for detailed description). Carbon pools are not
optimized here (fixed to original values). Parameters included in the
optimization are: a Log10 decomposition rate, b fraction of GPP
respired, c fraction of net primary production (NPP) allocated to
foliage, d log10 turnover rate of foliage, e log10 turnover rate of
wood, f log10 turnover rate of fine roots, g log10 mineralization rate
of litter, h log10 mineralization rate of soil organic matter (SOM),
I temperature dependence of respiratory and decomposition processes, j photosynthetic nitrogen use efficiency, k foliage carbon pool,
l fraction of autotrophic respiration partitioned to belowground organs
progress towards improving our understanding of the
underlying processes and mechanisms or towards identifying the sources of uncertainty that could be reduced by
new measurements or experiments (Brown 2010).
between processes and drivers). Using a different model
approach can often give a very different result, particularly
when the model is used for extrapolation in time or space
(Rastetter 2003). Here we highlight model uncertainty from
two sources: (1) that arising from an incorrect or inadequate description of a process (e.g., not accounting for soil
moisture effects in arid ecosystems) and (2) that arising
from fixed internal relations (e.g., prescribed carbon pool
sizes or allocation rates). A best practice approach to MDF
can allow the testing of such uncertainty (Code of best
practices #2, Table 1).
Model uncertainty
Any given process can usually be modeled in a variety of
ways (e.g., the mathematical structure—in terms of how
processes and states are connected to, or feed back upon,
each other—and the functional form of the relationship
123
592
Fig. 3 Time series of modeled net ecosystem productivity (NEP)
from the optimized model (black line) along with 90% confidence
intervals (gray area) for a model–data fusion run for 1 year of
synthetic data (1997) with observed gap frequency and data noise
(Experiment 4). Optimization was performed on initial pool sizes
only, with all parameters fixed to those used to generate the synthetic
data. The 90% confidence intervals are calculated based on a v2 test
for acceptance/rejection. Inset Accepted range of initial pool sizes
(i.e., the possible combination of pool sizes which could generate the
plotted iNEP) for each pool (Cf carbon in foliage, Cr carbon in roots,
Cw carbon in wood, Clit carbon in litter, Csom carbon in soil organic
matter) normalized against the respective prior range. See Richardson
et al. (2010) for initial range values used. Cw here ranged between the
initial Cw value ± 50%
Although almost a trivial example of model error, the
estimated annual ecosystem respiration, extrapolated from
nocturnal eddy covariance measurements, varied by 50%
or more depending on the model used in the study by
Richardson et al. (2006). At a larger scale, the C4MIP
project showed that different model approaches led to very
different answers in terms of the relative sensitivity of
terrestrial productivity versus respiration to future climate
(Friedlingstein et al. 2006). Optimized model parameters
and states are thus conditional on model structure (Franks
et al. 1997). MDF has been successfully used to assess the
error introduced by different process representations (Sacks
et al. 2007; Zobitz et al. 2008). However, systematic model
differences that result from model structure are probably
the most difficult to identify and quantify (Butts et al. 2004;
Abramowitz et al. 2007; Zobitz et al. 2008).
Most models are subject to the setting of fixed internal
relations, such as prescribed pool sizes. We tested the
affect of fixed internal relations in synthetic Experiment 4.
Here, we varied initial pool sizes, which are often set as
fixed values in MDF studies, but kept all other parameters
constant. This allowed us to test if different structural
relations, in the form of pool sizes, could give an equivalent fit to the data. Widely varying combinations of pool
123
Oecologia (2011) 167:587–597
Fig. 4 The 90% confidence intervals for NEP) for a model–data
fusion run using 1 year of synthetic data with observed gap frequency
(Experiment 5). Gray area Confidence interval obtained when
assuming observed data noise, black area confidence interval when
underestimating data error by 25%. Model parameters are optimized,
and initial pools are prescribed. The 90% confidence intervals are
calculated based on a v2 test for acceptance/rejection
sizes gave an equally good model performance (Fig. 3). In
particular, the lack of data on some pool sizes (e.g., labile
carbon, root carbon) allowed the model to use different
initial structural combinations to arrive at a similar final
aggregate flux. On the other hand, pools for which data
constraints were available (foliar and wood carbon) were
well constrained.
Fixed internal parameters and relations are often
assigned based on ‘‘expert opinion’’ or literature values
specific to other sites or species. For example, models often
assign a constant proportion of assimilate to belowground
allocation (e.g., Zhou et al. 2010), whilst there is evidence
that belowground allocation of new assimilate can vary by
up to 500% within the summer season alone (Högberg
2010). Carvalhais et al. (2008) found that the common
assumption that carbon pools are in an equilibrium state
can lead to large differences in estimated parameters, with
results often showing high sensitivity to initial pool sizes
(Santaren et al. 2007). Such fixed internal relations are
good examples of model structural error that can be
addressed in a MDF framework. By including fixed internal relations as variables in the optimization process, their
impact on model performance can be assessed
Confidence intervals on results
While MDF can be used to identify the model parameters
that yield a statistically ‘‘optimal’’ match between data and
model (and thus, in principle, eliminate mis-parameterization as a source of model error), this does not guarantee
Oecologia (2011) 167:587–597
Fig. 5 Modeled daily ecosystem respiration (Reco) its components
[daily autotrophic (Ra) and heterotrophic (Rh) respiration] from the
optimized model (filled circles) along with 90% confidence intervals
(gray area) for a model–data fusion run using 1 year of synthetic data
with observed gap frequency and data noise (Experiment 3). Model
parameters are optimized conditional on the data, and initial pools are
prescribed. The 90% confidence intervals are calculated based on a v2
test for acceptance/rejection
593
that the model output is correct or true. In this context,
statistically valid confidence intervals for predicted model
states and fluxes are needed to strengthen inferences based
on the model analysis.
Since many model–data fusion techniques are rooted in
the Bayesian perspective, it is reasonable (and expected)
that a confidence interval is conditional on both data and
the model, so that the 90% confidence is based on prior
information, data, and information provided by the model.
That said, large variabilities (up to tenfold) in uncertainty
estimates have been reported in previous experiments, even
when algorithms and methods that were, in principle, very
similar, were applied to the same synthetic data with the
same model (Trudinger et al. 2007; Fox et al. 2009). This
large discrepancy leads us to ask the question of what do
the uncertainty estimates really represent? Further work is
clearly warranted to identify the reasons why different
MDF algorithms yield different confidence intervals.
Retrieved parameters are also dependent on an accurate
quantification of the error structure in the observed data
(e.g., Sacks et al. 2007). In our synthetic experiment 5, we
underestimated the data error by 25%. This led to an
overestimation of the apparent confidence in model
parameters (as we show here for daily net ecosystem productivity, Fig. 4). Data uncertainty feeds forward (or is
‘‘mapped’’) to parameter uncertainty, and thus to uncertainty in model predictions. This lends weight to the
importance of adequately characterizing data uncertainty in
order to accurately quantify parameter uncertainty.
A variety of methods are available to estimate confidence ranges, such as bootstrapping approaches [Efron and
Tibshirani (1993); see Richardson and Hollinger (2005)
and Reichstein et al. (2003) for applications]. The simple
v2 statistic, calculated between model and data (Press et al.
2007), can also be used to evaluate whether (for a given
parameter set) data and model are consistent at a given
level of statistical confidence, thereby offering another
basis for generating confidence intervals on model output.
The benefit of such tests is that they take into account both
uncertainties and the degrees of freedom defined by the
number of parameters and observations. Accurately quantifying the degrees of freedom is complicated by difficulties in quantifying the true information content of data. The
current assumption is that all measurements are made at a
frequency that represents the frequency of information
variability in the variable being measured. It may be
advisable to test different approaches to quantifying
parameter and state uncertainty (Code of best practices #6,
Table 1). A conservative approach would then pick the
largest confidence band, keeping in mind that a ‘‘90%
confidence interval’’ is approach- and model-dependent.
123
594
Fig. 6 Acceptable values of modeled annual autotrophic and heterotrophic respiration (iRa, iRh) from the optimized model for a model–
data fusion run using 1 year of synthetic data with observed gap
frequency and data noise (Experiment 3). Each data point represents
an accepted model run (90% confidence intervals) based on a v2 test
for acceptance/rejection. Model parameters are optimized conditional
on the data, and initial pools are prescribed. The figure inset shows the
smoothed temporal correlation of accepted values of Ra and Rh over
the course of the year. Data correspond to those shown in Fig. 3
Inferences, model constraints, and validation
An unobserved variable is a variable in the MDF framework that is not directly observed and thus must be inferred, conditional on the model structure and model
parameters. A good model fit to observations does not
imply a good estimation of the unobserved processes
responsible for the observations (Fig. 5, see discussion
above), as compensating errors have the potential to generate the right overall answer for the wrong reason
(equifinality: Beven 2006; Luo et al. 2009). Notwithstanding, direct inferences about the behavior of unobserved variables are sometimes made on the basis of
optimized model output (e.g., Zhou et al. 2010). This is in
part often unavoidable. For example, the labile carbon pool
of forest ecosystems is an important aspect of carbon
cycling, but data on the size of this pool, at the ecosystem
scale, are scarce due to practical measurement difficulties.
Autotrophic and heterotrophic respirations are good
examples of components often treated in MDF frameworks
as unobserved processes (processes that are not quantified
by direct observations). This is due to the relative lack of
data constraints on belowground processes, which cannot
be constrained by the use of NEE data alone (Sacks et al.
2007; Zobitz et al. 2008). We use synthetic experiment 3 to
highlight the potential for uncertainty in both the annual
contribution to total respiratory fluxes and the temporal
change of that contribution. In our experiment, although
123
Oecologia (2011) 167:587–597
estimates of GPP were well constrained (Fig. 2), both
autotrophic and heterotrophic ecosystem respiration were
subject to large compensating uncertainties (Figs. 5, 6).
This shows that even soil respiration data are not capable
of constraining belowground partitioning between heterotrophic and autotrophic respiration. Furthermore, the
uncertainty associated with the partitioning of ecosystem
respiration into its auto- and heterotrophic components
varied non-linearly throughout the year (Figs. 5, 6). This
reflects the heteroscedastic nature of data errors and nonlinear scaling of respiration with temperature (Fig. 5). The
noisier the data, the lower the model confidence in the
contribution of autotrophic respiration to total soil respiration (Fig. 7). Unobserved variables incorporate all the
uncertainties previously discussed, thus any inference
should therefore be made with extreme caution.
While there are many examples of unobserved variables
that would, in practice, be impossible to measure, others
are simply most often treated as unobserved variables
because they are difficult or expensive to measure. As an
example, total soil respiration is commonly measured (e.g.,
Savage et al. 2008; Vargas et al. 2011), but in MDF
applications, the separate contributions of autotrophic and
heterotrophic respiration to total respiration are more often
treated as unobserved variables. However, methods such as
root exclusion and isotope partitioning (Hanson et al. 2000;
Kuzyakov 2006) can be used to distinguish the contribution
of autotrophic and heterotrophic components. Thus, these
are not necessarily unobserved variables. Similarly, many
Fig. 7 The inferred percentage contribution of autotrophic respiration to total soil respiration, when considering: no noise (a), 10%
fixed rate noise (b), and observed noise (c) (Experiments 1, 2, 3,
respectively). Whiskers 90% confidence interval, based on a v2 test
against measured data. Results represent 1 year of synthetic data at
Howland forest, optimizing model parameters, with initial carbon
pools fixed to original values
Oecologia (2011) 167:587–597
other unobserved variables can be constrained by using
more data streams. For example, as the information content
of tower eddy–covariance data can only constrain a limited
number of parameters (e.g., Knorr and Kattge 2005; Alton
and Bodin 2010), additional data streams are needed to
constrain more parameters even in very simple carbon flux
models (Richardson et al. 2010). Using different data
streams effectively constrains different but non-independent parts of the model, thus reducing equifinality. Additional data, which are not currently widely available, such
as information on the allocation of photosynthetic assimilates or transfer rates among different carbon pools, would
greatly aid in future MDF studies.
Of course, direct validation of results against independent observational data (e.g., Moore et al. 2008; Zobitz
et al. 2008; Richardson et al. 2010) is the acid test to
confirm confidence in model predictions (Code of best
practices #5, Table 1). In the worst-case scenario (no
observed data), the sensitivity of model results to
assumptions regarding parameter estimates and model
structure can be assessed. An operational best practice
would be to use multiple measured data streams (Code of
best practices #3, Table 1) both to constrain the optimization process and provide validation (using independent
training and testing data pools). With respect to validation,
previous studies have demonstrated the benefit of testing
the MDF framework itself by using synthetic data
(Braswell et al. 2005; Zhang et al. 2005; Trudinger et al.
2007; Lasslop et al. 2008; Fox et al. 2009). Even in the
presence of good constraint availability and the potential
for model validation, either through cross-holdout schemes
or outside the optimization domain, conducting a synthetic
experiment as performed in this study is essential to validate the MDF framework itself.
Conclusions
Despite the recent increase in the application of MDF to
questions in environmental sciences and ecology, there
remains no common agreement on best practices with
respect to the characterization and propagation of uncertainties and the validation and testing of model output.
Adequately addressing the uncertainty associated with
MDF results remains a critical challenge. The lack of a
proper consideration of model and/or data uncertainty,
however, invariably gives overconfident and potentially
misleading results. Uncertainty has historically been
regarded as the curse of environmental research (Pappenberger and Beven 2006). However, we would argue the
contrary that acknowledgment of the true inherent uncertainty only serves to improve the quality of both current
and future scientific endeavors. MDF offers a unique
595
framework by which to do this, and we would urge future
studies to apply an open treatment of uncertainty (Brown
2010).
Here, we have discussed the ways in which uncertainties
can be included in MDF analyses, based on the belief that
obtaining accurate uncertainty estimates is every bit as
important as obtaining the ‘‘right’’ result. In Table 1 we put
forward a ‘‘best practices’’ template for model–data fusion.
We encourage further discussion and debate of this
proposal.
Acknowledgments TFK and ADR acknowledge support from the
Northeastern States Research Cooperative, and from the Office of
Science (BER), U.S. Department of Energy, through the Terrestrial
Carbon Program under Interagency Agreement number DE-AI0207ER64355, and through the Northeastern Regional Center of the
National Institute for Climatic Change Research. MR acknowledges
support from the CARBO-Extreme project of the European Commission (FP7-ENV-2008-1-226701). MSC acknowledges support
from the Kearney Foundation of Soil Science.
References
Abramowitz G, Pitman A, Gupta H, Kowalczyk E, Wang P (2007)
Systematic bias in land surface models. J Hydrometeorol
8:989–1001
Alton P, Bodin P (2010) A comparative study of a multilayer and a
productivity (light-use) efficiency land-surface model over
different temporal scales. Agric For Meterorol 150:182–195
Baldocchi D (2008) Breathing of the terrestrial biosphere: lessons
learned from a global network of carbon dioxide flux measurement systems. Aust J Bot 56:1–26
Barrett DJ, Hill MJ, Hutley LB, Beringer J, Xu JH, Cook GD, Carter
JO, Williams RJ (2005) Prospects for improving savanna
biophysical models by using multiple-constraints model-data
assimilation methods. Aust J Bot 53:689–714
Beven K (2006) A manifesto for the equifinality thesis. J Hydrol
320:18–36
Beven KJ, Smith PJ, Freer JE (2008) So just why would a modeller
choose to be incoherent? J Hydrol 354:15–32
Braswell BH, Sacks WJ, Linder E, Schimel DS (2005) Estimating
diurnal to annual ecosystem parameters by synthesis of a carbon
flux model with eddy covariance net ecosystem exchange
observations. Glob Change Biol 11:335–355
Brown JD (2010) Prospects for the open treatment of uncertainty in
environmental research. Prog Phys Geogr 34:75–100
Butts MB, Payne JT, Kristensen M, Madsen H (2004) An evaluation
of the impact of model structure on hydrological modelling
uncertainty for streamflow simulation. J Hydrol 298:242–266
Carbone MS, Vargas R (2008) Automated soil respiration measurements: new information, opportunities and challenges. New
Phytol 177:295–297
Carvalhais N, Reichstein M,Seixas J,Collatz GJ,Santos Pereira J et al
(2008) Implications of the carbon cycle steady state assumptionfor biogeochemical modeling performanceand inverse
parameter retrieval. Glob Biochem Cycles 22:GB2007. doi:
10.1029/2007GB003033
Carvalhais N, Reichstein M, Ciasis P, Collatz GJ, Mohecha MD et al
(2010) Identification of vegetation and soil carbon pools out of
equilibrium in a process model via eddy covariance and
biometric constraints. Glob Change Biol 16:2813–2829
123
596
Cressie N, Calder CA, Clark JS, Hoef JMV, Wikle CK (2009)
Accounting for uncertainty in ecological analysis: the strengths
and limitations of hierarchical statistical modeling. Ecol Appl
19:553–570
Desai AR (2010) Climatic and phenological controls on coherent
regional interannual variability of carbon dioxide flux in a
heterogeneous landscape. J Geophys Res 115:G00J02. doi:
10.1029/2010JG001423
Desai AR, Richardson AD, Moffat AM, Kattge J, Hollinger DY et al
(2008) Cross-site evaluation of eddy covariance GPP and RE
decomposition techniques. Agric For Meteorol 148:821–838
Efron B, Tibshirani RJ (1993) An introduction to the bootstrap.
Chapman and Hall, New York
Fox A, Williams M, Richardson AD, Cameron D, Gove JH et al
(2009) The REFLEX project: comparing different algorithms
and implementations for the inversion of a terrestrial ecosystem
model against eddy covariance data. Agric For Meteorol
149:1597–1615
Franks SW, Beven KJ, Quinn PF, Wright IR (1997) On the sensitivity
of soil-vegetation-atmosphere transfer (SVAT) schemes: Equifinality and the problem of robust calibration. Agric For Meteorol
86:63–75
Franks SW, Beven KJ, Gash JHC (1999) Multi-objective conditioning
of a simple SVAT model. Hydrol Earth Sys Sci 3:488–489
Friedlingstein P et al (2006) Climate-carbon cycle feedback analysis:
Results from the C4MIP model intercomparison. J Climatol
19:3337–3353
Hanson PJ, Edwards NT, Garten CT, Andrews JA (2000) Separating
root and soil microbial contributions to soil respiration: a review
of methods and observations. Biogeochemistry 48:115–146
Hastings WK (1970) Monte-Carlo sampling methods using markov
chains and their applications. Biometrika 57:97
Högberg P (2010) Is tree root respiration more sensitive than
heterotrophic respiration to changes in soil temperature? New
Phytol 188:9–10
Knorr W, Kattge J (2005) Inversion of terrestrial ecosystem model
parameter values against eddy covariance measurements by
Monte Carlo sampling. Glob Change Biol 11:1333–1351
Kuzyakov Y (2006) Sources of CO2 efflux from soil and review of
partitioning methods. Soil Biol Biochem 38:425–448
Lasslop G, Reichstein M, Kattge J, Papale D (2008) Influences of
observation errors in eddy flux data on inverse model parameter
estimation. Biogeosciences 5:1311–1324
Luo Y, Clark JS, Hobbs T, Lakshmivarahan S, Latimer AM, Ogle K,
Schimel D, Zhou X (2008) Symposium 23. Toward ecological
forecasting. Bull Ecol Soc Am 89:467–474
Luo Y, Weng E, Wu X, Gao C, Zhou X, Zhang L (2009) Parameter
identifiability, constraint, and equifinality in data assimilation
with ecosystem models. Ecol Appl 19:571–574
Medvigy D, Wofsy SC, Munger JW, Hollinger DY, Moorcroft PR
(2009) Mechanistic scaling of ecosystem function and dynamics
in space and time: ecosystem demography model version 2.
J Geophys Res 114:G01002. doi:10.1029/2008JG000812
Moore D, Hu J, Sacks WJ, Schimel DS, Monson RK (2008)
Estimating transpiration and the sensitivity of carbon uptake to
water availability in a subalpine forest using a simple ecosystem
process model informed by measured net CO2 and H2O fluxes.
Agric For Meteorol 148:1467–1477
Pappenberger F, Beven KJ (2006) Ignorance is bliss: or seven reasons
not to use uncertainty analysis. Water Resour Res 42: W05302.
doi:10.1029/2005WR004820
Press WH, Teukolsky SA, Vetterling WT, Flannery BP (2007)
Numerical recipes: The art of scientific computing. Cambridge
University Press, Cambridge
Rastetter EB (2003) The collision of hypotheses: what can be learned
from comparisons of ecosystem models? In: Canham CD, Cole
123
Oecologia (2011) 167:587–597
JJ, Lauenroth WK (eds) Models in ecosystem science. Princeton
University Press, Princeton, pp 221–224
Raupach MR et al (2005) model-data synthesis in terrestrial carbon
observation: methods, data requirements and data uncertainty
specifications. Glob Change Biol 11:378–397
Reichstein M et al (2003) Inverse modeling of seasonal drought
effects on canopy CO2/H2O exchange in three Mediterranean
ecosystems. J Geophys Res 108(D23):4726. doi:10.1029/2003
JD003430
Richardson AD, Hollinger DY (2005) Statistical modeling of
ecosystem respiration using eddy covariance data: maximum
likelihood parameter estimation, and Monte Carlo simulation of
model and parameter uncertainty, applied to three simple
models. Agric For Meteorol 131:191–208
Richardson AD, Braswell BH, Hollinger DY, Burman P, Davidson
EA et al (2006) Comparing simple respiration models for eddy
flux and dynamic chamber data. Agric For Meteorol 141:219–
234
Richardson AD, Mahecha MD, Falge E, Kattge J, Moffat AM et al
(2008) Statistical properties of random CO2 flux measurement
uncertainty inferred from model residuals. Agric For Meteorol
148:38–50
Richardson A, Williams M, Hollinger DY, Moore DJP, Dail DB et al
(2010) Estimating parameters of a forest ecosystem C model
with measurements of stocks and fluxes as joint constraints.
Oecologia 164:25–40
Sacks WJ, Schimel DS, Monson RK (2007) Coupling between carbon
cycling and climate in a high-elevation, subalpine forest: a
model-data fusion analysis. Oecologia 151:54–68
Santaren D, Peylin P, Viovy N, Ciais P (2007) Optimizing a processbased ecosystem model with eddy-covariance flux measurements: a pine forest in southern France. Glob Biogeochem Cyc
21:GB2013. doi:10.1029/2006GB002834
Savage K, Davidson EA, Richardson AD (2008) A conceptual and
practical approach to data quality and analysis procedures for
high-frequency soil respiration measurements. Funct Ecol
22:1000–1007
Taylor JR (1997) An introduction to error analysis: the study of
uncertainties in physical measurements. University Science
Books, Sausalito
Trudinger CM, Raupach MR,Rayner PJ,Kattge J,Liu Q et al (2007)
OptIC project: an intercomparison of optimization techniques for
parameter estimation in terrestrial biogeochemical models.
J Geophys Res 112:G02027. doi:10.1029/2006JG000367
Vargas R, Carbone M, Reichstein M, Baldocchi D (2011) Frontiers
and challenges in soil respiration research: from measurements
to model-data integration. Biogeochemistry 102:1–13
Wang YP, Baldocchi D, Leuning R, Falge E, Vesala T (2006)
Estimating parameters in a land-surface model by applying
nonlinear inversion to eddy covariance flux measurements from
eight FLUXNET sites. Glob Change Biol 12:1–19
Wang YP, Trudinger CM, Enting IG (2009) A review of applications
of model–data fusion to studies of terrestrial carbon fluxes at
different scales. Agric For Meteorol 149:1829–1842
Weng E, Luo Y (2011) Relative information contributions of model
vs. data to constraints of short- and long-term forecasts of forest
carbon dynamics. Ecol Appl 21:1490–1505
Williams M, Schwarz PA, Law BE, Irvine J, Kurpius MR (2005) An
improved analysis of forest carbon dynamics using data assimilation. Glob Change Biol 11:89–105
Williams M, Richardson AD, Reichstein M, Stoy PC, Peylin P et al
(2009) Improving land surface models with FLUXNET data.
Biogeosciences 6:1341–1359
Wu XW, Luo YQ, Weng ES, White L, Ma Y, Zhou XH (2009)
Conditional inversion to estimate parameters from eddy-flux
observations. J Plant Ecol 2:55–68
Oecologia (2011) 167:587–597
Xie H, Eheart JW, Chen Y, Bailey BA (2009) An approach for
improving the sampling efficiency in the Bayesian calibration of
computationally expensive simulation models. Water Resour
Res 45:W06419. doi:10.1029/2007WR006773
Xu T, White L, Hul D, Luo Y (2006) Probabilistic inversion of a
terrestrial ecosystem model: analysis of uncertainty in parameter
estimation and model prediction. Glob Biogeochem Cycles
20:GB2007. doi:10.1029/2005GB002468
Zhang Q, Xiao X, Braswell B, Linder E, Baret F, Moore B (2005)
Estimating light absorption by chlorophyll, leaf and canopt in a
deciduous broadleaf forest using MODIS data and a radiative
transfer model. Remote Sens Environ 99:359–371
Zhang L, Luo Y, Yu G, Zhang L (2010) Estimated carbon residence
times in three forest ecosystems of eastern China: applications of
597
probabilistic inversion. J Geophys Res 115:G01010. doi:
10.1029/2009JG001004
Zhou X et al (2010) Concurrent and lagged impacts of an
anomalously warm year on autotrophic and heterotrophic
components of soil respiration: a deconvolution analysis. New
Phytol 187:184–198
Zobitz JM, Moore D, Sacks WJ, Monson RK, Bowling DR, Schimel
DS (2008) Integration of process-based soil respiration models
with whole ecosystem CO2 measurements. Ecosystems 11:250–
269
Zobitz JM, Desai, AR, Moore DJP, Chadwick MA (2011) A primer
for data assimilation with ecological models using markov chain
Monte Carlo (MCMC). Oecologia. doi:10.1007/s00442-0112107-9
123