Title: Calibration Guidelines
1Calibration Guidelines
Model development
Model testing
9. Evaluate model fit 10. Evaluate optimal
parameter values 11. Identify new data to
improve parameter estimates 12. Identify new data
to improve predictions 13. Use deterministic
methods 14. Use statistical methods
1. Start simple, add complexity carefully 2. Use
a broad range of information 3. Be well-posed
be comprehensive 4. Include diverse observation
data for best fit 5. Use prior information
carefully 6. Assign weights that reflect
observation error 7. Encourage convergence by
making the model more accurate 8. Consider
alternative models
Potential new data
Prediction uncertainty
2Model DevelopmentGuideline 1 Apply principle of
parsimony(start simple, add complexity
carefully)
Start simple. Add complexity as warranted by the
hydrogeology, the inability of the model to
reproduce observations, the predictions, and
possibly other things. But my system is so
complicated!! DATA ARE LIMITED. SIMULATING
COMPLEXITY NOT SUPPORTED BY THE DATA CAN BE
USELESS AND MISLEADING.
3Model complexity ? accuracy
- Neither a million grid nodes nor hundreds of
parameters guarantee a model capable of accurate
predictions. - Here, we see that a model that fits all the data
perfectly can be produced by adding many
parameters, but the resulting model has poor
predictive capacity. It is fitting observation
error, not system processes. - We dont know the best level of complexity. We
do know that we dont want to start matching
observation error. Observation error is evaluated
when determining weights (Guideline 6).
4Model complexity ? accuracy Model simplicity ?
accuracy
- General situation Tradeoff between model fit and
prediction accuracy with respect to the number of
parameters
5Can 20 of the system detail explain 90 of the
dynamics?
- The principle of parsimony calls for keeping the
model as simple as possible while still
accounting for the main system processes and
characteristics evident in the observations, and
while respecting other system information. - Begin calibration by estimating very few
parameters that together represent most of the
features of interest. - The regression methods provide tools for more
rigorously evaluating the relation between the
model and the data, compared to trial and error
methods. - It is expected (but not guaranteed) that this
more rigorous evaluation produces more accurate
models.
6Flow through Highly Heterogeneous Fractured Rock
at Mirror Lake, NH Tiedeman, et al., 1998
720 of the system detail does explain 90 of the
dynamics!
MODFLOW Model with only 2 horizontal hydraulic
conductivity parameters
Conceptual Cross Section Through Well Field
Fracture zoneswith high T
Fracture zones with high T
Fractures with low T
8Apply principle of parsimony to allaspects of
model development
- Start by simulating only the major processes
- Use a mathematical model only as complex as is
warranted. - When adding complexity, test
- Whether observations support the additional
complexity - Whether the additional complexity affects the
predictions - This can require substantial restraint!!!
9Advantages of starting simple and building
complexity as warranted
- Transparency Easier to understand the simulated
processes, parameter definition, parameter
values, and their consequences. Can test whether
more complexity matters. - Refutability Easier to detect model error.
- Helps maintain big picture view consistent with
available data. - Often consistent with detail needed for accurate
prediction. - Can build prediction scenarios with detailed
features to test the effect f those features. - Shorter execution times
10Issues of computer execution time
- Computer execution time for inverse models can be
approximated using the time for forward models
and the number of parameters estimated (NP) as - Tinverse 2(NP) ? Tforward ? (1NP)
- (1NP) is the number of solutions per
parameter-estimation iteration - 2(NP) is an average number of parameter
estimation iterations - To maintain overnight simulations, try for
- Tforward lt about 30 minutes
- Tricks
- Buffer sharp K contrasts where possible
- Consider linear versions of the problem as much
as possible (for gw problems replace the water
table with assigned thickness unless the
saturated thickness varies alot over time
replace nonlinear boundary conditions such as EVT
and RIV Packages of MODFLOW with GHB Package
during part of calibration) - Parallel runs
11Calibration Guidelines
Model development
Model testing
9. Evaluate model fit 10. Evaluate optimal
parameter values 11. Identify new data to
improve parameter estimates 12. Identify new data
to improve predictions 13. Use deterministic
methods 14. Use statistical methods
1. Start simple, add complexity carefully 2. Use
a broad range of information 3. Be well-posed
be comprehensive 4. Include diverse observation
data for best fit 5. Use prior information
carefully 6. Assign weights that reflect
observation error 7. Encourage convergence by
making the model more accurate 8. Consider
alternative models
Potential new data
Prediction uncertainty
12Model DevelopmentGuideline 2 Use a broad range
of information (soft data) to constrain the
problem
- Soft data is that which cannot be directly
included as observations in the regression - Challenge to incorporate soft data into model so
that it (1) characterizes the supportable
variability of hydrogeologic properties, and (2)
can be represented by a manageable number of
parameters. - For example, in ground-water model calibration,
use hydrology and hydrogeology to identify likely
spatial and temporal structure in areal recharge
and hydraulic conductivity, and use this
structure to limit the number of parameters
needed to represent the system. - Do not add features to the model to attain model
fit if they contradict other information about
the system!!
13Example Parameterization for simulation of
ground-water flow in fractured dolomite(Yager,
USGS Water Supply Paper 2487, 1997)
How to parameterize hydraulic conductivity in
this complex system? Yager took advantage of data
showing that the regional fractures that dominate
flow are along bedding planes.
14Example Parameterization for simulation of
ground-water flow in fractured dolomite
- Transmissivity estimated from aquifer tests is
roughly proportional to the number of fractures
intersecting the pumped well. - Thus, assume all major fractures have equal T,
and calculate T for each model layer from the
number of fractures in the layer. - The heterogeneous T field can then be
characterized by a single model parameter and
multiplication arrays.
15Example Death Valley Regional Flow System
Surficial Deposits Volcanic Rocks Carbonate
Rocks Granitic Rocks pC Igneous, Meta.
50 KM
16Death Valley flow system Simulated
Hydraulic-Conductivity Distribution
Total of 9 K Parameters for this extremely
complex hydrogeology!!!
Layer 1 500m
Layer 2 750m
Layer 3 1250m
3-layer model DAgnese . 1997, 1999
17Data management, analysis, and visualization
- Data management, analysis, and visualization
problems can be daunting. It is difficult to
allocate project time between these efforts and
modeling in an effective manner, because - There are many kinds of data (point well data, 2D
and 3D geophysics, cross sections, geologic maps,
etc) and the subsurface is often very complex.
Capabilities for integrating these data exist,
but can be cumbersome. - The hardware and software change often. Thus far,
products have been useful, but not dependable or
comprehensive. - Low end Rockworks US2000. High end
Earthvision US100,000
US20,000/yr
GUIs provide some capabilities
18Data management, analysis, and visualization
- For MODFLOW, graphical interfaces have been
produced commercially since mid-80s. - Now, 5 major interfaces Argus ONE, GMS,
Groundwater Vistas, PMWin, Visual Modflow. GMS
funded in part by DOD, Argus application by USGS.
These interfaces work very well for many systems. - 700-7000
- Have various limitations regarding fully 3D data
management
Argus
GMS
GW Vistas
PMWin
Visual Modflow
19Calibration Guidelines
Model development
Model testing
9. Evaluate model fit 10. Evaluate optimal
parameter values 11. Identify new data to
improve parameter estimates 12. Identify new data
to improve predictions 13. Use deterministic
methods 14. Use statistical methods
1. Start simple, add complexity carefully 2. Use
a broad range of information 3. Be well-posed
be comprehensive 4. Include diverse observation
data for best fit 5. Use prior information
carefully 6. Assign weights that reflect
observation error 7. Encourage convergence by
making the model more accurate 8. Consider
alternative models
Potential new data
Prediction uncertainty
20Model DevelopmentGuideline 3 Be Well-Posed Be
Comprehensive
Well posed Dont spread observation data too
thinly For a well-posed problem, estimated
parameters are supported by the calibration
observations, and the regression converges to
optimal values. In earth systems, observations
are usually sparse, so being well-posed often
leads to models with few parameters. Comprehensive
Include many system aspects. Characterize as
many system attributes as possible using defined
model parameters. Leads to many parameters.
Is achieving Guideline 3 possible? Challenge
Bridge the gap. Develop a useful model that has
complexity the observation data can support and
the predictions need.
21- Be Well-Posed and Be Comprehensive
- Often harder to be well posed than to be
comprehensive. - Easy to add lots of complexity to a model.
- Harder to limit complexity to what is supported
by the observations and most important to
predictions. - Keeping the model well-posed can be facilitated
by - Scaled sensitivities, parameter correlation
coefficients, leverage statistics - Independent of model fit. Can use before model
is calibrated - Cooks D, DFBetas (influence statistics)
- Advantage -- integrate sensitivities and
parameter correlations. - Caution -- dependent on model fit. Use cautiously
with uncalibrated model.
22Sensitivities
- Derivatives of dependent variables with respect
to model parameters. Sensitivity of a simulated
value yi to parameter bj is expressed as ?yi/
?bj (can be thought of as ?yi/ ?bj) - Use to estimate parameters by nonlinear
regression. - Use to identify important observations and
parameters. Scale to get meaningful comparisons.
Otherwise the different possible units of the yi
and bj obscure what we want to know.
Scaled Sensitivities
- Can use during model calibration to help
determine which parameters to estimate and what
new data would be beneficial. - Independent of model fit, so can use in early
stages of calibration -- poor model fit wont
affect their values.
23- Dimensionless scaled sensitivity (dss)
- Scale ?y/ ?b to make it dimensionless.
- Use dss to compare importance of different
observations to estimation of a single parameter
larger absolute values of dss indicate greater
importance.
Composite scaled sensitivity (css)
- An averaged measure of all dss for a single
parameter - Indicates importance of observations as a whole
to a single parameter can use to help choose
which parameters to estimate by regression. - If css for a parameter is more than about 2
orders of magnitude smaller than the largest css,
regression may have difficulty estimating the
parameter and converging.
24Dimensionless Scaled Sensitivities Support of
each observation for each parameter(example from
Death Valley)
- Estimation of parameter K4 seems to be dominated
by 4 observations 3 heads and 1 flow. - Scaled sensitivities neglect parameter
correlation, so some observations may be more
important than indicated. In ground-water
problems, flows are very important for reducing
correlations.
Heads obs 1-501 Flows obs 502-517
3 dominant head obs
1 dominant flow obs
25Composite Scaled Sensitivities Support of whole
observation set for each parameter
- CSS for initial Death Valley model with only 9
parameters. - Graph clearly reveals relative support the
observations as a whole provide towards
estimating each parameter. - Observations provide much information about RCH
and 2 or 3 of the K parameters little
information about ANIV or ETM
Supportable model complexity
? The observations provide enough information to
add complexity to the K and RCH parameterization
26Composite Scaled Sensitivities Support of whole
observation set for each parameter
Supportable model complexity
Good way to show the observation support as the
number of defined parameters becomes large. This
graph is from the final Death Valley model.
- Black bars parameters estimated by regression.
- Grey bars not estimated by regression because of
parameter correlation, insensitivity, or other
reasons.
27Parameter Correlation Coefficients (pcc)
- A measure of whether or not the calibration data
can be used to independently estimate a pair of
parameters. - If pcc gt 0.95, then it may not be possible to
estimate the 2 parameters uniquely using the
available regression data. In this case, changing
parameter values in a coordinated manner will
likely produce very similar conditions.
28Parameter correlations DVRFS model
- pcc gt0.95 for 4 parameter pairs in the
three-layer DVRFS model with - all 23 parameters active
- no prior information
- 501 head observations
- 16 flow observations.
Parameter pair Parameter pair Correlation
GHBgs K7 -0.99
K2 RCH3 0.98
RCH3 Q2 0.97
K1 RCH3 0.96
- With head data alone, all parameters except
vertical anisotropy are perfectly correlated --
Multiply all by any positive number, get
identical head distribution. By Darcys Law. - The flow observations reduce the correlation to
what is shown above.
29Influence Statistics
- Like DSS, they help indicate if parameter
estimates are largely affected by just a few
observations - Like DSS, they depend on the type, location, and
time of the observation - Unlike DSS, they depend on model fit to the
observed value. - Unlike DSS, they include the effect of pcc
(parameter correlation coefficient) (Leverage
does this, too) - Cooks D a measure of how a set of parameter
estimates would change with omission of an
observation, relative to how well the parameters
are estimated given the entire set of
observations.
30Cooks D Which observations are most important
to estimating all the parameters?(3-layer Death
Valley example)
Accounts for sensitivities, parameter
correlations, and model fit
- Estimation dominated by 10 of the observations
- 5 obs very important 3 heads, 2
flows. - Importance of flows is better reflected by Cooks
D than scaled sensitivities. In gw problems,
flows often resolve extreme correlation. Need
flows to uniquely estimate parameter values. - Although dependent on model fit, relative
valuesof Cooks D can be useful for uncalibrated
models.
flow obs (502-517)
31Sensitivity Analysis for 2 parameters
- CSS
- DSS
- Leverage
- Cooks D
- DFBETAS
- Conclusion flow qleft has a small
sensitivity but is critical to uncoupling
otherwise completely correlated parameters.
32Which statistics address which relations??
Observations Parameters - Predictions
dss pss css
ppr pcc leverage Param
eter cv AIC BIC DFBETAS Cooks D opr
Observations ---------------- Predictions
33Problems with Sensitivity Analysis Methods
- Nonlinearity of simulated values with respect to
the parameters - Inaccurate sensitivities
34Nonlinearity
Nonlinearity sensitivities differ for different
parameter values.
- Scaled sensitivities change for different
parameter values because (1) the sensitivities
are different and (2) the scaling. dss
(?y/?b)bw1/2 - Consider decisions based on scaled sensitivities
to be preliminary. Test by trying to estimate
parameters. If conclusions drawn from scaled
sensitivities about what parameters are important
and can be estimated change dramatically for
different parameter values, the problem may be
too nonlinear for this kind of sensitivity
analysis and regression to be useful.
- Parameter correlation coefficients commonly
differ for different parameter values. - Extreme correlation is indicated if pcc1.0 for
all parameter values regression can look okay
but beware! (see example in Hill and Tiedeman,
2003)
(pcc)
From Poeter and Hill, 1997. See book p. 58
35Inaccurate sensitivities
- How accurate are the sensitivities?
- Most accurate sensitivity-equation method.
MODFLOW-2000. Generally 5-7 digits - Less accurate Perturbation methods. UCODE_2005
or PEST. Often only 2-3 digits Both programs
can use model-produced sensitivities if
available. - When does it NOT matter?
- Scaled sensitivities, regression often do not
require accurate sensitivities. Regression
convergence improves with more accurate
sensitivities for problems on the edge. Mehl and
Hill, 2002 - When does it matter?
- Parameter correlation coefficients. Hill and
Østerby, 2003 - Values of 1.00 and 1.00 reliably indicate
parameter correlation smaller absolute values do
not guarantee lack of correlation unless the
sensitivities are known to be sufficiently
accurate. - Parameter correlation coefficients have more
problems as sensitivity accuracy declines for all
parameters, but it is most severe for pairs of
parameters for which one parameter or both
parameters have small composite scaled
sensitivity.
36Calibration Guidelines
Model development
Model testing
9. Evaluate model fit 10. Evaluate optimal
parameter values 11. Identify new data to
improve parameter estimates 12. Identify new data
to improve predictions 13. Use deterministic
methods 14. Use statistical methods
1. Start simple, add complexity carefully 2. Use
a broad range of information 3. Be well-posed
be comprehensive 4. Include diverse observation
data for best fit 5. Use prior information
carefully 6. Assign weights that reflect
observation error 7. Encourage convergence by
making the model more accurate 8. Consider
alternative models
Potential new data
Prediction uncertainty
37Model Development Guideline 4 Include many
kinds of data as observations (hard data) in the
regression
- Adding different kinds of data generally provides
more information about the properties of the
simulated system. - In ground-water flow model calibration
- Flow data are important. With only head data, if
all major K and Recharge parameters are being
estimated, extreme values of parameter
correlation coefficients will likely occur
(Darcys Law). - Advective transport (or concentration
first-moment data) can provide valuable
information about the rate and direction of
ground-water flow. - In ground-water transport model calibration
- Advective transport (or concentration
first-moment data) important because they are
more stable numerically and the misfit increases
monotonically as the fit to observations becomes
worse. (Barth and Hill, 2005a,b, Journal of
Contaminant Hydrology)
38Here, model fit does not change with changes in
the parameter values unless overlap occurs
From Barth and Hill (2005a). Book p. 224
39Effect of having many observation data
typesMiddle Rio Grande Basin (Sanford , 2004
Plummer , 2004)
Water-Level Observations
Carbon-14 Observations
- These data sets allowed for estimation of 59
parameters representing K, recharge, and
anisotropy, with no large correlation
coefficients.
40Contoured or kriged data values as
observations?(book p. 284)
- Has the advantage of creating additional
observations for the regression. - However, a significant disadvantage is that the
interpolated values are not necessarily
consistent with processes governing the true
system, e.g. the physics of ground-water flow for
the true system. For example, interpolated values
could be unrealistically smooth across abrupt
hydrogeologic boundaries in the true subsurface. - This can cause estimated parameter values to
representative of the true system poorly.
? Proceed with Caution !!!!
41Calibration Guidelines
Model development
Model testing
9. Evaluate model fit 10. Evaluate optimal
parameter values 11. Identify new data to
improve parameter estimates 12. Identify new data
to improve predictions 13. Use deterministic
methods 14. Use statistical methods
1. Start simple, add complexity carefully 2. Use
a broad range of information 3. Be well-posed
be comprehensive 4. Include diverse observation
data for best fit 5. Use prior information
carefully 6. Assign weights that reflect
observation error 7. Encourage convergence by
making the model more accurate 8. Consider
alternative models
Potential new data
Prediction uncertainty
42Model DevelopmentGuideline 5 Use prior
information carefully
- Prior information allows some types of soft data
to be included in objective function (e.g. T from
aquifer test) - Prior information penalizes estimated parameter
values that are far from expected values
through an additional term in the objective
function. - What are the expected values?
HEADS FLOWS PRIOR
43Ground-Water Modeling
- Hydrologic and hydrogeologic data less accurate
- Relate to model inputs
- Dependent variable observations more accurate
- Relate to model outputs - calibration
Ground-Water Model -- Parameters
Predictions
Prediction uncertainty
Societal decisions
44Suggestions
- Begin with no prior information, to determine the
information content of the observations. - Insensitive parameters (parameters with small
css) - Can include in regression using prior information
to maintain a well-posed problem - Or during calibration exclude them to reduce
execution time. Include them when calculating
prediction uncertainty and associated measures
(Guidelines 12 14). - Sensitive parameters
- Do not use prior information to make unrealistic
optimized parameter values realistic. - Figure out why model calibration data together
cause regression to converge to unrealistic
values (see Guideline 9).
45Highly parameterized models
- parameters gt observations
- Methods
- Pilot points (de Marsily, RamaRao, LaVenue)
- Pilot points with smoothing (Tikhonov)
- Pilot points with regularization (Alcolea,
Doherty) - Sequential self calibration (Gomez-Hernandez,
Hendricks Franssen) - Representer (Valstar)
- Moment mehod (Guadagnini, Neuman)
- Most common current usage PEST regularization
capability, by John Doherty
46Why highly parameterize?
- Can easily get close fits to observations
- Intuitive appeal to resulting distributions
- We know the real field varies
- Highly parameterized methods can be used to
develop models with variable distributions - Mostly used to represent K can use if for other
aspects of the system
47Why not highly parameterize?
- Are the variations produced by highly
parameterized fields real? - Perhaps NO if they are produced because of
- Data error (erroneous constraint)
- Lack of data (no constraint)
- Instability
- How can we know?
- Here, consider synthetic problem.
- Start with no observation error
- Add error to observations
48(From Hill and Tiedeman, 2006, Wiley)
11 Observations 10 heads (), 1 flow (to
river) 6 Parameters HK_1 HK_2
(multiplier) RCH_1, RCH_2 K_RB (prior) VK_CB
(prior) Steady state
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . .
Top layer really homogeneous. Consider using 36
pilot points to represent it.
Model Layer 1
Model Layer 2
49Initial run with no data error
Here, use the constrained minimization
regularization capability of PEST
Variogram inputs Nugget0.0 a3x107
Variance0.1 Interpolation inputs Search
radius36,000 Max points36
Consistent with the constant value that actually
occurs.
Equals fit achieved with a single HK_1 parameter
given a correct model and observations with no
error
Fit Criterion phi1x10-5
50Introducing variability
The variogram and interpolation input reflects a
much better understanding of the true
distribution than would normally be the case. To
create a more realistic situation, use ß, the
regularization weight factor
S(b) (y - yobs)T w (y - yobs) ß
(?p - 0)T w (?p - 0) To the extent that ß can
be small, variability is allowed in the K
distribution. ? Increased s2lnK
51No data error. Correct model.
Perform regression starting from values other
than the true values
Percent error for each parameter calculated as
100?(btrue-best)/btrue
For HK_1, best exp(lnK) lnKmean of the ln
Ks of the 36 pilot points Also
report s2lnK
If ß is restricted to be close to 1.0, the
estimates are close to the true and s2lnK is
small, as expected. What happens if ß can be
small?
52No data error. Correct model.
Perform regression starting from values other
than the true values
Percent error for each parameter calculated as
100?(btrue-best)/btrue
For HK_1, best exp(lnK) lnKmean of the ln
Ks of the 36 pilot points
s2ln(k) 0.1
ß
53No data error. Correct model.
Perform regression starting from another set of
values
Same good fit to obs
54No data error. Correct model.
Parameter estimates depend on starting parameter
values
- Disturbing.
- Means that in the following results as ß becomes
small discrepancies may not be caused by
observation error.
55Data error. Correct model.
- Parameter error
- Distribution of K in layer 1
56Parameter error
Data error. Correct model.
s2 ln(k) 1.2
Not possible to determine for which phi values
will be accurate. Small parameter error ?
Accurate predictions? Depends on effect of
the variability.
57(No Transcript)
58Lessons
- Inaccurate solutions?
- Possibly. Variability can come from actual
variability, or from data error or instability of
the HP method. - Exaggerated uncertainty?
- Possibly, if include variability caused by data
error and instabilities. True of Moore and
Doherty method?? - How can we take advantage highly parameterized
methods? - Use hydrogeology to detect unrealistic solutions.
- Analyze observation errors and accept that level
of misfit. Set phi accordingly. - Consider weighting of regularization equations.
Use b1? - Check how model fit changes as phi changes.
- Use sensitivity analysis to identify extremely
correlated parameters and parameters dominated by
one observation. Use parsimonious overlays.
59Calibration Guidelines
Model development
Model testing
9. Evaluate model fit 10. Evaluate optimal
parameter values 11. Identify new data to
improve parameter estimates 12. Identify new data
to improve predictions 13. Use deterministic
methods 14. Use statistical methods
1. Start simple, add complexity carefully 2. Use
a broad range of information 3. Be well-posed
be comprehensive 4. Include diverse observation
data for best fit 5. Use prior information
carefully 6. Assign weights that reflect
observation error 7. Encourage convergence by
making the model more accurate 8. Consider
alternative models
Potential new data
Prediction uncertainty
60Observation error???Measurement error vs.
model error
- Should weights account for only measurement
errors, or also some types of model error? - A useful definition of observation error that
allows for inclusion of some types of model error
is - Any error with an expected value of zero that is
related to aspects of the observation that are
not represented in the model (could be, for
example, the model configuration and/or the
simulated processes). - Unambiguous measurement errors Errors associated
with the measuring device and the spatial
location of the measurement. - More ambiguous errors Heads measured in wells
that partially penetrate a model layer. Here,
even in a model that perfectly represents the
true system, observed and simulated heads will
not necessarily match. The weights could be
adjusted to account for this mismatch between
model and reality if the expected value of the
error is zero.
61Model DevelopmentGuideline 6 Assign weights
that reflect observation error
- Model calibration and uncertainty evaluation
methods that do not account for observation
error can be misleading and are not worth using. - Observation errors commonly are accounted for
by weighting - Can use large weights (implied small
observation error) to investigate solution
existence. - For uncertainty evaluation the weights need to be
realistic
62Model DevelopmentGuideline 6 Assign weights
that reflect observation error
- Strategy for uncorrelated errors
- Assign weights equal to 1/s2, where s2 is the
best estimate of the variance of the measurement
error (details given in Guideline 6 of the book).
- Values entered can be variances (s2), standard
deviations (s) or coefficients of variation (CV).
Model calculates variance as needed.
- Advantages
- Weight more accurate observations more heavily
than less accurate observations intuitively
appealing - This produces squared weighted residuals that are
dimensionless, so they can be summed in the
objective function. - Use information that is independent of model
calibration, so statistics used to calculate the
weights generally are not changed during
calibration. - This weighting strategy is required for common
uncertainty measures to be correct. - Without this the regression becomes difficult and
arbitrary.
63If weights do not reflect observation error,
regression becomes difficult and arbitrary
Interested in predictions for large values of x.
?
yb0b1x
y
?
?
?
?
?
?
?
?
?
?
x
No, not if the model is correct and the other
data are important.
64Determine weights by evaluating errors associated
with the observations
- Often can assume (1) a normal distribution and
(2) different error components are additive. - If so, two things are important.
- add variances, not standard deviations or
coefficients of variation! - Mean needs to equal zero!
- Quantify possible magnitude of error using
- range that is symmetric about the observation or
prior information - probability with which the true value is expected
to occur within the range. - Examples
- Head observation with three sources of error
- Streamflow loss observation
65Head observation with three sources of error
- (1) Head measurement is thought to be good to
within 3 feet. - (2) Well elevation is accurate to 1.0 feet.
- (3) Well located in the model within 100 feet.
Local gradient is 2. - Quantify good to within 3 feet as there is a
95-percent chance the true value falls within 3
feet of the measurement. - Use a normal probability table to determine that
a 95-percent confidence interval is a value
1.96 times the standard deviation, s. - This means that 1.96 x s 3 ft., so s 1.53 ft.
- Quantify well elevation error similarly to get
s0.51 ft. - Quantify located in the model within 100 feet.
Locally the gradient is 2 as there is a
95-percent chance the true value falls within
plus and minus 2 feet - Using the procedure above, s1.02 ft.
- Calculate observation statistic.
Add
variances (1.53)2 (0.51)2 (1.02)2 3.64 ft2
s1.91 ft. Specify variance or
standard deviation in the input file.
66Streamflow loss observation
- Streamflow gain observation derived by
subtracting flow measurements with known error
range and probability. - Upstream and downstream flow measurements 3.0
ft3/s and 2.5 ft3/s ? loss observation 0.5
ft3/s. The first flow measurement is considered
fair, the second is good. Carter and Anderson
(1963) suggest that a good streamflow measurement
has a 5 error.
- Quantify the error. 5 forms a 90 confidence
interval on the fair upstream flow plus and
minus 5 forms a 95 confidence interval on the
good downstream flow. Using values from a
normal probability table, the standard deviations
of the two flows are 0.091 and 0.64. - Calculate observation statistic
Add variances (0.091)2 (0.64)21/2
0.0124.
Coefficient of variation 0.0124/0.5 0.22, or
22 percent. Specify variance, standard
deviation or coefficient of variation in the
input file
67What if weights are set to unrealisticallylarge
values?
Objective function surfaces(contours of
objective function calculated for combinations of
2 parameters)
With flow weighted using a reasonable coefficient
of variation of 10
With flow weighted using an unreasonable
coefficient of variation of 1
Heads only
From Hill and Tiedeman, 2002. Book p. 82.
68Unrealistically large weightsHazardous
consequences for predictions
Unreasonable weighting results in misleading
calculated confidence interval excludes true
value of 1737 m
From Hill and Tiedeman, 2002. Book p. 303