Methods for Developing Input Distributions for Probabilistic Risk Assessments - PowerPoint PPT Presentation

1 / 111
About This Presentation
Title:

Methods for Developing Input Distributions for Probabilistic Risk Assessments

Description:

When the cost of intervention is smaller than the cost of analysis ... Hypothesis 4: Doing uncertainty analysis actually reduces overall resource ... – PowerPoint PPT presentation

Number of Views:158
Avg rating:3.0/5.0
Slides: 112
Provided by: itre
Category:

less

Transcript and Presenter's Notes

Title: Methods for Developing Input Distributions for Probabilistic Risk Assessments


1
Methods and Applications of Uncertainty and
Sensitivity Analysis
H. Christopher Frey, Ph.D. Professor
Department of Civil, Construction, and
Environmental Engineering North Carolina State
University Raleigh, NC 27695 Prepared for
Workshop on Climate Change Washington,
DC March 7, 2005
2
Outline
  • Why are uncertainty and sensitivity analysis
    needed?
  • Overview of methods for uncertainty analysis
  • Model inputs
  • Empirical data
  • Expert judgment
  • Model uncertainty
  • Scenario uncertainty
  • Overview of methods for sensitivity analysis
  • Examples
  • Technology assessment
  • Emissions Factors and Inventories
  • Air Quality Modeling
  • Risk Assessment
  • Findings
  • Recommendations

3
Why are uncertainty and sensitivity analysis
needed?
  • Strategies for answering this question
  • what happens when we ignore uncertainty and
    sensitivity?
  • what do decision makers want to know that
    motivates doing uncertainty and sensitivity
    analysis?
  • what constitutes best scientific practice?
  • Program and research managers may not care about
    all three, but might find at least one to be
    convincing (and useful)

4
When is ProbabilisticAnalysis Needed or Useful?
  • Consequences of poor or biased estimates are
    unacceptably high
  • A (usually conservative) screening level analysis
    indicates a potential concern, but carries a
    level of uncertainty
  • Determining the value of collecting additional
    information
  • Uncertainty stems from multiple sources
  • Significant equity issues are associated with
    variability
  • Ranking or prioritizing significance of multiple
    pathways, pollutants, sites, etc.
  • Cost of remediation or intervention is high
  • Scientific credibility is important
  • Obligation to indicate what is known and how well
    it is known

5
When is a Probabilistic Approach Not Needed?
  • When a (usually conservative) screening level
    analysis indicates a negligible problem
  • When the cost of intervention is smaller than the
    cost of analysis
  • When safety is an urgent and/or obvious issue
  • When there is little variability or uncertainty

6
Myths Barriers to Use of Methods
  • Myth it takes more resources to do uncertainty
    analysis, we have deadlines, we dont know what
    to do with it, lets just go with what we have
  • Hypothesis 1 poorly informed decisions based
    upon misleading deterministic/point estimates can
    be very costly, leading to a longer term and
    larger resource allocation to correct mistakes
    that could have been avoided or to find better
    solutions
  • Hypothesis 2 Uncertainty analysis helps to
    determine when a robust decision can be made
    versus when more information is needed first
  • Hypothesis 3 Uncertainty and sensitivity
    analysis help identify key weaknesses and focus
    limited resources to help improve estimates
  • Hypothesis 4 Doing uncertainty analysis
    actually reduces overall resource requirements,
    especially if it is integrated into the process
    of model development and applications

7
Role of Modeling in Decision-Making
  • Modeling should provide insight
  • Modeling should help inform a decision
  • Modeling should be in response to clearly defined
    objectives that are relevant to a decision.

8
Questions that Decision-Makers and Stakeholders
Typically Ask
  • How well do we know these numbers?
  • What is the precision of the estimates?
  • Is there a systematic error (bias) in the
    estimates?
  • Are the estimates based upon measurements,
    modeling, or expert judgment?
  • How significant are differences between two
    alternatives?
  • How significant are apparent trends over time?
  • How effective are proposed control or management
    strategies?
  • What is the key source of uncertainty in these
    numbers?
  • How can uncertainty be reduced?

9
Application of Uncertainty to Decision Making
  • Risk preference
  • Risk averse
  • Risk neutral
  • Risk seeking
  • Utility theory
  • Benefits of quantifying uncertainty Expected
    Value of Including Uncertainty
  • Benefits of reducing uncertainty Expected Value
    of Perfect Information (and others)

10
Variability and Uncertainty
  • Variability refers to the certainty that
  • different members of a population will have
    different values (inter-individual variability)
  • values will vary over time for a given member of
    the population (intra-individual variability)
  • Uncertainty refers to lack of knowledge
    regarding
  • True value of a fixed but unknown quantity
  • True population distribution for variability
  • Both depend on averaging time

11
Variability and Uncertainty
  • Sources of Variability
  • Stochasticity
  • Periodicity, seasonality
  • Mixtures of subpopulations
  • Variation that could be explained with better
    models
  • Variation that could be reduced through control
    measures

12
Variability and Uncertainty
  • Sources of Uncertainty
  • Random sampling error for a random sample of data
  • Measurement errors
  • Systematic error (bias, lack of accuracy)
  • Random error (imprecision)
  • Non-representativeness
  • Not a random sample, leading to bias in mean
    (e.g., only measured loads not typical of daily
    operations)
  • Direct monitoring versus infrequent sampling
    versus estimation, averaging time
  • Omissions
  • Surrogate data (analogies with similar sources)
  • Lack of relevant data
  • Problem and scenario specification
  • Modeling

13
Overview of State of the Science
  • Statistical Methods Based Upon Empirical Data
  • Statistical Methods Based Upon Judgment
  • Other Quantitative Methods
  • Qualitative Methods
  • Sensitivity Analysis
  • Scenario Uncertainty
  • Model Uncertainty
  • Communication
  • Decision Analysis

14
Statistical MethodsBased Upon Empirical Data
  • Frequentist, classical
  • Statistical inference from sample data
  • Parametric approaches
  • Parameter estimation
  • Goodness-of-fit
  • Nonparametric approaches
  • Mixture distributions
  • Censored data
  • Dependencies, correlations, deconvolution

15
Statistical MethodsBased Upon Empirical Data
  • Variability and Uncertainty
  • Sampling distributions for parameters
  • Analytical solutions
  • Bootstrap simulation

16
Propagating Variability and Uncertainty
  • Analytical techniques
  • Exact solutions (limited applicability)
  • Approximate solutions
  • Numerical methods
  • Monte Carlo
  • Latin Hypercube Sampling
  • Other sampling methods (e.g., Hammersley,
    Importance, stochastic response surface method,
    Fourier Amplitude Sensitivity Test, Sobols
    method, Quasi-Monte Carlo methods, etc.)

17
Monte Carlo Simulation
  • Probabilistic approaches are widely used
  • Monte Carlo (and similar types of) simulation are
    widely used.
  • Why?
  • Extremely flexible
  • Inputs
  • Models
  • Relatively straightforward to conceptualize

18
Tiered Approach to Analysis
  • Purpose of Analyses (examples)
  • Screening to prioritize resources
  • Regulatory decision-making
  • Research planning
  • Types of Analyses
  • Screening level point-estimates
  • Sensitivity Analysis
  • One-Dimensional Probabilistic Analysis
  • Two-Dimensional Probabilistic Analysis
  • Non-probabilistic approaches

19
MethodsBased Upon Expert Judgment
  • Expert Elicitation
  • Heuristics and Biases
  • Availability
  • Anchoring and Adjustment
  • Representativeness
  • Others (e.g., Motivational, Expert, etc.)
  • Elicitation Protocols
  • Motivating the expert
  • Structuring
  • Conditioning
  • Encoding
  • Verification
  • Documentation
  • Individuals and Groups
  • When Experts Diasagree

20
An Example of Elicitation ProtocolsStanford/SRI
Protocol
21
Key Ongoing Challenges
  • Expert Judgment vs. Data
  • Perception that judgment is more biased than
    analysis of available data
  • Unless data are exactly representative, they too
    could be biased
  • Statistical methods are objective in that the
    results can be reproduced by others, but this
    does not guarantee absence of bias
  • A key area for moving forward is to agree on
    conditions under which expert judgment is an
    acceptable basis for subjective probability
    distributions, even for rulemaking situations

22
Appropriate Use of Expert Judgment in Regulatory
Decision Making
  • There are examplese.g.,
  • analysis of health effects for EPA standards
  • Uncertainty in benefit/cost analysis (EPA, OMB)
  • Probabilistic risk analysis of nuclear facilities
  • Key components of credible use of expert
    judgment
  • Follow a clear and appropriate protocol for
    selecting experts and for elicitation
  • For the conditioning step, consider obtaining
    input via workshop, but for encoding, work
    individually with experts preferably at their
    location
  • Document (explain) the basis for each judgment
  • Compare judgments identify key similarities and
    differences
  • Evaluate the implications of apparent differences
    with respect to decision objectives do not
    combine judgments without first doing this
  • Where possible, allow for iteration

23
Statistical MethodsBased Upon Expert Judgment
  • Bayesian methods can incorporate expert judgment
  • Prior distribution
  • Update with data using likelihood function and
    Bayes Theorem
  • Create a posterior distribution
  • Bayesian methods can also deal with various
    complex situations
  • Conditional probabilities (dependencies)
  • Combining information from multiple sources
  • Appears to be very flexible
  • Computationally, can be very complex
  • Complexity is a barrier to more widespread use

24
Other Quantitative Methods
  • Interval Methods
  • Simple intervals
  • Probability bounds
  • Produce optimally narrow bounds cannot be any
    narrower and still enclose all possible outcomes,
    including dependencies among inputs
  • Bounds can be very wide in comparison to
    confidence intervals

25
Other Quantitative Methods
  • Fuzzy methods
  • Representation of vagueness, rather than
    uncertainty
  • Approximate/semi-quantitative
  • Has been applied in many fields
  • Meta-analysis
  • Quantitatively combine, synthesize, and summarize
    data and results from different sources
  • Requires assessment of homogeneity among studies
    prior to combining
  • Produces data with larger sample sizes than the
    constituent inputs
  • Can be applied to summary data
  • If raw data are available, other methods may be
    preferred

26
Scenario Uncertainty
  • A need for formal methods
  • Creativity, brainstorming, imagination
  • Key dimensions (e.g., human exposure assessment)
  • Pollutants
  • Transport pathways
  • Exposure routes
  • Susceptible populations
  • Averaging time
  • Geographic extent
  • Time Periods
  • Activity Patterns
  • Which dimensions/combinations matter, which ones
    dont?
  • Uncertainty associated with mis-specification of
    a scenario systematic error
  • Scenario definition should be considered when
    developing and applying models

27
Model Uncertainty
  • Model Boundaries (related to scenario)
  • Simplifications
  • Aggregation
  • Exclusion
  • Resolution
  • Structure
  • Calibration
  • Validation, Partial validation
  • Extrapolation

28
Model Uncertainty
  • Methods for Dealing with Model Uncertainty
  • Compare alternative models, but do not combine
  • Weight predictions of alternative models (e.g.,
    probability trees)
  • Meta-models that degenerate into alternative
    models (e.g., Y a(x-t)n , where n determines
    linear/nonlinear and t determines threshold or
    not)

29
Weighting vs. Averaging
Each Model has Equal Weight
Model B
Model A
Probability Density
Output of Interest
Average of Both Models
Neither Model Supports This Range of Outcomes
Probability Density
Output of Interest
30
Sensitivity Analysis
  • Objectives of Sensitivity Analysis (examples)
  • Help identify key sources of variability (to aid
    management strategy)
  • Critical control points?
  • Critical limits?
  • Help identify key sources of uncertainty (to
    prioritize additional data collection to reduce
    uncertainty)
  • What causes worst/best outcomes?
  • Evaluate model behavior to assist
    verification/validation
  • To assist in process of model development
  • Local vs. Global Sensitivity Analysis
  • Model Dependent vs. Model Independent Sensitivity
    Analysis
  • Applicability of methods often depends upon
    characteristics of a model (e.g., nonlinear,
    thresholds, categorical inputs, etc.)

31
Examples of Sensitivity Analysis Methods
  • Mathematical Methods
  • Assess sensitivity of a model output to the range
    of variation of an input.
  • Statistical Methods
  • Effect of variance in inputs on the output
    distribution.
  • Graphical Methods
  • Representation of sensitivity in the form of
    graphs, charts, or surfaces.

32
Sensitivity Analysis Methods (Examples)
  • Nominal Range Sensitivity Analysis
  • Differential Sensitivity Analysis
  • Conditional Analysis
  • Correlation coefficients (sample, rank)
  • Linear regression (sample, rank, variety of basis
    functions possible)
  • Other regression methods
  • Analysis of Variance (ANOVA)
  • Categorical and Regression Trees (CART) (a.k.a.
    Hierarchical Tree-Based Regression)
  • Sobols method
  • Fourier Amplitude Sensitivity Test (FAST)
  • Mutual Information Index
  • Scatter Plots

33
Sensitivity Analysis Displays/Summaries
  • Scatter plots
  • Line plots/conditional analyses
  • Radar plots
  • Distributions (for uncertainty or variability in
    sensitivity)
  • Summary statistics
  • Categorical and regression trees
  • Apportionment of variance

34
Guidance on Sensitivity Analysis
  • Guidance for Practitioners, with a focus on food
    safety process risk models (Frey et al., 2004)
  • When to perform sensitivity analysis
  • Information needed depending upon objectives
  • Preparation of existing or new models
  • Defining the case study/scenarios
  • Selection of sensitivity analysis methods
  • Procedures for application of methods
  • Presentation and interpretion of results

35
Summary of Evaluation Results for Selected
Sensitivity Analysis Methods
36
Example of Guidance on Selection of Sensitivity
Analysis Methods
Source Frey et al., 2004, www.ce.ncsu.edu/risk/
37
Example of Guidance on Selection of Sensitivity
Analysis Methods
38
Communication
  • Case Studies (scenarios)
  • Graphical Methods
  • Influence Diagrams
  • Decision Tree
  • Others
  • Summary statistics/data
  • Evaluation of effectiveness of methods for
    communication (e.g., Bloom et al., 1993 Ibrekk
    and Morgan, 1987)

39
Example Case Studies
  • Technology Assessment
  • Emission Factors and Inventories
  • Air Quality Modeling
  • Risk Assessment

40
Role of Technology Assessment in Regulatory
Processes (Examples)
  • Assessment of ability of technology to achieve
    desired regulatory or policy goals (emissions
    control, safety, efficiency, etc.)
  • Evaluation of regulatory alternatives (e.g.,
    based on model cost estimates)
  • Regulatory Impact Analysis assessment of costs

41
An Example of Federal Decision MakingProcess
Technology RDD
42
A Probabilistic Framework for FederalProcess
Technology Decision-Making
43
Methodology for Probabilistic Technology
Assessment
  • Process simulation of process technologies in
    probabilistic frameworks
  • Integrated Environmental Control Model (IECM) and
    derivatives
  • Probabilistic capability for ASPEN chemical
    process simulator
  • Quantification of uncertainty in model inputs
  • Statistical analysis
  • Elicitation of expert judgment
  • Monte Carlo simulation
  • Statistical methods for sensitivity analysis
  • Decision tree approach to comparing technologies
    and evaluating benefits of additional research

44
Conceptual Diagram of Probabilistic Modeling
Engineering Performance and Cost Model of a New
Process Technology
Input Uncertainties
Output Uncertainties
Performance
Performance Inputs
Emissions
Cost Inputs
Cost
45
Comparison of Probabilistic and Point-Estimate
Results for an IGCC System
46
Example of a Probabilistic Comparison of
Technology Options
Uncertainty in the difference in cost between two
technologies, taking into account correlations
between them
47
Example Engineering Study ofCoal-Gasification
Systems
  • DOE/METC Engineers
  • Briefing Packets
  • - Part 1 Uncertainty Analysis (9 pages)
  • - Part 2 Process Area Technical Background
  • - Lurgi Gasifier 12 p., 16 ref.
  • - KRW Gasifier 19 p., 25 ref.
  • - Desulfurization 9 p., 19 ref.
  • - Gas Turbine 23 p., 36 ref.
  • - Part 3 Questionnaire
  • Follow-Up

48
Examples of the Judgmentsof One Expert
Fines Carryover
Carbon Retention
Air/Coal Ratio
49
Examples of the Judgmentsof Multiple Experts
50
Do Different Judgments Really Matter?
Specific Sources of Disagreement
- Sorbent Loading - Sorbent
Attrition Qualitative Agreement in Several
Cases
51
Technology AssessmentFindings (1)
  • Interactions among uncertain inputs, and
    nonlinearities in model, contribute to positive
    skewness in model output uncertainties
  • Uncertainties in inputs are often positively
    skewed (physical, non-negative quantities)
  • The mean value of a probabilistic estimate is
    often worse (lower performance, higher cost)
    than the best guess deterministic estimate, and
    the probability of worse outcomes is typically
    greater than 50 percent.
  • A system approach is needed to account for
    interactions among process areas
  • Deterministic analysis leads to apparent cost
    growth and performance shortfall because it
    does not account for simultaneous interactions
    among positively skewed inputs
  • Uncertainty analysis requires more thought
    pertaining to developing input assumptions, but
    provides more insight into potential sources of
    cost growth and performance shortfall

52
Technology AssessmentFindings (2)
  • A decision model provides a framework for
    evaluating judgments regarding the outcomes of
    additional research and prioritizing additional
    research
  • Able to quantify the probability of pay-offs as
    well as downside risks
  • Able to compare competing options under
    uncertainty and identify robust choices
  • Trade-offs when comparing technologies can be
    evaluated probabilistically (e.g., efficiency,
    emissions, and cost)
  • It is possible to combine approaches for
    quantifying uncertainty in one assessment,
    consistent with objectives

53
Technology AssessmentFindings (3)
  • Thinking about uncertainties leads to better
    understanding of what matters most in the
    assessment
  • Often, only a relatively small number of inputs
    contribute substantially to uncertainty in a
    model output
  • Reducing uncertainty in only a few key inputs can
    substantially reduce downside risk and increase
    the pay-offs of new technology
  • Conversely, for those inputs to which the output
    is not sensitive, it is not critical to devote
    resources to refinement
  • When basing inputs on expert judgments, only
    those disagreements that really matter to the
    decision need become the focus of further
    discussion and evaluation
  • Bottom Line Probabilistic analysis helps
    improve decisions and avoid unpleasant
    surprises.

54
Emission Factors and Inventories
  • Significance to Regulatory Processes
  • Assessment of capability of technology to
    reduce/prevent emissions
  • Evaluation of regulatory alternatives
  • Regulatory Impact Analysis including
    benefit/cost analysis
  • Component of air quality management at various
    temporal and spatial scales
  • Component of human and ecological exposure and
    risk assessment
  • Modeling Aspects
  • Some emission factors are estimated using models
  • Emission Inventories are linear models
  • Specialized models for some emission factors and
    inventories (e.g., Mobile6, NONROAD, MOVES)

55
Motivations for Probabilistic Emission Factors
and Inventories
  • How good are the estimates?
  • What are the key sources of uncertainty in the
    estimates that should be targeted for
    improvement?
  • Likelihood of meeting an emissions budget?
  • Which emission sources are the most significant?
  • What is the inter-unit variability in emissions?
  • What is the uncertainty in mean emissions for a
    group/fleet of sources?
  • What are the implications of uncertainty in
    emissions for air quality management, risk
    management of human exposures?
  • Consideration of geographic extent and averaging
    time
  • Estimation for future scenarios versus
    retrospective estimates of past emissions or
    assessment of current emissions

56
Motivations for Probabilistic Analysis
  • That a perfect assessment of uncertainty cannot
    be done, however, should not stop researchers
    from estimating the uncertainties that can be
    addressed quantitatively (p. 150, NRC, 2000)
  • EPA, along with other agencies and industries,
    should undertake the necessary measures to
    conduct quantitative uncertainty analyses of the
    mobile-source emissions models in the modeling
    toolkit. (p. 166, NRC, 2000)

57
Current Practice for Qualifying Uncertaintyin
Emission Factors and Inventories
  • Qualitative ratings for emission factors (AP-42)
  • Data Attribute Rating System (DARS) (not really
    used in practice)
  • Both methods are qualitative
  • No quantitative interpretation
  • Some sources of uncertainty (i.e.
    non-representativeness) difficult to quantify
  • Qualitative methods can complement quantitative
    methods

58
Statistical Methodological Approach
  • Compilation and evaluation of database
  • Visualization of data by developing empirical
    cumulative distribution functions
  • Fitting, evaluation, and selection of alternative
    probability distribution models
  • Characterization of uncertainty in the
    distributions for variability (e.g., uncertainty
    in the mean)
  • Propagation of uncertainty in activity and
    emissions factors to estimate uncertainty in
    total emissions
  • Calculation of importance of uncertainty

59
Summary of Approaches to Emission Factor and
Inventory Uncertainty
  • Probabilistic Methods
  • Empirical, Parametric
  • Mixture distributions
  • Censored distributions (non-detects)
  • Vector autoregressive time series (intra- and
    inter-unit correlation)
  • Bootstrap simulation
  • Expert Judgment
  • Monte Carlo simulation
  • Sensitivity analysis
  • Software tools
  • AUVEE Analysis of Uncertainty and Variability
    in Emissions Estimation
  • AuvTool standalone software

60
Summary of Probabilistic Emissions Case Studies
at NCSU
  • Case Studies (examples)
  • Point sources
  • Power Plants
  • Natural gas-fired engines (e.g., compressor
    stations)
  • Mobile sources
  • On-Road Highway Vehicles
  • Non-Road Vehicles (e.g., Lawn Garden,
    Construction, Farm, Industrial)
  • Area sources
  • Consumer/Commercial Product Use
  • Natural Gas-Fueled Internal Combustion Engines
  • Gasoline Terminal Loading Loss
  • Cutback Asphalt Paving
  • Architectural Coatings
  • Wood Furniture Coatings
  • Pollutants
  • NOx
  • VOC
  • Urban air toxics (e.g., Houston case study)

61
Example Results Lawn Garden Equipment
Based on Frey and Bammi (2002)
62
Probabilistic CO Emission Factors for On-Road
Light Duty Gasoline Vehicles (Mobile5)
Based on Frey and Zheng (2002)
63
MOVES
  • Conceptual Basis for MOVES, the successor to
    Mobile6 and NONROAD (www.epa.gov/otaq/ngm.htm)
  • Shootout
  • NCSU report on modal/binning approach
  • NCSU recommended approaches for quantification of
    inter-vehicle variability and fleet average
    uncertainty in modal emission rates and estimates
    of emissions for driving cycles (details in our
    report to EPA)
  • EPA requested further assessment of an
    approximate analytical procedure for propagating
    error (report by Frey to EPA)
  • At last report, EPA was considering Monte Carlo
    simulation

64
Probabilistic AP-42 Emission Factors for Natural
Gas-fueled Engines (July 2000 Version)
aUnits are lb/106 BTU. bMLE is used for 2SLB
engine, MoMM is used for 4SLB engine,
WWeibull distribution, GGamma
distribution. cCalculated based upon bootstrap
simulation results.
Based on Frey and Li (2003) (submitted)
65
Summary of Probabilistic Emission Inventories for
Selected Air Toxics
66
Key Sources of Uncertainty
67
Example of Benzene Emission Factor Category 3b
Nonwinter Storage Losses at a Bulk Terminal
Empirical Distribution
68
Example of Benzene Emission Factor Category 3b
Fitted Lognormal Distribution
69
Example of Benzene Emission Factor Category 3b
Confidence Interval in the CDF
70
Example of Benzene Emission Factor Category 3b
Uncertainty in the Mean
0.06
Uncertainty in mean -73 to 200
71
Using AuvTool to Fit a Distribution for
Variability
72
Using AuvTool for Bootstrap Simulation
73
Using AuvTool to Quantify Uncertainty in the Mean
74
Uncertainty in Total Emission InventoryAUVEE
Prototype Software
75
Summary ofProbabilistic Emission Inventory
76
Identification of Key Sources of Uncertaintyin
an Inventory
77
Detection Limits and Air Toxic Emission Factor
Data
  • Many air toxic emission factor data contain one
    or more measurements below a detection limit
  • Detection limits can be unique to each
    measurement because of differences in sample
    volumes and analytical chemistry methods among
    sites or contractors
  • A database can contain some non-detected data
    with detection limits larger than detected values
    measured at other sites

78
Methodology Conventional Methods for Censored
Data
  • Conventional approaches to estimate the mean
  • -Remove non detected values (biased)
  • -Replace values below DL with zero
    (underestimate)
  • -Replace values below DL with DL/2 (biased)
  • -Replace values below DL with DL
    (overestimate)
  • Cause biased estimates of the mean
  • Does not provide adequate insights regarding
  • Population distribution
  • Unbiased statistics
  • Uncertainty in statistics

79
Methodology Quantification of the Inter-Unit
Variability in Censored Data
  • Maximum Likelihood Estimation (MLE) is used to
    fit parametric distributions to censored data
  • MLE is asymptotically unbiased
  • Fitted distribution is the best estimate of
    variability
  • Can estimate mean and other statistics from the
    fitted distribution
  • Can quantify uncertainty caused by random
    sampling error using Bootstrap simulation

80
Results Fitted Lognormal Distribution, No
Censoring

81
Results Fitted Lognormal Distribution, 30
Censoring
82
Results Fitted Lognormal Distribution, 60
Censoring

83
Example Case Study Formaldehyde Emission Factor
from External Coal Combustion
  • 14 data points including 5 censored values
  • Each censored data point has a different
    detection limit
  • Some detected data values are less than some
    detection limits
  • There is uncertainty regarding the empirical
    cumulative probability of such detected data
    values

84
Results of Example Case Study Empirical
Cumulative Probability
85
Results of Example Case Study Lognormal
Distribution Representing Inter-Unit Variability
86
Results of Example Case Uncertainty in
Inter-Unit Variability
87
Results of Example Case Uncertainty in the Mean
(Basis to Develop Probabilistic Emission
Inventory)
Uncertainty in mean -77 to 208
88
Mixtures of Distributions
Percent of data in 50 CI 92 Percent of data
in 95 CI 100
89
Case Study
  • Charlotte modeling domain
  • 32 units from 9 different coal-fired power plants
  • 1995 and 1998 data used
  • Propagation of uncertainty investigated using
    July 12 July 16 1995 meteorological data
  • Data available for emission and activity factors
  • Vector autoregressive time-series modeling of
    emissions from each unit

90
Time Series and Uncertainty
Different uncertainty ranges for different hours
of day
91
Emission Factors and InventoriesFindings (1)
  • Visualization of data used to develop an
    inventory is highly informative to choices of
    empirical or parametric distribution models for
    quantification of variability or uncertainty
  • A key difficulty in developing probabilistic
    emission factors inventories is to find the
    original data used by EPA and others.
  • When data are found, they are typically poorly
    documented.
  • The time required to assemble databases when
    original data could not be found was substantial
  • Test methods used for some emission sources are
    not representative of real world operation,
    implying the need for real world data and/or
    expert judgment when estimating uncertainty
  • Uncertainty in measurement methods is not
    adequately reported. There is a need for more
    systematic reporting of the precision and
    accuracy of measurement/test methods
  • Emissions databases should not be arbitrarily
    fragmented into too many subcategories.
    Conversely, subcategories should be created when
    there is a good (empirical) basis for doing so.

92
Emission Factors and InventoriesFindings (2)
  • Uncertainties in emission factors are typically
    positively skewed, unless the uncertainties are
    relatively small (e.g., less than about plus or
    minus 30 percent)
  • Uncertainty estimates might be sensitive to the
    choice of parametric distribution models if there
    is variation in the goodness-of-fit among the
    alternatives. However, in such cases, there is
    typically a preferred best fit. When several
    alternative models provide equivalent fits,
    results are not sensitive to the choice of the
    model
  • The quantifiable portion of uncertainty
    attributable to random sampling error can be
    large and should be accounted for when using
    emission factors and inventories
  • Variability in emissions among units could be a
    basis for assessing the potential of emissions
    trading programs

93
Emission Factors and InventoriesFindings (3)
  • Intra-unit dependence in hourly emissions is
    significant for some sources (e.g., power
    plants), including hourly and daily lag effects
  • Inter-unit dependence in emissions is important
    for some sources, such as power plants
  • Range of variability and uncertainty is typically
    much greater as the averaging time decreases
  • Even for sources with continuous emissions
    monitoring data, there is uncertainty regarding
    predictions of future emissions that can be
    informed by analysis of historical data
  • Prototype software demonstrates the feasibility
    of increasing the convenience of performing
    probabilistic analysis
  • Uncertainties in total inventories are often
    attributable to just a few key emission sources

94
Mobile5 and Mobile6Findings
  • Range of variability and uncertainty in
    correction factors (e.g., temperature) dominate
    and are large
  • Uncertainties in average emissions are large in
    some cases (e.g., -80 to 220 percent)
    normality assumptions are not valid
  • Asymmetry in uncertainties associated with
    non-negative quantities and large inter-unit
    variability
  • Sensitivity analysis was used to identify key
    sources of uncertainty and recommend future data
    collection priorities in order to reduce
    uncertainty
  • There is a proliferation of driving cycles. Some
    of them are redundant and, therefore, unnecessary
  • When comparing model predictions with validation
    data, or when comparing models with each other,
    their prediction uncertainty ranges should be
    considered
  • It is difficult to do an uncertainty analysis for
    a model such as Mobile5 or Mobile6 after the
    fact, but would be much easier if integrated into
    the data management and modeling approach.

95
Air Quality Modeling
  • Widely used for
  • Assessment of regulatory options
  • Identification and evaluation of control
    strategies (which pollutants, how much control,
    where to control?)
  • Emissions permit applications
  • Identification of contributing factors to local,
    urban, and regional air quality problems
  • Human exposure assessment

96
PROBABILISTIC MODELING
Input Uncertainties
Output Uncertainties
Emissions
Peak Ozone
Chemistry
Variable-Grid Urban Airshed Model (UAM-V)
Local Ozone
Meteorology
Local NOx
Initial Boundary Conditions
Local VOC
97
Case Study of Hanna et al. (2001)
  • OTAG Modeling Domain (eastern U.S.)
  • UAM-V model with Carbon Bond-IV mechanism
  • Uncertainty in peak ozone concentrations
  • Assessment of effects of 50 reductions in each
    of NOx and VOC emissions
  • Quantification of uncertainty in 128 inputs
  • Literature review for chemical kinetic rate
    constants
  • Expert elicitation for emissions, meteorological,
    and initial and boundary condition inputs
  • Monte Carlo simulation with n100
  • Correlations used to identify key sources of
    uncertainty

98
Key Findings of Case Study of Hanna et al. (2001)
  • It was feasible to perform Monte Carlo simulation
    on UAM-V for the OTAG domain and a seven day
    episode
  • Simulation results include base case uncertainty
    estimates for ozone concentrations and estimates
    of differences in ozone concentrations because of
    emissions reduction strategies
  • There was less uncertainty in estimates of
    differences in concentration than in absolute
    estimates of total concentration
  • Reductions in NOx emissions led to higher
    estimated reductions in O3 than did reductions in
    VOC emissions. This is consistent with the
    expectation that most of the domain is
    NOx-limited.
  • Key uncertainties include NOx photolysis rate,
    several meteorological variables, and biogenic
    VOC emissions
  • Compared to Hanna et al. (1998), there was more
    disaggregation of uncertainty estimates for
    emission sources, and this may tend to weaken the
    sensitivity to any one source. It is possible,
    however, that model outputs would be more
    sensitive to an aggregated collection of
    emissions sources.
  • There is a need for improved methods of
    uncertainty estimation, particularly for the
    chemical mechanism and the meteorological fields,
    and for better accounting of correlations and
    dependencies (e.g., temperature dependence of
    biogenic emissions).

99
Case Study of Abdel-Aziz and Frey (2004)
  • Focus was on evaluating implications of
    uncertainties in hourly coal-fired power plant
    NOx emissions with respect to ozone for the
    Charlotte, NC domain
  • Key questions
  • (1) what is the uncertainty in ozone predictions
    solely attributable to uncertainty in coal-fired
    utility NOx emissions?
  • (2) can uncertainties in maximum ozone levels be
    attributed to specific power plant units?
  • (3) how likely is it that National Ambient Air
    Quality Standards (NAAQS) will be exceeded? and
  • (4) how important is it to account for inter-unit
    correlation in emission uncertainties

100
Probability of Exceeding NAAQS Comparison of
1-hour and 8-hour Standards

101
Location of Power Plant Impact
Analysis of Correlation in Emissions versus Ozone
Levels in a Specific Grid Cell Can Detect
Influence of a Specific Plant

102
Key Findings from Abdel-Aziz and Frey (2004) study
  • The uncertainty in maximum 1-hour ozone
    predictions is potentially large enough to create
    ambiguity regarding compliance with the NAAQS for
    any given emissions management strategy.
  • Control strategies can be developed to achieve
    attainment with an acceptable degree of
    confidence, such as 90 or 95 percent.
  • There was a substantial difference in results
    when comparing independent versus dependent
    units thus, it can be important to a decision
    to account for dependencies between units
  • Probabilistic air quality modeling results
    provide insight regarding where to site
    monitoring stations and regarding the number of
    stations needed
  • Under the old 1-hour standard, uncertainties in
    the maximum domain-wide ozone levels could be
    traced to an individual power plant, thereby
    implying that control strategies must include
    that plant.
  • Under the new 8-hour standard, uncertainties in
    maximum ozone levels are attributable to many
    plants, implying the need for a more widespread
    control strategy

103
Risk Assessment Modeling
  • Risk assessment is growing in importance as a
    basis for regulatory decision making
  • E.g., Phase 2 of MACT standards
  • Urban air toxics
  • Food safety and international trade
  • (etc.)

104
Human Exposure and Risk Analysis
  • Over the last 10-15 years, there has been growing
    acceptance and incorporation of probabilistic
    approaches to dealing with inter-individual
    variability and uncertainty
  • EPA has issued various guidance
  • International guidance e.g., FAO/WHO

105
Example of Probabilistic Techniques in an
Exposure and Risk Model SHEDS
106
Listeria Monocytogenes Model
107
E. Coli O157 in Ground Beef Risk Assessment Model
Scope Hazard Identification
  • Slaughter
  • Production
  • Preparation
  • Dose-response

Dose-Response Assessment Morbidity Mortality
108
Findings Based Upon Risk Assessment
  • Risk assessment applications often differ from
    others because of distinction between
    inter-individual variability and uncertainty
  • A two-dimensional probabilistic simulation
    framework is used
  • Expert judgment is inherent in the process of
    fitting distributions to data for variability and
    is often used to estimate uncertainty in the
    parameters of such distributions
  • Sensitivity analysis is critical to
    interpretation of risk assessment results
  • Assists in risk management decision-making
  • Prioritize future work to improve the assessment
  • There was a lack of practical guidance regarding
    sensitivity analysis of risk models that has been
    addressed by recent work

109
General Recommendations (1)
  • Uncertainty and sensitivity analysis should be
    used to answer key decision maker and stakeholder
    questions, e.g.,
  • prioritize scarce resources toward additional
    research or data collection
  • make choices among alternatives in the face of
    uncertainty,
  • evaluate trends over time, etc.
  • Where relevant to decision making, uncertainty
    and sensitivity analysis should be included as
    functional requirements from the beginning and
    incorporated into model and input data
    development
  • There should be minimum reporting requirements
    for uncertainty in data (e.g., summary statistics
    such as mean, standard deviation, sample size)
  • Federal agencies should continue to improve
    documentation and accessibility of models and
    data for public peer review

110
General Recommendations (2)
  • Foster greater acceptance of appropriate methods
    for including, documenting, and reviewing expert
    judgment in regulatory-motivated modeling and
    analysis
  • There is a need for flexibility since there are
    many possible approaches to analysis of
    uncertainty and sensitivity. Specific choices
    should be appropriate to assessment objectives,
    which are typically context-specific
  • Human resources for modeling, including
    uncertainty and sensitivity analysis, should be
    appropriately committed.
  • Adequate time and budget to do the job right the
    first time (could save time and money in the long
    run)
  • Adequate training and peer review
  • Promote workshops and other training
    opportunities, and periodic refinement of
    authoritative compilations of techniques and
    recommended practice

111
General Recommendations (3)
  • Software tools substantially facilitate both
    uncertainty and sensitivity analysis (e.g.,
    Crystal Ball) but in some ways also limit what is
    done in practice there is a long-term need for
    software tools appropriate to specific types of
    applications
  • Some areas need more research e.g., best
    techniques for communication, real-world
    information needs for decision makers
  • The relevance of analyses to decision making
    needs to be emphasized and considered by analysts
  • Decision makers need or should have access to
    information on why/how they should use
    probabilistic results
  • A multi-disciplinary compilation of relevant case
    studies and insights from them is a useful way to
    help convince others of the value of doing
    uncertainty and sensitivity analysis
  • Uncertainty and sensitivity analysis should be an
    open and transparent process that can be subject
    to scrutiny and peer review
Write a Comment
User Comments (0)
About PowerShow.com