Reasoning and Decision Making Under Uncertainty - PowerPoint PPT Presentation

About This Presentation
Title:

Reasoning and Decision Making Under Uncertainty

Description:

Fred-Complains: Fred complains about his grade (short FC, has states true and false) ... Fred/John Complains Problem. Problem 12 Assignment3 Fall 2002 ... – PowerPoint PPT presentation

Number of Views:58
Avg rating:3.0/5.0
Slides: 16
Provided by: Christo9
Learn more at: https://www2.cs.uh.edu
Category:

less

Transcript and Presenter's Notes

Title: Reasoning and Decision Making Under Uncertainty


1
Reasoning and Decision Making Under Uncertainty
  • Uncertainty, Rules of Probability
  • Bayes Theorem and Naïve Bayesian Systems
  • Bayesian Belief Networks
  • Structure and Concepts
  • D-Separation
  • How do they compute probabilities?
  • How to design BBN using simple examples
  • Other capabilities of Belief Network short!
  • Applications of BBN transparencies 135ff
  • Netica Demo
  • Develop a BBN for HD homework

2
Causes of not knowing thingsprecisely
Default Logic and Reasoning
Incompleteness
Uncertainty
Belief Networks
Vagueness
Bayesian Technology
Fuzzy Sets
3
Example 2
BN probability of a variable only depends on its
direct successors e.g. P(b,e,a,j,m)
P(b)P(e)P(ab,e)P(ja)P(ma)0.010.020.950
.10.7
4
Basic Properties of Belief Networks
  • Simplifying Assumption Let X1,,Xn be the
    variables of a belief network and all variables
    have binary states
  • P(X1,,Xn) P P(XiParents(Xi)) allows to
    compute all atomic events
  • P(X1,,Xp-1) P(X1,,Xp-1,Xp) P(X1,,Xp-1,Xp)
  • P(XY) a P(X,Y) where a 1/P(Y)
  • Remark These 3 equations are sufficient to
    compute any probability in a belief network
    however, using this approach is highly
    inefficient e.g. with n20 computing P(X1X2)
    would require the addition of 218219
    probabilities. Therefore, more efficient ways to
    compute probabilities are needed e.g. if X1 and
    X2 are independent, only P(X1) needs to be
    computed. Another way to speedup computations is
    using probabilities that are already known and do
    not need to be computed and taking advantage of
    the fact that probabilities add up to 1

n
i1
5
Fred Complains / John Complains Problem
  • Assume that John and Fred are students taking
    courses together for which they receive a grade
    of A, B, or C. Moreover, sometimes Fred and John
    complain about their grades. Assume you have to
    model this information using a belief network
    that consists of the following variables
  • Grade-John Johns grade for the course (short
    GJ, has states A, B, and C)
  • Grade-Fred Johns grade for the course (short
    GF, has states A, B, and C)
  • Fred-Complains Fred complains about his grade
    (short FC, has states true and false)
  • John-Complains John complains about his grade
    (short JC, has states true and false)
  • If Fred gets an A in the course he never
    complains about the grade if he gets a B he
    complains about the grade in 50 of the cases, if
    he gets a C he always complains about the grade.
    If Fred does not complain, then John does not
    complain. If Johns grade is A, he also does not
    complain. If, on the other hand, Fred complains
    and Johns grade is B or C, then John also
    complains. Moreover P(GJA)0.1, P(GJB)0.8,
    P(GJC)0.1 and P(GFA)0.2, P(GFB)0.6,
    P(GFC)0.2.
  • Design the structure of a belief network
    including probability table that involves the
    above variables (if there are probabilities
    missing make up your own probabilities using
    common sense)
  • Using your results from the previous step,
    compute P(GFCJCtrue) by hand! Indicate every
    step that is used in your computations and
    justify transformation you apply when computing
    probabilities!

6
Example FC/JC Network Design
GF
GJ
  1. Specify Nodes and States
  2. Specify Links
  3. Determine Probability Tables
  4. Use Belief Network

FC
JC
  • Nodes GF and GJ have states A,B,C
  • Nodes FC and JC have states true,false
  • Notations in the following,
  • we use FC as a short notation for FCtrue and
  • Use FC as a short notation for FCfalse
  • Similarly, we use JC as a short notation for
    JCtrue and
  • Use JC as a short notation for JCfalse.
  • We also write P(A,B) for P(A ?B).

7
Example FC/JC Network Design
GF
GJ
  1. Specify Nodes and States
  2. Specify Links
  3. Determine Probability Tables
  4. Use Belief Network

FC
JC
  • Next probability tables have to be specified for
    each node in the network for
  • each value of a variable conditional
    probabilities have to be specified that
  • depend on the variables of the parents of the
    node for that above
  • example these probabilities are P(GF), P(GJ),
    P(FCGF), P(JCFC,GJ)
  • P(GJA)0.1, P(GJB)0.8, P(GJC)0.1
  • P(GFA)0.2, P(GFB)0.6, P(GFC)0.2
  • P(FCGFA)0, P(FCGFB)0.5, P(FCGFC)1
  • P(JCGJA,FC)0, P(JCGJA,FC)0,
    P(JCGJB,FC)1,
  • P(JCGJB,FC)0, P(JCGJC,FC)1,
    P(JCGJC,FC)0.

8
D-Separation
  • Belief Networks abandon the simple independence
    assumptions of naïve Bayesian systems and replace
    them by a more complicated notion of independence
    called d-separation.
  • Problem Given evidence involving a set of
    variables E when are two sets of variables X and
    Y of a belief network independent (d-separated)?
  • Why is this question important? If X and Y are
    d-separated (given E)
  • P(XYE)P(XE)P(YE) and
  • P(XEY)P(XE)
  • D-separation is used a lot in belief network
    computations (see P(DS1,S2) example to be
    discussed later) particularly to speed up
    belief network computations.

9
D-Separation All paths between members of X and
Y must match one of the following 4 patters
Y
X
E(in E, not in E)
(1a)
(1b)
(2)
(3)
10
D-Separation
A
D
C
B
E
  • a)  Which of the following statements are implied
    by the indicated network structure answer yes
    and no and give a brief reason for your answer!
    6
  • i) P(A,BC) P(AC)P(BC)
  •  yes, because
  • ii) P(C,ED) P(CD)P(ED)
  •  
  •  no, because
  • iii) P(CA)P(C)
  •  no, because

11
Fred/John Complains Problem Problem 12
Assignment3 Fall 2002
  • P(FC)P(FCGFA)P(GFA) P(FCGFB)P(GFB)
    P(FCGFC)P(GFC) 00.2 0.5x0.6 1x0.2
    0.5
  • P(JC) (problem description) P(FC,GJB)
    (FC,GJC) (d-separation of FC and GJ)
    P(FC)0.8 P(FC)0.1P(FC)0.90.45
  • P(JCFC) P(JC,GJAFC) P(JC,GJBFC)
    P(JC,GJAFC) P(GJAFC)P(JCGJA,FC)
    (GJ and FC are d-separated)
    P(GJA)P(JCGJA,FC) P(GJB)P(JCGJB,FC)
    P(GJA) P(JCGJA,FC) 0.10 0.8x1 0.1x1
    0.9
  • P(JCGFC) P(JC,FCGFC) P(JC,FCGFC)
    P(FCGFC)P(JCFC,GFC) P(FCGFC)P(JCFC,GF
    C) (given FC JC and GF are d-separated)
    P(FCGFC)P(JCFC) P(FC)GFC)P(JCFC)
    1(JCFC) 0 0.9
  • P(GFCJC) (Bayes Theorem) P(JCGFC)
    P(GFC) / P(JC) 0.90.2/0.450.4

(1)
(3)
(4)
(2)
Remark In the example P(GFB) and P(GFBJC) are
both 0.6, but P(GFC) is 0.2 whereas
P(GFCJC)0.4
12
Compute P(DS1,S2)!!
S1
D
B
S2
  • All 3 variables of B have binary states T,F
  • P(D) is a short notation for P(DT) and P(S2D)
    is a short notation for P(S2TDF).
  • Bs probability tables contain P(D)0.1,
    P(S1D)0.95, P(S2D)0.8, P(S1D)0.2,
    P(S2D)0.2
  • Task Compute P(DS1,S2)

13
Computing P(DS1,S2)
  • P(DS1,S2)P(D)P(S1D)P(S2D)/P(S1,S2) because
    S1D indep S2D
  • P(DS1,S2)P(D)P(S1D)P(S2D)/P(S1,S2) S1D
    indep S2D
  • (12) 1(P(D)P(S1D)P(S2D)
    P(D)P(S1D)P(S2D))/P(S1,S2)
  • P(S1,S2) P(D)P(S1D)P(S2D)
    P(D)P(S1D)P(S2D)g
  • P(DS1,S2) a / a b with
  • aP(D)P(S1D)P(S2D) and b P(D)P(S1D)P(S2
    D)
  • For the example a0.10.950.80.076 and b
    0.90.20.20.036
  • P(DS1,S2)0.076/0.1120.678

S1
D
S2
14
How do Belief Network Tools Perform These
Computations?
  • Basic Problem How to compute P(VariableEvidence)
    efficiently?
  • The asked probability has to be transformed
    (using definitions and rules of probability,
    d-separation,) into an equivalent expression
    that only involves known probabilities (this
    transformation can take many many steps
    especially if the belief network contains many
    variables and long paths between the
    variables).
  • For a given expression a large number of
    transformation can be used (e.g. P(A,B,C))
  • In general, the problem has been shown to be
    NP-hard
  • Popular algorithms to solve this problem include
    Junction Trees (Netica), Loop Cutset, Cutset
    Conditioning, Stochastic Simulation, Clustering
    (Hugin),

15
Other Capabilities of Belief Network Tools
  • Learning belief networks from empirical data
  • Support for continuous variables
  • Support to map continuous variables into nominal
    variables
  • Support for popular density functions
  • Support for utility computations and decision
    support
  • (many other things)
Write a Comment
User Comments (0)
About PowerShow.com