Probabilistic models - PowerPoint PPT Presentation

About This Presentation
Title:

Probabilistic models

Description:

Experiment: a procedure involving chance ... Multinomial distribution. Dirichlet distribution. Extreme value distribution (EVD) ... Multinomial distribution ... – PowerPoint PPT presentation

Number of Views:44
Avg rating:3.0/5.0
Slides: 30
Provided by: hat89
Category:

less

Transcript and Presenter's Notes

Title: Probabilistic models


1
Probabilistic models
  • Haixu Tang
  • School of Informatics

2
Probability
  • Experiment a procedure involving chance that
    leads to different results
  • Outcome the result of a single trial of an
    experiment
  • Event one or more outcomes of an experiment
  • Probability the measure of how likely an event
    is

3
Example a fair 6-sided dice
  • Outcome The possible outcomes of this experiment
    are 1, 2, 3, 4, 5 and 6
  • Events 1 6 even
  • Probability outcomes are equally likely to
    occur.
  • P(A)  The Number Of Ways Event A Can Occur  /
    The Total Number Of Possible Outcomes
  • P(1)P(6)1/6 P(even)3/61/2

4
Probability distribution
  • Probability distribution the assignment of a
    probability P(x) to each outcome x.
  • A fair dice outcomes are equally likely to occur
    ? the probability distribution over the all six
    outcomes P(x)1/6, x1,2,3,4,5 or 6.
  • A loaded dice outcomes are unequally likely to
    occur ? the probability distribution over the all
    six outcomes P(x)f(x), x1,2,3,4,5 or 6, but
    ?f(x)1.

5
Example DNA sequences
  • Event Observing a DNA sequence Ss1s2sn si ?
    A,C,G,T
  • Random sequence model (or Independent and
    identically-distributed, i.i.d. model) si
    occurs at random with the probability P(si),
    independent of all other residues in the
    sequence
  • P(S)
  • This model will be used as a background model (or
    called null hypothesis).

6
Conditional probability
  • P(i?) the measure of how likely an event i
    happens under the condition ?
  • Example two dices D1, D2
  • P(iD1) ?probability for picking i using dicer D1
  • P(iD2) ?probability for picking i using dicer D2

7
Joint probability
  • Two experiments X and Y
  • P(X,Y) ? joint probability (distribution) of
    experiments X and Y
  • P(X,Y)P(XY)P(Y)P(YX)P(X)
  • P(XY)P(X), X and Y are independent
  • Example experiment 1 (selecting a dice),
    experiment 2 (rolling the selected dice)
  • P(y) yD1 or D2
  • P(i, D1)P(i D1)P(D1)
  • P(i D1)P(i D2), independent events

8
Marginal probability
  • P(X)?YP(XY)P(Y)
  • Example experiment 1 (selecting a dice),
    experiment 2 (rolling the selected dice)
  • P(y) yD1 or D2
  • P(i) P(i D1)P(D1)P(i D2)P(D2)
  • P(i D1)P(i D2), independent events
  • P(i) P(i D1)(P(D1)P(D2)) P(i D1)

9
Probability models
  • A system that produces different outcomes with
    different probabilities.
  • It can simulate a class of objects (events),
    assigning each an associated probability.
  • Simple objects (processes) ? probability
    distributions

10
Example continuous variable
  • The whole set of outcomes X (x?X) can be
    infinite.
  • Continuous variable x?x0,x1
  • P(x0xx1) -gt0
  • P(x-dx/2 x xdx/2) f(x)dx ?f(x)dx1
  • f(x) probability density function (density,
    pdf)
  • P(x?y) ?yx0f(x)dx cumulated density function
    (cdf)

x1
dx
x0
x1
11
Mean and variance
  • Mean
  • m?xP(x)
  • Variance
  • ?2 ?(k-m)2P(k)
  • ? standard deviation

12
Typical probability distributions
  • Binomial distribution
  • Gaussian distribution
  • Multinomial distribution
  • Dirichlet distribution
  • Extreme value distribution (EVD)

13
Binomial distribution
  • An experiment with binary outcomes 0 or 1
  • Probability distribution of a single experiment
    P(1)p and P(0) 1-p
  • Probability distribution of N tries of the same
    experiment
  • Bi(k 1s out of N tries)

14
Gaussian distribution
  • N -gt ?, Bi -gt Gaussian distribution
  • Define the new variable u (k-m)/ ?
  • f(u)

15
Multinomial distribution
  • An experiment with K independent outcomes with
    probabilities ?i, i 1,,K, ??i 1.
  • Probability distribution of N tries of the same
    experiment, getting ni occurrences of outcome i,
    ?ni N.
  • M(N?)

16
Example a fair dice
  • Probability outcomes (1,2,,6) are equally
    likely to occur
  • Probability of rolling 1 dozen times (12) and
    getting each outcome twice
  • 3.4?10-3

17
Example a loaded dice
  • Probability outcomes (1,2,,6) are unequally
    likely to occur P(6)0.5, P(1)P(2)P(5)0.1
  • Probability of rolling 1 dozen times (12) and
    getting each outcome twice
  • 1.87?10-4

18
Dirichlet distribution
  • Outcomes ?(?1, ?2,, ?K)
  • Density D(?a)
  • (a1, a2,, aK) are constants ? different a gives
    different probability distribution over ?.
  • K2 ? Beta distribution

19
Example dice factories
  • Dice factories produces all kinds of dices ?(1),
    ?(2),, ?(6)
  • A dice factory distinguish itself from the others
    by parameters a(a1,a2 ,a3 , a4 , a5 , a6)
  • The probability of producing a dice ? in the
    factory a is determined by D(?a)

20
Extreme value distribution
  • Outcome the largest number among N samples from
    a density g(x) is larger than x
  • For a variety of densities g(x),
  • pdf
  • cdf

21
Probabilistic model
  • Selecting a model
  • Probabilistic distribution
  • Machine learning methods
  • Neural nets
  • Support Vector Machines (SVMs)
  • Probabilistic graphical models
  • Markov models
  • Hidden Markov models
  • Bayesian models
  • Stochastic grammars
  • Model ? data (sampling)
  • Data ? model (inference)

22
Sampling
  • Probabilistic model with parameter ? ? P(x ?)
    for event x
  • Sampling generate a large set of events xi with
    probability P(xi ?)
  • Random number generator ( function rand() picks a
    number randomly from the interval 0,1) with the
    uniform density
  • Sampling from a probabilistic model ?
    transforming P(xi ?) to a uniform distribution
  • For a finite set X (xi?X), find i s.t.
    P(x1)P(xi-1) lt rand(0,1) lt P(x1)P(xi-1)
    P(xi)

23
Inference (ML)
  • Estimating the model parameters (inference) from
    large sets of trusted examples
  • Given a set of data D (training set), find a
    model with parameters ? with the maximal
    likelihood P(? D)

24
Example a loaded dice
  • loaded dice to estimate parameters ?1, ?2,, ?6,
    based on N observations Dd1,d2,dN
  • ?ini / N, where ni is of i, is the maximum
    likelihood solution (11.5)
  • Inference from counts

25
Bayesian statistics
  • P(XY)P(YX)P(X)/P(Y)
  • P(? D) P(?)?P(D ?)/P(D)
  • P(?)?P(D ?)/ ??(P(D ?)P (?)
  • P(?) ? prior probability P(?D) ? posterior
    probability

26
Example two dices
  • Fair dice 0.99 loaded dice 0.01, P(6)0.5,
    P(1)P(5)0.1
  • 3 consecutive 6es
  • P(loaded36s)P(loaded)P(36sloaded)/P(36s
    ) 0.01(0.53 / C)
  • P(fair36s)P(fair)P(36sfair)/P(36s)
    0.99 ((1/6)3 / C)
  • Likelihood ratio P(loaded36s) / P(fair36s)
    lt 1

27
Inference from counts including prior knowledge
  • Prior knowledge is important when the data is
    scarce
  • Use Dirichlet distribution as prior
  • P(? n) D(?a)?P(n?)/P(n)
  • Equivalent to add ai as pseudo-counts to the
    observation I (11.5)
  • We can forget about statistics and use
    pseudo-counts in the parameter estimation!

28
Entropy
  • Probabilities distributions P(xi) over K events
  • H(x)-? P(xi) log P(xi)
  • Maximized for uniform distribution P(xi)1/K
  • A measure of average uncertainty

29
Mutual information
  • Measure of independence of two random variable X
    and Y
  • P(XY)P(X), X and Y are independent ?
    P(X,Y)/P(X)P(Y)1
  • M(XY)?x,y P(x,y)logP(x,y)/P(x)P(y)
  • 0 ? independent
Write a Comment
User Comments (0)
About PowerShow.com