DCSP-5: Noise - PowerPoint PPT Presentation

1 / 75
About This Presentation
Title:

DCSP-5: Noise

Description:

... (xi) I(xi) = - Si P(xi) log2 P(xi) A binary symmetric ... the uncertainly of disorder of a system). ... by the average number of bits needed for ... – PowerPoint PPT presentation

Number of Views:31
Avg rating:3.0/5.0
Slides: 76
Provided by: Jian136
Category:
Tags: dcsp | binary | noise | number | system

less

Transcript and Presenter's Notes

Title: DCSP-5: Noise


1
DCSP-5 Noise
  • Jianfeng Feng
  • Department of Computer Science Warwick Univ., UK
  • Jianfeng.feng_at_warwick.ac.uk
  • http//www.dcs.warwick.ac.uk/feng/dcsp.html

2
Assignment2015
  • Q1 you should be able to do it after last
  • week seminar
  • Q2 need a bit reading (my lecture notes)
  • Q3 standard
  • Q4 standard

3
Assignment2015
  • Q5 standard
  • Q6 standard
  • Q7 after todays lecture
  • Q8 load jazz, plot soundsc
  • load Tunejazz plot
  • load NoiseJazz plot

4
Recap
  • Fourier Transform for a periodic signal
  • sim(n w t), cos(n w t)
  • For general function case,

5
Recap this is all you have to remember (know)?
  • Fourier Transform for a periodic signal
  • sim(n w t), cos(n w t)
  • For general function case,

6
Can you do FT for cos(2 p t)?
Dirac delta function

7
Dirac delta function
For example, take F0 in the equation above, we
have It
makes no sense !!!!
8
Dirac delta function A photo with the highest
IQ (15 NL)
Shordiger
Heisenberg
Ehrenfest
Bragg
Dirac
Compton
Bone
Pauli
Debije
De Boer
M Curie
Einstein
Planck
Lorentz
Langevin
9
Dirac delta function A photo with the highest
IQ (15 NL)
Shordiger
Heisenberg
Ehrenfest
Bragg
Dirac
Compton
Bone
Pauli
Debije
De Boer
M Curie
Einstein
Planck
Lorentz
Langevin
10
Dirac delta function
The (digital) delta function, for a given
n0
n00 here
d(t)
11
Dirac delta function
The (digital) delta function, for a given
n0 Dirac delta function d(x) (you
could find a nice movie in Wiki)
n00 here
d(t)
12
Dirac delta function
Dirac delta function d(x)
The FT of cos(2pt) is
-1 0 1
Frequency
13
A final note (in exam or future)
  • Fourier Transform for a periodic signal
  • sim(n w t), cos(n w t)
  • For general function case (it is true, but need
    a bit further work),

14
Summary
Will come back to it soon (numerical) This trick
(FT) has changed our life
and
will continue to do so
15
This Weeks Summary
  • Noise
  • Information Theory

16
Noise in communication systems probability and
random signals
Noise
  • I imread('peppers.png')
  • imshow(I)
  • noise 1randn(size(I))
  • Noisy imadd(I,im2uint8(noise))
  • imshow(Noisy)

17
Noise in communication systems probability and
random signals
Noise
  • I imread('peppers.png')
  • imshow(I)
  • noise 1randn(size(I))
  • Noisy imadd(I,im2uint8(noise))
  • imshow(Noisy)

18
Noise
  • Noise is a random signal (in general).
  • By this we mean that we cannot predict its value.
  • We can only make statements about the probability
    of it taking a particular value

19
pdf
  • The probability density function (pdf) p(x) of a
    random variable x is the probability that x takes
    a value between x0 and x0 dx.
  • We write this as follows
  • p(x0 )dx P(x0 ltxlt x0 dx)

P(x)
x0 x0 dx
20
pdf
  • Probability that x will take a value lying
    between x1 and x2 is
  • The probability is unity. Thus

21
IQ distribution
22
pdf
  • A density satifying the equation is termed
    normalized.
  • The cumulative distribution function (CDF) F(x)
    is the probability x is less than x0
  • My IQ is above 85 (F(my IQ)85).

23
pdf
  • From the rules of integration
  • P(x1ltxltx2) P(x2) --P(x1)
  • pdf has two classes continuous and discrete

24
  • Continuous distribution
  • An example of a continuous distribution is the
    Normal, or Gaussian distribution
  • where m, s is the mean and standard variation
    value of p(x).
  • The constant term ensures that the distribution
    is normalized.

25
  • Continuous distribution.
  • This expression is important as many actually
    occurring noise source can be described by it,
    i.e. white noise or coloured noise.

26
Generating f(x) from matlab
Xrandn(1,1000) Plot(x)
  • X1, x2, . X1000,
  • Each xi is independent
  • Histogram

27
Discrete distribution.
  • If a random variable can only take discrete
    value, its pdf takes the forms of lines.
  • An example of a discrete distribution is the
    Poisson distribution

28
Discrete distribution.
29
Mean and variance
  • We cannot predicate value a random variable
  • We can introduce measures that summarise what we
    expect to happen on average.
  • The two most important measures are the mean (or
    expectation) and the standard deviation.
  • The mean of a random variable x is defined to be

30
Mean and variance
  • In the examples above we have assumed that the
    mean of the Gaussian distribution to be 0, the
    mean of the Poisson distribution is found to be
    l.

31
Mean and variance
  • The mean of a distribution is, in common sense,
    the average value.
  • Can be estimated from data
  • Assume that x1, x2, x3, ,xN are sampled from
    a distribution
  • Law of Large Numbers EX (x1x2xN)/N

32
Mean and variance
mean
  • The more data we have, the more accurate we can
    estimate the mean
  • (x1x2xN)/N against N for randn(1,N)

33
Mean and variance
  • The variance is defined as The variance s is
    defined to be
  • The square root of the variance is called
    standard deviation.
  • Again, it can be estimated from data

34
Mean and variance
  • The standard deviation is a measure of the spread
    of the probability distribution around the mean.
  • A small standard deviation means the distribution
    are close to the mean.
  • A large value indicates a wide range of possible
    outcomes.
  • The Gaussian distribution contains the standard
    deviation within its definition (m,s)

35
Mean and variance
  • Communication signals can be modelled as a
    zero-mean, Gaussian random variable.
  • This means that its amplitude at a particular
    time has a PDF given by Eq. above.
  • The statement that noise is zero mean says that,
    on average, the noise signal takes the values
    zero.

36
Mean and variance
http//en.wikipedia.org/wiki/Nations_and_intellige
nce
37
Einsteins IQ
Einsteins IQ160 What about yours?
Above Average 34.1
Exceptionally Gifted 0.13
Low Intelligence 13.6
High Intelligence 13.6
Superior Intelligence 2.1
Mentally Inadequate 23
Average 34.1
38
SNR
  • Signal to noise ratio is an important quantity in
    determining the performance of a communication
    channel.
  • The noise power referred to in the definition is
    the mean noise power.
  • It can therefore be rewritten as
  • SNR 10 log10
    ( S / s2)

39
Correlation or covariance
  • Cov(X,Y) E(X-EX)(Y-EY)
  • correlation coefficient is normalized covariance
  • Coef(X,Y) E(X-EX)(Y-EY) / s(X)s(Y)
  • Positive correlation, Negative correlation
  • No correlation (independent)

40
Stochastic process signal
  • A stochastic process is a collection of random
    variables xn, for each fixed n, it is a
    random variable
  • Signal is a typical stochastic process
  • To understand how xn evolves with n, we will
    look at auto-correlation function (ACF)
  • ACF is the correlation between k steps

41
Stochastic process
gtgt clear all close all n200 for
i110 x(i)randn(1,1) y(i)x(i) end for
i1n-10 y(i10)randn(1,1)
x(i10).8x(i)y(i10) end plot(xcorr(x)/ma
x(xcorr(x))) hold on plot(xcorr(y)/max(xcorr(y)),
'r')
  • two signals are generated y (red) is simply
    randn(1,200)

  • x (blue) is generated xi10.8xi yi10
  • For y, we have g(0)1, g(n)0, if n is not 0
    having no memory
  • For x, we have g (0)1, and g (n) is not zero,
    for some n having memory

42
white noise wn
  • White noise is a random process we can not
    predict at all (independent of history)
  • In other words, it is the most violent noise
  • White noise draws its name from white light which
    will become
  • clear in the next few lectures

43
white noise wn
  • The most noisy noise is a white noise since its
    autocorrelation is zero, i.e.
  • corr(wn, wm)0 when
  • Otherwise, we called it colour noise since we
    can predict some outcome of wn, given wm,
    mltn

44
Why do we love Gaussian?
Sweety Gaussian
45
Sweety Gaussian

Yes, I am junior Gaussian

Herr Gauss Frau Gauss
Juenge Gauss
  • A linear combination of two Gaussian random
    variables is Gaussian again
  • For example, given two independent Gaussian
    variable X and Y with mean zero
  • aXbY is a Gaussian variable with mean zero and
    variance a2 s(X)b2s(Y)
  • This is very rare (the only one in continuous
    distribution) but extremely useful panda in the
    family of all distributions

46
DCSP-6 Information Theory
  • Jianfeng Feng
  • Department of Computer Science Warwick Univ., UK
  • Jianfeng.feng_at_warwick.ac.uk
  • http//www.dcs.warwick.ac.uk/feng/dcsp.html

47
Data Transmission
48
Data Transmission
How to deal with noise? How to transmit
signals?
49
Data Transmission
50
Data Transmission
  • Transform I
  • Fourier Transform
  • ASK (AM), FSK(FM), and PSK
  • (skipped, but common knowledge)
  • Noise
  • Signal Transmission

51
Today
  • Data transmission
  • Shannon Information and Coding Information
    theory,
  • coding of information for efficiency and error
    protection

52
Information and coding theory
  • Information theory is concerned with
  • description of information sources
  • representation of the information from a source
  • (coding) ,
  • transmission of this information over channel.

53
Information and coding theory
Information and coding theory
  • The best example
  • how a deep mathematical theory
  • could be successfully applied to
  • solving engineering problems.

54
Information and coding theory
  • Information theory is a discipline in applied
    mathematics involving the
  • quantification of data
  • with the goal of enabling as much data as
    possible to be reliably
  • stored
  • on a medium and/or
  • communicated
  • over a channel.

55
Information and coding theory
  • The measure of data, known as
  • information entropy,
  • is usually expressed by the average number of
    bits needed for storage or communication.

56
Information and coding theory
  • The field is at the crossroads of
  • mathematics,
  • statistics,
  • computer science,
  • physics,
  • neurobiology,
  • electrical engineering.

57
Information and coding theory
  • Impact has been crucial to success of
  • voyager missions to deep space,
  • invention of the CD,
  • feasibility of mobile phones,
  • development of the Internet,
  • the study of linguistics and of human
    perception,
  • understanding of black holes,
  • and numerous other fields.

58
Information and coding theory
  • Founded in 1948 by Claude Shannon in his
    seminal work
  • A Mathematical Theory of Communication

59
Information and coding theory
  • The bible paper cited more than 60,000

60
Information and coding theory
  • The most fundamental results of this theory are
  • 1. Shannon's source coding theorem
  • the number of bits needed to represent the
    result of
  • an uncertain event is given by its entropy
  • 2. Shannon's noisy-channel coding theorem
  • reliable communication is possible over
    noisy
  • channels if the rate of communication is
    below a
  • certain threshold called the channel
    capacity.
  • The channel capacity can be approached by
    using appropriate encoding and decoding systems.

61
Information and coding theory
  • The most fundamental results of this theory are
  • 1. Shannon's source coding theorem
  • the number of bits needed to represent the
    result of
  • an uncertain event is given by its entropy
  • 2. Shannon's noisy-channel coding theorem
  • reliable communication is possible over
    noisy
  • channels if the rate of communication is
    below a
  • certain threshold called the channel
    capacity.
  • The channel capacity can be approached by
    using appropriate encoding and decoding systems.

62
Information and coding theory
  • Consider to predict the activity of Prime
    minister tomorrow.
  • This prediction is an information source.

63
Information and coding theory
  • Consider to predict the activity of Prime
    Minister tomorrow.
  • This prediction is an information source X.
  • The information source X O, R has two
    outcomes
  • He will be in his office (O),
  • he will be naked and run 10 miles in London (R).

64
Information and coding theory
  • Clearly, the outcome of 'in office' contains
    little information
  • it is a highly probable outcome.
  • The outcome 'naked run', however contains
    considerable information
  • it is a highly improbable event.

65
Information and coding theory
  • An information source is a probability
    distribution, i.e. a set of probabilities
    assigned to a set of outcomes (events).
  • This reflects the fact that the information
    contained in an outcome is determined not only
    by the outcome, but by how uncertain it is.
  • An almost certain outcome contains little
    information.
  • A measure of the information contained in an
    outcome was introduced by Hartley in 1927.

66
Information
  • Defined the information contained in an outcome
    xi in xx1, x2,,xn
  • I(xi) - log2 p(xi)

67
Information
  • The definition above also satisfies the
    requirement that the total information in in
    dependent events should add.
  • Clearly, our prime minister prediction for two
    days contain twice as much information as for one
    day.

68
Information
  • The definition above also satisfies the
    requirement that the total information in in
    dependent events should add.
  • Clearly, our prime minister prediction for two
    days contain twice as much information as for one
    day XOO, OR, RO, RR.
  • For two independent outcomes xi and xj,
  • I(xi and xj) - log2 P(xi and xj)
  • - log2 P(xi) P(xj)
  • - log2 P(xi) - log2P(xj)

69
Entropy
  • The measure entropy H(X) defines the information
    content of the source X as a whole.
  • It is the mean information provided by the
    source.
  • We have
  • H(X) Si P(xi) I(xi) - Si P(xi) log2
    P(xi)
  • A binary symmetric source (BSS) is a source with
    two outputs whose probabilities are p and 1-p
    respectively.

70
Entropy
  • The prime minister discussed is a BSS.
  • The entropy of the BBS source is
  • H(X) -p log2 p - (1-p) log2 (1-p)

71
Entropy
  • .
  • When one outcome is certain, so is the other, and
    the entropy is zero.
  • As p increases, so too does the entropy, until
    it reaches a maximum when p 1-p 0.5.
  • When p is greater than 0.5, the curve declines
    symmetrically to zero, reached when p1.

72
Next Week
  • Application of Entropy in coding
  • Minimal length coding

73
Entropy
  • We conclude that the average information in BSS
    is maximised when both outcomes are equally
    likely.
  • Entropy is measuring the average uncertainty
    of the source.
  • (The term entropy is borrowed from
    thermodynamics. There too it is a measure of the
    uncertainly of disorder of a system).
  • Shannon
  • My greatest concern was what to call it.
  • I thought of calling it information, but the
    word was overly used, so I decided to call it
    uncertainty.
  • When I discussed it with John von Neumann, he had
    a better idea.
  • Von Neumann told me, You should call it entropy,
    for two reasons.
  • In the first place your uncertainty function has
    been used in statistical mechanics under that
    name, so it already has a name.
  • In the second place, and more important, nobody
    knows what entropy really is, so in a debate you
    will always have the advantage.

74
Entropy
In Physics thermodynamics
  • The arrow of time (Wiki)
  • Entropy is the only quantity in the physical
    sciences that seems to imply a particular
    direction of progress, sometimes called an arrow
    of time.
  • As time progresses, the second law of
    thermodynamics states that the entropy of an
    isolated systems never decreases
  • Hence, from this perspective, entropy measurement
    is thought of as a kind of clock

75
Entropy
Write a Comment
User Comments (0)
About PowerShow.com