Professional Documents
Culture Documents
xx
(f) = 0 for f W; by the sampling theorem
Where {X(n/2W)} denote the samples of the process at the
sampling rate (Nyquist rate) of f = 2W samples/s
Thus applying the sampling theorem we may convert the
output of an analog source into an equivalent discrete-
time source
Digital Communications Chapter 2: Source Coding
7
|
.
|
\
|
|
.
|
\
|
|
.
|
\
|
=
=
W 2
n
t W 2
W 2
n
t W 2 sin
)
w 2
n
X( X(t)
n
Sem. I, 2012/13
Mathematical Models
The output is statistically characterized by the joint pdf
p(x
1,
x
2
, x
m
) for m 1 where x
n
=X(n/2W); 1 n m, are
the random variables corresponding to the samples of X(t)
Note that the samples {X(n/2W)} are in general continuous
and cannot be represented digitally without loss of
precision
Quantization may be used such that each sample is a
discrete value, but it introduces distortion (More on this
later)
Digital Communications Chapter 2: Source Coding
8
Sem. I, 2012/13
Overview
Source coding and mathematical models
Logarithmic measure of information
Average mutual information
Digital Communications Chapter 2: Source Coding
9
Sem. I, 2012/13
Logarithmic Measure of Information
Consider two discrete random variables X and Y such that
X = {x
1
, x
2
,x
n
} and Y = {y
1
,y
2
,.y
m
}
Suppose we observe Y = y
j
and wish to determine,
quantitatively, the amount of information Y = y
j
provides
about the event
X = x
i
, i = 1,2,..n
Note that if X and Y are statistically independent Y = y
j
provides no information about the occurrence of X = x
i
If they are fully dependent Y = y
j
determines the
occurrence of X = x
i
The information content is simply that provided by X = x
i
Digital Communications Chapter 2: Source Coding
10
Sem. I, 2012/13
Logarithmic Measure of Information
A suitable measure that satisfies these conditions is the
logarithm of the ratio of the conditional probabilities
P{X = x
i
Y = y
j
} = P{x
i
y
j
} and P{X = x
i
} = p(x
i
)
Information content provided by the occurrence of Y = y
j
about the event X = x
i
is defined as
I(x
i
, y
j
) is called the mutual information between x
i
and y
j
Unit of the information measure is the nat(s) if the natural
logarithm is used and bit(s) if base 2 is used
Note that ln a = ln 2. log
2
a = 0.69315 log
2
a
Digital Communications Chapter 2: Source Coding
11
) P( x
) y P( x
y x I
i
j i
j i
/
log ) , ( =
Sem. I, 2012/13
Logarithmic Measure of Information
If X and Y are independent I(x
i
, y
j
) = 0
If the occurrence of Y = y
j
uniquely determines the
occurrence of X = x
i
,
I(x
i
, y
j
) = log (1/ p(x
i
))
i.e., I(x
i
, y
j
) = -log p(x
i
) = I(x
i
) self information of event X
Note that a high probability event conveys less
information than a low probability event
For a single event x where p(x) = 1, I(x) = 0 the
occurrence of a sure event does not convey any information
Digital Communications Chapter 2: Source Coding
12
Sem. I, 2012/13
Logarithmic Measure of Information
Example
1. A binary source emits either 0 or 1 every
s
seconds with
equal probability. Information content of each output is
then given by
I(x
i
) = -log
2
p = -log
2
1/2 = 1 bit, x
i
= 0,1
2. Suppose successive outputs are statistically
independent (DMS) and consider a block of binary digits
from the source in time interval k
s
Possible number of k-bit block = 2
k
= M, each of which is
equally probable with probability 1/M = 2
-k
I(x
i
) = -log
2
2
-k
= k bits in k
s
sec
(Note the additive property)
Digital Communications Chapter 2: Source Coding
13
Sem. I, 2012/13
Logarithmic Measure of Information
Now consider the following relationships
and since
Information provided by y
j
about x
i
is the same as that
provided by x
i
about Y = y
j
Digital Communications Chapter 2: Source Coding
14
) (y p ) (x p
) y , (x p
log
) (y p ) (x p
) p(y ) y (x p
log
) x p(
) y x ( p
log ) y , I(x
j i
j i
j i
j j i
i
j i
j i
= = =
) p(x
) y , p(x
) x p(y
i
j i
i j
=
) , (
) (
) (
) , (
i j
j
i j
j i
x y I
y p
x y p
y x I = =
Sem. I, 2012/13
Logarithmic Measure of Information
Using the same procedure as before we can also define
conditional self-information as
Now consider the following relationships
Which leads to
Note that is self information about X = x
i
after having
observed the event Y = y
j
Digital Communications Chapter 2: Source Coding
15
) I(x ) y , I(x
) p(x
) p(x ) y p(x
log
i j i
i
i j i
=
(
(
) ( ) ( ) , (
j i i j i
y x I x I y x I =
) (
j i
y x I
) ( log
) (
1
log ) (
2 2 j i
j i
j i
y x p
y x p
y x I = =
Sem. I, 2012/13
Logarithmic Measure of Information
Further, note that
and thus
Indicating that the mutual information between two events
can either be positive or negative
Digital Communications Chapter 2: Source Coding
16
when y x I
j i
0 ) , (
) ( ) (
j i i
y x I x I >
0 ) ( 0 ) (
j i i
y x I and x I
Sem. I, 2012/13
Overview
Source coding and mathematical models
Logarithmic measure of information
Average mutual information
Digital Communications Chapter 2: Source Coding
17
Sem. I, 2012/13
Average Mutual Information
Average mutual information between X and Y is given by
Similarly, average self-information is given by
where X represents the alphabet of possible output letters
from a source
H(X) represents the average self information per source
letter and is called the entropy of the source
Digital Communications Chapter 2: Source Coding
18
= = = =
= =
n
i
m
j
j i
j i
j i j i j
n
i
m
j
i
y p x p
y x p
y x p y x I y x p I
1 1
2
1 1
0
) ( ) (
) , (
log ) , ( ) , ( ) , ( ) ; ( Y X
= = ) ( log ) ( ) ( ) ( ) (
2 i i i i
x p x p x I x p H X
Sem. I, 2012/13
Average Mutual Information
If p(x
i
) = 1/n for all i, then the entropy of the source
becomes
In general, H(X) log n for any given set of source letter
probabilities
Thus, the entropy of a discrete source is maximum when the output
letters are all equally probable
Digital Communications Chapter 2: Source Coding
19
n log
n
1
log
n
1
H(X)
2 2
= =
Sem. I, 2012/13
Average Mutual Information
Consider a binary source where the letters {0,1} are
independent with probabilities
p(x
0
=0)=q or p(x
1
=1)=1-q
The entropy of the source is given by
H(X)= -q log
2
q (1-q) log
2
(1-q) = H(q)
whose plot as a function is shown below
Digital Communications Chapter 2: Source Coding
20
Binary Entropy Function
Sem. I, 2012/13
Average Mutual Information
In a similar manner as above, we can define conditional
self-information or conditional entropy as follows
Conditional entropy is the information or uncertainty in X
after Y is observed
From the definition of average mutual information one can
show that
Digital Communications Chapter 2: Source Coding
21
) / (
1
log ) , ( ) / (
1 1
j i
n
i
m
j
j i
y x p
y x p H
= =
= Y X
| |
) (
) ( log ) / ( log ) , (
1 1
X (X/Y)
Y) I(X,
H H
x p y x p y x p
n
i
m
j
i j i j i
+ =
=
= =
Sem. I, 2012/13
Average Mutual Information
Thus I(X,Y) = H(X) H(X/Y) and H(X) H(X|Y), with
equality when X and Y are independent
H(X/Y) is called the equivocation
It is interpreted as the amount of average uncertainty
remaining in X after observation of Y
H(x) Average uncertainty prior to observation
I(X;Y) Average information provided about the set X by
the observation of the set Y
Digital Communications Chapter 2: Source Coding
22
Sem. I, 2012/13
Average Mutual Information
The above results can be generalized for more than
two random variable
Suppose we have a block of k random variables
x
1
,x
2
,..x
k
with joint probability P(x
1
,x
2
,..x
k
); the
entropy of the block will then be given by
Digital Communications Chapter 2: Source Coding
23
) ...... , ( log ) ...... , ( .... ) (
2 1
1
1
2
2
2 1 jk j j
n
j
n
j
nk
jk
jk j j
x x x P x x x P X H
=
Sem. I, 2012/13
Continuous Random Variables - Information Measure
If X and Y be continuous random variables with joint pdf
f (x,y) and marginal pdfs f (x) and f (y) the average Mutual
information between X and Y is defined as
The concept of self information does not exactly carry over
to continuous random variables since these would require
infinite number of binary digits to exactly represent them
and thus making their entropies infinite
Digital Communications Chapter 2: Source Coding
24
dy dx
y f x f
x f x y f
x y f x f Y X I
=
) ( ) (
) ( ) / (
log ) / ( ) ( ) , (
Sem. I, 2012/13
Continuous Random Variables
We can however define differential entropy for continuous
random variable as
And the average conditional entropy as
Average mutual information is then given by I(X,Y)= H(X) -
H(X/Y) or I(X,Y) = H(Y) - H(Y/X)
Digital Communications Chapter 2: Source Coding
25
= dx x f x f x H ) ( log ) ( ) (
= dxdy y x f y x f Y X H ) , ( log ) , ( ) / (
Sem. I, 2012/13
Continuous Random Variables
If X is discrete and Y is continuous, the density of Y is
expressed as
The mutual information about X = x
i
provided by the
occurrence of the event Y = y is given by
The average mutual information between X and Y is
Digital Communications Chapter 2: Source Coding
26
=
=
n
i
i i
x p x y f y f
1
) ( ) / ( ) (
) (
) / (
log
) ( ) (
) ( ) / (
log ) ; (
y f
x y f
x p y f
x p x y f
Y x I
i
i
i i
i
= =
) (
) / (
log ) ( ) / ( ) ; (
1
y f
x y f
x p x y f Y X I
i
n
i
i i
=