Professional Documents
Culture Documents
Marius Pesavento
NTS
COURSE ORGANIZATION
Website: http://www.nts.tu-darmstadt.de/
NTS
RECOMMENDED TEXTBOOKS
NTS
COURSE OUTLINE
I
I
I
I
I
Multi-access channels
Broadcast channels
Relay channels
Cyclic codes
Convolutional codes
Turbo-codes
NTS
Basic theory of channel coding, linear block coding, Reed-Muller codes, Golay
code
NTS
Let X be a continuous random variable with the cumulative density function (cdf)
FX (x) = Probability{X x} = P(X x)
Probability density function (pdf):
fX (x) =
where
FX (x)
x
x0
FX (x0 ) =
fX (x) dx
NTS
fX (x) dx = 1
Simple interpretation:
fX (x) = lim0
P{x /2 X x + /2}
f (x)
X
x1 x2
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 7
SURFACE =Probability{x
1< X< x 2}
NTS
EXAMPLE 1
FX (x)
1
0
f (x)
1/T
0
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 8
NTS
EXAMPLE 2
Let the real-valued random variable X has the so-called Gaussian (normal)
distribution
2
2
1
fX (x) = p
e (xX ) /2X
2
2X
where X2 = var{X } is the variance and X is the mean. The corresponding
distribution function is given by
Z x
2
2
1
FX (x) = p
e (X ) /2X d
2
2X
NTS
FX (x)
1
0
1/2
(22X)
f (x)
X
2 X
0
x
NTS
Let X now be a discrete random variable which takes the values xi (i = 1, ... , I )
with the probabilities P(xi ) (i = 1, ... , I ), respectively.
For discrete variables, we define the probability mass function
P(xi ) = Probability(X = xi )
P(xi ) = 1
i=1
NTS
, x = 0
0, x =
6 0
Z
,
(x) dx = 1
NTS
Using the definition of the unit step function, we can express the cdf as
FX (x) =
I
X
P(xi )u(x xi )
i=1
NTS
I
X
P(xi )(x xi )
i=1
I
X
fX (x) dx =
P(xi )(x xi ) dx
i=1
I Z
X
i=1
P(xi )(x xi ) dx =
I
X
P(xi ) = 1
i=1
NTS
EXAMPLE 1
FX (x)
1.0
0.5
0
f (x)
(0.5)
0
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 15
(0.5)
1
x
NTS
EXAMPLE 2
FX (x)
1
1/6
0
f (x)
(1/6)
0
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 16
x
NTS
STATISTICAL EXPECTATION
X = E{X } =
x fX (x) dx
X = E{X } =
xfX (x) dx =
I Z
X
i=1
I
X
P(xi )(x xi ) dx
i=1
xP(xi )(x xi ) dx =
I
X
xi P(xi )
i=1
NTS
STATISTICAL EXPECTATION
I
X
g (xi ) P(xi )
i=1
NTS
NTS
EXAMPLE
Compute the mean and variance of the random variable uniformly distributed in
the interval [0, 1]
f (x)
X
1
1
1
x
x dx = =
2
2
0
Z
X
=
0
X2
1
2
x dx
=
0
2X
1
x 3
1
1 1
1
= = =
3
4
3 4
12
0
NTS
JOINT DISTRIBUTION
Joint pdf:
fX ,Y (x, y ) =
2 FX ,Y (x, y )
x y
NTS
JOINT DISTRIBUTION
x0
y0
FX ,Y (x0 , y0 ) =
fX ,Y (x, y ) dx dy
Also,
Z
fX ,Y (x, y ) dx = fY (y ) ,
fX ,Y (x, y ) dy = fX (x)
NTS
CONDITIONAL DISTRIBUTION
In practical problems, we are often interested in the pdf of one random variable X
conditioned by the fact that a second random variable Y has some specific value
y . It is obvious that
P(X x; Y y ) = P(X x|Y y )P(Y y )
Then, conditional cdf is defined as
FX (x|y ) = P(X x|Y y ) =
FX ,Y (x, y )
FY (y )
FX ,Y (x, y )
FX (x)
NTS
CONDITIONAL DISTRIBUTION
fX ,Y (x, y )
fY (y )
fX ,Y (x, y )
fY (y |x) =
fX (x)
fX (x|y ) =
NTS
NORMALIZATION CONDITION
fX ,Y (x, y )
dx
fY (y )
Z
1
=
fX ,Y (x, y ) dx = 1
fY (y )
fX (x|y ) dx =
Conditional expectation:
Z
E{g (X )|y } =
g (x)fX (x|y ) dx
NTS
STATISTICAL INDEPENDENCE
NTS
EXAMPLE
Let
fX ,Y (x, y ) =
4xy , 0 x 1, 0 y 1
0,
otherwise
2x , 0 x 1,
0,
otherwise
2y , 0 y 1,
0,
otherwise
=
fY (y )
1
y 2
y dy = 4x
2
0
NTS
E{XY } X Y
NTS
If the variables X and Y are statistically independent then for any functions h
and g , E{h(X )g (Y )} = E{h(X )}E{g (Y )}.
NTS
EXTENSION TO
MULTIVARIAT DISTRIBUTIONS
We may also consider multiple (more than two) random variables X1 , ... , Xn .
NTS
MULTIVARIAT DISTRIBUTIONS
Introducing vectors
X =
[X1 , X2 , ... , Xn ]T
[x1 , x2 , ... , xn ]T
P(X x)
fX (x)
N FX (x)
x1 x2 xn
Normalization condition:
Z
fX (x) dx = 1
NTS
MULTIVARIAT DISTRIBUTIONS
g (x) fX (x) dx
NTS
Jointly Gaussian random variables have the following joint multivariate pdf:
T 1
1
1
fX (x) =
e 2 (xX ) R (xX )
n
1/2
( 2) det{R}
NTS
X2
X Y
X Y
Y2
,
E{(X X )(Y Y )}
X Y
NTS
+
2(1 2 )
X Y
X2
Y2
The maximum of this function is located in the point {x = X ; y = Y } and the
maximal value is
1
p
max {fX ,Y (x, y )} = fX ,Y (X , Y ) =
2X Y 1 2
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 35
NTS
e (xX ) /2X
e (y Y ) /2Y
=
2X
2Y
= fX (x) fY (y )
i.e., the variables X and Y become statistically independent. This is a very
important result showing that any uncorrelated Gaussian random variables are also
statistically independent! Note that in the case of non-Gaussian random variables,
this is not true.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 36
NTS
Contour plots of the bivariate Gaussian pdf with the parameters X = Y = 0 and
X = Y = 1.
3
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
3
3
0
x
NTS
After the occurrence of the event, there is no uncertainty anymore, but there
is gain in the amount of information.
NTS
N
X
P(xi ) log(P(xi ))
i=1
N
X
P(xi ) log
i=1
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 47
1
P(xi )
with
[H(X )] = Bit/Symbol
NTS
Example
0
I
I
0.5
NTS
N X
M
X
P(xi , yl ) log(P(xi , yl ))
i=1 l=1
NTS
Conditional Entropy
N X
M
X
i=1 l=1
N
X
i=1
P(xi )
M
X
l=1
NTS
Useful properties
Hence the entropy, conditional entropy, and joint entropy are related quantities.
Another important property: Conditioning reduces entropy:
H(X |Y ) H(X )
with the equality if and only if X and Y are statistically independent.
NTS
Mutual information
NTS
Mutual information
The amount of mutual information exchanged between two sources X and Y can
be obtained by averaging of I (xi ; yl ) as
I (X ; Y ) =
N X
M
X
P(xi |yl )
P(xi )
P(xi , yl )
P(xi )P(yl )
P(xi , yl ) log
i=1 l=1
N X
M
X
i=1 l=1
P(xi , yl ) log
with [I (X ; Y )] = Bit/Symbol
NTS
Mutual information
NTS
Channel capacity
The input probabilities P(xi ) are independent of the channel. We can then
maximize the mutual information I (X ; Y ) w.r.t. P(xi ). The channel capacity can
be then defined as the maximum mutual information in any single use of the
channel, where the maximization is over P(xi ) (i = 1, ... , M):
C = max I (X ; Y )
{P(xi )}
with [C ] = Bits/Symbol
NTS
Example
x1
y1
1- p
1- p
x2
y2
CB = 1 + p logp + (1 p) log(1 p)
= 1 HB (X )
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 56
NTS
H B (X), Bit/Zeichen
C B , Bit/Zeichen
0.5
0.5
NTS
Channel coding/decoding
The inevitable presence of noise in a channel causes errors between the output
and input data sequences of a digital communication system. To reduce these
errors we will resort to channel coding.
Channel encoder maps the incoming source data into a channel input sequence. It
adds redundancy to these data to protect it from errors.
Channel decoder inversely maps the channel output sequence into an output data
sequence in a way that the overall effect of the channel noise on the system is
minimized.
NTS
NTS
If
H(X ) > C
then it is impossible to transmit the source outputs over the channel with an
arbitrary small probability of error.
The latter theorem follows from the direct combination of source-coding and
channel-coding theorems.
NTS
Continuous sources
The mutual information between two continuous random sources X and Y with
the joint symbol pdf fX ,Y (x, y ) is given by
Z Z
fX (x|y )
I (X ; Y ) =
fX ,Y (x, y ) log
dx dy
fX (x)
Z Z
fX ,Y (x, y )
dx dy
=
fX ,Y (x, y ) log
fX (x)fY (y )
NTS
What is the relationship between the discrete and continuous mutual information?
It can be shown that the definitions of mutual information in the continuous and
discrete cases are essentially similar.
This property enables to use the continuous mutual information to define the
capacity in the case of continuously distributed (infinite alphabet) sources.
NTS
NTS
Bandbegrentzter AWGN-Kanal
H(f)
x(t)
y(t)
+
AWGN
n(t)
Leistungsdichtespectrum
des Rauschens
B f
-B
idealer Tiefpass
Bandbreite B
SN (f)
N 0 /2
f
NTS
Capacity
Shannons bound:
C
P
= lim B log 1 +
B
N0 B
P
P
= loge
' 1.44
N0
N0
NTS
i = 1, ... , K
( K
X
)
Xi2
i=1
K
X
i=1
E{Xi2 } =
K
X
Pi P
i=1
NTS
Water-filling
Pi =
K
X
i=1
( PN,i )+ P
i=1
x,
0,
if x 0
if x < 0
NTS
Water-filling
NTS
NTS
P=
P1
0
..
.
0
P2
..
.
..
.
0
0
..
.
PK
= diag{P1 , ... , PK }
PK
i=1
Pi = P. and
NTS
We use the Lagrange multiplier method. The Lagrangian function can be written
as
K
K
K
X
X
1X
Pi
L(P1 , ... , PK ) =
log 1 +
+ 0 (P
Pi ) +
i Pi
2
PN,i
i=1
i=1
i=1
where 0 , ... , K are the Lagrange multipliers. Differentiating L(P1 , ... , PK ) w.r.t.
Pi , we have
!
K
K
X
X
L
1
Pi
=
log e ln 1 +
+ 0 (P
Pi ) +
i Pi
Pi
Pi 2
PN,i
i=1
i=1
1/PN,i
1
= log e
0 + i
2
1 + Pi /PN,i
log e
1
=
0 + i
2 Pi + PN,i
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 71
NTS
Pi? = P
Pi? 0
(zero gradient)
(complementary slackness)
(constraint satisfaction)
i=1
log e
1
?0 + ?i
2 Pi? + PN,i
?i Pi?
?i
Thus
Pi?
0 and
PK
?
i=1 Pi
Pi? (?0
0; i = 1, ... , K
= P as well as
log e
1
log e
1
) and ?0
2 Pi? + PN,i
2 Pi? + PN,i
NTS
PK
i=1
Pi? = P as well as
log e
1
log e
1
) = 0 and ?0
2 Pi? + PN,i
2 Pi? + PN,i
Thus if
?0 <
log e 1
,
2 PN,i
then from the last equation we have Pi? > 0 which by slackness conditions implies
that
log e
1
.
?0 =
?
2 Pi + PN,i
and thus for ? = log e/(2?0 )
Pi? = ( ? PN,i ) .
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 73
NTS
PK
i=1
Pi? = P as well as
log e
1
log e
1
) = 0 and ?0
2 Pi? + PN,i
2 Pi? + PN,i
Reversely if
?0
log e 1
,
2 PN,i
log e 1
log e
1
>
,
?
?
2 Pi
2 Pi + PN,i
NTS
NTS
Ergodic capacity
NTS
Ergodic capacity
1
PH 2
log 1 +
2
PN
Note that the conventional capacity is instantaneous, that is, it characterizes the
maximal achievable rate for particular given realization of the gain H of the
channel.
How can we characterize the maximal achievable rate in average rather than for
some particular channel gain?
NTS
Ergodic capacity
Based on this fact, the ergodic capacity is defined as the instantaneous capacity C
averaged over the channel realizations:
CE = EH {C }
where EH {} denotes statistical expectation over the random channel gain.
NTS
Ergodic capacity
Assume that we know the channel gain pdf fH (h). In this case, we can compute
the ergodic capacity as
Z
CE =
fH (h) C (h) dh
NTS
Outage
Outage capacity: the transmission rate Cpout which does not exceed the
instantaneous capacity C in pout 100 percents of channel realizations.
The quantity pout is called outage probability.
Outage is defined as the event where, for some particular channel realization, the
chosen transmission rate is higher than the instantaneous capacity (that is, where
no error-free transmission is possible).
In the cases of small pout (roughly speaking, pout 0.1), outage-induced errors
can be cured by means of channel coding.
NTS
Outage
NTS
Outage
The outage capacity can be characterized as follows. Let the pdf of the
instantaneous capacity C = C (H) be fC (c) where fC (c) = 0 for c < 0. Then, the
outage capacity is defined by the equation
Z
p = P(C < Cpout ) =
Cpout
fC (c) dc
0
NTS
Channel coding
Channel encoding and decoding is used to correct errors that may occur during
the signal transmission over the channel.
NTS
NTS
Row-vector notations
m = [m1 , ... , mk ]
b = [b1 , ... , bnk ]
c = [b1 , ... , bnk , m1 , ... , mk ] = [b, m]
Block codes use the message bits to generate parity-check bits according to the
equation:
b = mP
where P is the k (n k) coefficient matrix. Noting that c = [b, m], we get that
c = [b, m] = [mP, m] = m[P, Ik ] = mG
where G is the k n generator matrix.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 85
NTS
Hamming codes
1
1
1
0
0
1
1
1
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
1
NTS
Message
0000
0001
0010
0011
0100
0101
...
Codeword
0000000
1010001
1110010
0100011
0110100
1100101
and so on ...
Hamming Weight
0
3
4
3
3
4
...
NTS
MULTI-ANTENNA CHANNELS
antennas
channel
Tx
antennas
Rx
In the frequency flat fading case, the signal in the mth receive antenna
Ym (t) =
N
X
m = 1, ... , M
n=1
where Hmn is the channel coefficient between the mth receive and nth transmit
antennas, Xn is the signal sent from the nth transmit antenna, and Zm is the noise
in the mth receive antenna.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 88
NTS
MIMO channel
H=
H11
H21
..
.
H12
H22
..
.
..
.
H1N
H2N
..
.
HM1
HM2
HMN
NTS
SIMO channel
1 antenna
antennas
Tx
Rx
In the frequency flat fading case, the signal in the nth receive antenna
Yn (t) = Hn (t)X (t) + Zn (t),
n = 1, ... , N
where Hn is the channel coefficient between the nth receive antenna and the
transmit antenna, and X (t) is the signal sent from the transmit antenna.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 90
NTS
SIMO channel
NTS
MISO channel
antennas
1 antenna
Tx
Rx
In the frequency flat fading case, the signal in the receive antenna
Y (t) =
N
X
n=1
where Hn is the channel coefficient between the receive antenna and the nth
transmit antenna, and Xn (t) is the signal sent from the nth transmit antenna.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 92
NTS
MISO channel
NTS
Let us consider the MIMO case assuming that z NC (0, 2 I). Then, the equation
y = Hx + z
describes a vector Gaussian channel. In the case of known channel at the
transmitter, the capacity can be computed by decomposing this channel into a set
of parallel independent scalar Gaussian sub-channels.
Singular value decomposition (SVD) of H:
H = UVH
where the M M matrix U and the N N matrix V are unitary, that is,
UH U = UUH = I and VH V = VVH = I.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 94
NTS
A = UVH =
m
n
= U
111
000
000 11
111
00
00
11
i ui viH
111
000
VH
n<m
m
111 11
000
00
00
11
A =
111
000
VH
n>m
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 95
NTS
NTS
z , UH z we become a system
NTS
i = 1, ... , no
NTS
NTS
NTS
no
X
i=1
Pi 2
log 1 + 2 i
bits/s
2
2i
+
Pno
and the water level is obtained from the total power constraint i=1
Pi P
Each i corresponds to an eigenmode of the channel, also called eigenchannel.
NTS
N Tx
wireless
MIMO
channel
M Rx
System model
+
M Rx
max
subject to:
(1)
(2)
and
are unitary:
for
i = 1,,r;
Optimization problem
Capacity:
max
subject to (1)
(2)
Water-filling principle
Water-filling principle
Water-filling principle
Water-filling principle
Water-filling principle
At high SNR, the water level is high and the policy of allocating equal amounts of
power to each channel is asymptotically optimal. In this case,
r
X
P2i
log 1 +
r 2
i=1
2
r
X
Pi
' B
log
r 2
C ' B
i=1
r
X
i=1
log
2i
r
bits/s
NTS
It can be proved that among the channels with the same power gain, the channels
with the equal spread of singular values result in the highest capacity.
This means that well-conditioned channel matrices are preferable in the high SNR
regime.
NTS
In this regime, the optimal policy is to allocate power to the channel with the
strongest eigenmode:
P2max
C ' B log 1 +
2
and ill-conditioned (rank-one) channel matrices are preferable.
Using the property log(1 + x) ' xloge that is valid for x 1, we have
C '
BP2max log e
2
NTS
Let us now obtain the MIMO channel capacity based on general considerations
assuming that H is fixed, while the other values (x, y and z) are random. In such
a case, no assumption on the channel knowledge at the transmitter is used at this
time, but the receiver is assumed to know H.
Capacity via mutual information:
C = max I (x; y) = max[H(y) H(y|x)]
p(x)
p(x)
NTS
Result: (Telatar, 1995; Foschini and Gans, 1998): Consider the model
y = Hx + z
where x NC (0, P), z NC (0, 2 I), and H is fixed. Let B be the channel
bandwidth in Hz. Then, the MIMO channel capacity is equal to
1
C = B log det I + 2 HPHH bits/s
NTS
Result 1
Let X1 , ... , Xn have a multivariate complex circular Gaussian distribution with the
mean and covariance matrix P:
fX (x) =
1
()n det{P}
e (xX )
P1 (xX )
Then
H(X) = H(X1 , ... , Xn ) = log ((e)n det{P})
Proof:
Z
H(X) =
fX (x)(x x )H P1 (x x )dx
+ ln (()n det{P})
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 120
NTS
Result 1 (proof)
NTS
Result 2
Let the random vector x Cn have zero mean and covariance ExxH = P. Then
H(X) = H(X1 , ... , Xn ) log {(e)n det{P}} with equality if and only if
X NC (0, P).
R
Proof:: Let g (x) be a pdf with covariance [P]ij = g (x)xi xj dx and let P (x) be
the complex circular Gaussian pdf NC (0, P).
NTS
Result 2 (proof)
Then the Kullback-Leibler D(g (x)||P (x)) distance between the two pdfs is given
as
Z
g (x)
0 D(g (x)||P (x)) = g (x) log(
)dx
P (x)
Z
= Hg (X) g (x) log(P (x)) dx
| {z }
quadratic form
|
Z
= Hg (X)
{z
second moment of X
NTS
It has been shown that for all random vectors with the covariance matrix R, the
entropy of y is maximized when y is zero-mean circularly-symmetric complex
Gaussian. But this is only true when the input vectors x are zero-mean
circularly-symmetric complex Gaussian, and, therefore, it is the optimal
distribution for X.
Using these facts, the capacity formula can be proved by obtaining explicit
expressions for H(Y) and H(Y|X).
NTS
= H(Y) H(Y|X)
= H(Y) H(HX + Z|X)
= H(Y) H(HX|X) H(Z|X)
| {z }
0
= H(Y) H(Z)
NTS
From Result 2 we know that the entropy H(Y) is maximized for the complex
circular Gaussian input distribution, thus
max
I (X; Y)
max
H(Y) H(Z)
NTS
P = P,
I=1
and, therefore
1
C = B log det I + 2 HPHH
|H|2 P
= B log 1 +
2
This is the classical Shannon capacity formula for a bandlimited channel!
NTS
If the channel matrix H is known at the transmitter then the in general unequal
powers should be chosen, and P is not a scaled identity matrix.
Eigenchannels and power allocation using water-filling should be used as discussed
above.
NTS
If the channel matrix H is unknown at the transmitter, then if follows from the
symmetry reasons that P should be scaled identity matrix. Using the power
constraint
tr{P} = P
we obtain that P has to be chosen as
P = (P/N)I
Indeed, the power constraint is satisfied because
tr{P} = tr {(P/N)I} = (P/N)tr{I} = P
NTS
NTS
(
M )
P
= B log
1+ 2
P
= MB log 1 + 2
NTS
Consider the general MIMO channel capacity formula. Let the eigendecomposition
of the positive semi-definite Hermitian matrix HPHH be
HPH
r
X
H
i ui uH
i = UU
i=1
NTS
1
1
H
= B log det I + 2 U U = B log det I + 2
)
( r
r
X
Y
= B log
(1 + i / 2 ) = B
log 1 + i / 2
i=1
i=1
NTS
The latter formula interprets the capacity of the MIMO channel as the sum of
capacities of r parallel SISO channels.
Assuming the case of uninformed transmitter (P = (P/N)I), r can be interpreted
as the rank of H full rank channels are preferable!
NTS
Assume M = N and let the Frobenius norm of H be given. What type of channel
will maximize the MIMO capacity?
Result: The capacity is maximized in the case when H is orthogonal:
HH H = HHH = I
where is a constant. In this case,
C
N
P
P
I = B log 1 + 2
= B log det
1+ 2
N
N
P
= NB log 1 + 2
N
NTS
1
H
= B log 1 + 2 Ph h
P
2
= B log 1 + 2 khk
Hence, the SIMO channel comprises only one spatial data pipe. The addition of
receive antennas yields only a logarithmic (rather than linear) increase of capacity.
NTS
Consider a MISO row-vector channel h with one receive and N transmit antennas.
The capacity formula becomes
1
H
C = B log 1 + 2 hPh
1
1/2 2
= B log 1 + 2 khP k
The situation is similar to that in the SIMO case. The increase in capacity is only
logarithmic (rather than linear).
NTS
The channel matrix H is no longer fixed, but is treated as random. The capacity
formula can be averaged over H:
1
EH {C } = B EH log det I + 2 HPHH
Result (Telatar, 1999): Let H be a Gaussian random matrix with i.i.d. elements.
Then, the average capacity is maximized subject to the power constraint
tr{P} P when
P
P= I
N
That is, to maximize the average capacity, the antennas should transmit
uncorrelated streams with the same power an intuitively appealing fact.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 138
NTS
Introduce P = P + Poff
C
1
max EH log det{I + 2 HP HH +
P:trPP
1
=
max EH log det{I + 2 HP HH
P :trP P
1
1
H
HP
H
}
off
2
1
+ 2 HPoff HH }
1
H
+ 2 HPoff H }
NTS
1
1
H
H
1
1
H
H
=
max log det{EH I + 2 HP H + EH 2 HPoff H }
P :trP P
|
{z
}
=0
where the last term in the second equation is identical zero due to the statistical
independence of the entries in H.
We conclude that restricting the transmit covariance to exhibit the diagonal
structure P = P does not reduce the achievable capacity.
NTS
Thus
C
1
H
=
max EH log det{ I + 2 HP H }
P :trP P
We can show that due to the i.i.d. property of H the objective function is
symmetric w.r.t. the input variable, i.e. exchanging the order of the entries
P1 , ... , PN does not change the function value. Further the function is concave.
We conclude that the optimal power allocation strategy in this case is to equally
distribute the power among the transmitted symbols, e.g. to choose
P1 = P2 = ... = PN .
NTS
Note that the latter choice of P coincides with our earlier choice of this matrix in
the case of fixed channel and uninformed transmitter.
Choosing P = (P/N) I, the maximal average capacity (which is commonly
referred to as ergodic capacity) becomes
P
H
CE = B EH log det I + 2 HH
N
Ergodic capacity has an important advantage w.r.t. fixed-channel capacity as it
gives an average rather than an instantaneous picture.
NTS
Using the parallel SISO channel interpretation and denoting the singular values of
H as i , we obtain
" r
#
X
Pi2
CE = B EH
log 1 + 2
N
i=1
r
X
P 2
= B
EH log 1 + 2 i
N
i=1
Please, note the difference with the water-filling capacity. In contrast to it, in the
latter expression equal powers are used for each eigen-channel.
NTS
r
X
i=1
SNRi2
EH log 1 +
N
Assume M = N and i.i.d. Rayleigh fading. Then, using random matrix theory, it
can be obtained that for any SNR
lim
CE
= const
N
NTS
Outage capacity
A value Cout which is larger than the capacity C in pout percents of channel
realizations. In other words,
Pr(Cout > C ) = pout
If one wants to transmit with Cout bits per second, then the channel capacity is
less than Cout with the probability pout . Hence, the transmission is impossible (the
system is in outage) in pout 100 percents of time.
Alternatively, we can write
Pr(Cout C ) = 1 pout
and, hence, in (1 pout ) 100 percents of time the transmission is possible as the
system is not in outage.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 145
NTS
Outage capacity
tr{P}P
where we additionally use the opportunity to minimize the outage probability by
means of a proper choice of P. This particular choice, of course, depends on the
statistics of the random channel matrix H.
NTS
10
Outage capacity (bits / s / Hz)
10
10
10
pout = 0.01
pout = 0.1
pout = 0.5
10
10
10
20
SNR (dB)
30
10
10
pout = 0.01
pout = 0.1
pout = 0.5
40
10
10
10
20
SNR (dB)
30
40
NTS
MULTIUSER CHANNELS
In multiuser channels, one user can interfere to another user. This type of
interference is called multiuser interference (MUI).
Multiple-access channels
Broadcast channels
Relay channels
NTS
Multiple-access channel
NTS
Broadcast channel
NTS
Relay channel
NTS
Multiple-access channels
Z (i) NC (0, 2 )
In the point-to-point (single user) case, the rate limit is the channel capacity. The
achievable rate region is, therefore, given by:
P
R < B log 1 + 2
In the two-user case, we should extend this concept to a capacity region C which
is a set of all pairs (R1 , R2 ) such that users 1 and 2 can simultaneously reliably
communicate at rates R1 and R2 , respectively.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 152
NTS
Multiple-access channels
Since the two users share the same bandwidth, there is a tradeoff between the
rates R1 and R2 : if one user wants to communicate at a higher rate, then the
other user may need to lower its rate.
Example of tradeoff: In orthogonal multiple access schemes such as OFDM, the
tradeoff can be achieved by varying the number of subcarriers allocated to each
user.
NTS
Rate region
Different scalar performance measures can be obtained from the capacity region:
I
is the maximum common rate at which both users can simultaneously reliably
communicate.
I
max (R1 + R2 )
(R1 ,R2 )C
NTS
Rate region
If we have two users with the powers P1 and P2 , then the capacity region for the
two-user channel is defined by the following inequalities:
P1
R1 < B log 1 + 2
P2
R2 < B log 1 + 2
P1 + P2
R1 + R2 < B log 1 +
2
The first two constraints say that the rate of each individual used cannot exceed
the capacity of the point-to-point link with the other user absent.
The last constraint says that the total throughput cannot exceed the capacity of
the point-to-point link with a single user defined as the sum of the two users.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 155
NTS
Rate region
That is, not only the rates R1 and R1 are limited, but their sum is limited as well.
This means that the signal of each user may be viewed as an interference for
another user.
Result: The two-user capacity region is a pentagon.
NTS
NTS
P2
P1 + 2
P2
P1
1+ 2 1+
P1 + 2
P1
P2
P1 P2
= B log 1 + 2 +
+
P1 + 2
2 (P1 + 2 )
P 2 + P1 2 + P2 2 + P1 P2
= B log 1 + 1
2 (P1 + 2 )
P1 + P2
= B log 1 +
2
R1 + R2 = B log
NTS
Each user should encode its data using a capacity achieving channel code. The
receiver should decode the information of both users in two stages:
I
In the first stage, the data of user 2 are decoded treating user
1 as AWGN.
2
Then, the maximum rate of user 2 can achieve R2 = B log 1 + P1P+
.
2
NTS
If one reverses the order of cancellation then one can achieve point B rather than
A.
All other rate points on the segment AB can be obtained by time-sharing between
the multiple-access strategies of points A and B.
The segment AB contains all the optimal operating points of the channel, in the
sense that any point in the capacity region is dominated by some point on AB.
That is, for any point within the capacity region that corresponds to the rates R1
and R2 we can always find a point on the segment AB whose rates R1 and R2
satisfy:
R1 R1 ,
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 160
R2 R2
NTS
Pareto-optimal
One can always increase the user rates to move to a point on the segment AB,
and there is no reason not to do this.
NTS
To maximize the sum capacity Csum , any point on AB is equally fine. Note
that we have already computed the sum of R1 and R2 in the point A. Hence,
P1 + P2
Csum = B log 1 +
2
NTS
P1 + 2
or
R1 = B log 1 +
P1
P2 + 2
P1
P2 + 2
P2
, R2 = B log 1 + 2
, R2 = B log 1 +
P2
P1 + 2
That is, one of the rates in the CDMA case is always lower than in the case of
successive cancellation!
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 164
NTS
2
2 (P1 + 2 )(P2 + 2 )
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 165
NTS
Y (i) =
K
X
Xk (i) + Z (i),
Z (i) NC (0, 2 )
k=1
Similar to the two-user case, in the case of K users, all of them share the same
bandwidth, and there is a tradeoff between the rates Rk (k = 1, 2, ... , K ). If one
(or more) users want to communicate at higher rate(s), then the other user(s)
may need to lower their rate(s).
NTS
Pk + Pi
Rk + Ri < B log 1 +
, k, i = 1, ... , K
2
Pk + Pi + Pi
Ri + Rk + Rl < B log 1 +
, k, i, l = 1, ... , K
2
!
PK
K
X
k=1 Pk
Rk < B log 1 +
2
k=1
NTS
is the maximum sum rate that can be achieved by a single transmitter with the
total power of the users in S and with no other users in the system.
NTS
K
X
max
(R1 ,...,RK )C
PK
Csum = B log 1 +
RK
k=1
k=1
2
Pk
and that there are exactly K ! corner points in the capacity region, each one
corresponding to a different successive cancellation order among the users.
NTS
KP
Csum = B log 1 + 2
Observe that the sum capacity is unbounded as the number of users grows. In
contrast, in the conventional CDMA receiver (decoding each user treating all the
other users as noise), the sum rate will be only
P
BK log 1 +
(K 1)P + 2
which approaches
BKP
log e ' B log e
(K 1)P + 2
as K . The growing interference is a limiting factor here!
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 170
NTS
max
(R,R,...,R)C
This rate for each user can be obtained by orthogonal multiplexing where each
user is allocated a fraction 1/K of the total degrees of freedom (for example, of
the total bandwidth B).
Note that Csym = CKsum
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 171
NTS
Broadcast channels
k = 1, 2;
Zk (i) NC (0, 2 )
where hk is the fixed complex channel gain corresponding to the kth user.
The broadcast case is often referred to as downlink.
The transmit power constraint: the average power of the transmit signal is P.
As in the multi-access (uplink) channel case, we can define the capacity region C
as the region of rates (R1 , R2 ) at which both users can simultaneously reliably
communicate.
NTS
Broadcast channels
k = 1, 2
For any k, this upper bound on Rk can be attained by using all the transmit
power to communicate to user k (with the rate of the remaining user being zero).
Thus, we have two extreme points:
P|h1 |2
R1 = B log 1 +
, R2 = 0
2
P|h2 |2
R2 = B log 1 +
, R1 = 0
2
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 173
NTS
Further, we can share the degrees of freedom (time and bandwidth) between the
users in an orthogonal manner to obtain any rate pair on the line joining these two
extreme points.
Hence, for the symetric ase of |h1 | = |h2 | the capacity region is a triangle.
NTS
NTS
In the symmetric case |h1 | = |h2 | , |h| sum rate can be shown to be bounded by
the single-user capacity:
P|h|2
R1 + R2 < B log 1 +
2
The latter conclusion follows from the triangle form of the capacity region.
NTS
As have been already mentioned, the rate pairs in the capacity region can be
achieved by sharing the degrees of freedom (bandwidth and time) between the
two users. What are the alternative ways to achieve the boundary of the capacity
region?
The channel symmetry suggests an alternative natural approach:
I
Let the channel of the user 2 be stronger than that of user 1 (|h1 | < |h2 |).
Thus, if user 1 can successfully decode its data from Y1 , then user 2 (which
has higher SNR) should also be able to decode the data of user 1 from Y2 .
Then, user 2 can subtract the data of user 1 from its received signal Y2 to
better decode its own data; i.e., it can perform successive interference
cancellation.
NTS
Consider the following transmission strategy that superposes the signals of two
users, much like in a spread-spectrum CDMA system. The transmitted signal is
the sum of two signals:
X (i) = X1 (i) + X2 (i)
where Xk (i) is the signal intended for user k.
NTS
Superposition coding
Weaker user 1 decodes it own signal by treating the signal for user 2 as noise.
Stronger user 2 performs successive interference cancellation: it first decodes the
data of user 1 by treating X2 as noise, subtracts the so-determined signal of user 1
from Y2 , and then extracts its own data. As a result, for any possible power split
of P = P1 + P2 , the following rate pair can be achieved
P1 |h1 |2
R1 = B log 1 +
P2 |h1 |2 + 2
P2 |h2 |2
R2 = B log 1 +
2
This strategy is commonly referred to as superposition coding.
NTS
Orthogonal scheme
On the other hand, in orthogonal schemes, for any power split P = P1 + P2 and
degree-of-freedom split [0, 1], the following rates are jointly achieved
P1 |h1 |2
R1 = B log 1 +
2
P2 |h2 |2
R2 = (1 )B log 1 +
(1 ) 2
Here, can be interpreted, for example, as the fraction of bandwidth (e.g. both
bandwidth B and noise power reduced by factor ) assigned to user 1. Alternative
can be interpreted as a fraction of time assigned to user 1 (e.g bits per 1
seconds becomes bits per second and signal power P1 is consumed in fraction
second of time).
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 180
NTS
Assume that superposition coding is used and that the power is split such that
P P1 + P2 . In this case if user 1 can decode its data treating the data of user 2
as noise, then also user 2 can decode the data of user 1, substract it from its
received signal and decoding its own data. Hence the the following rates pairs are
suported.
P1 |h1 |2
(P1 + P2 )|h1 |2
P2 |h1 |2
) = B log(1 +
) B log(1 +
)
2
2
2
P2 |h1 | +
2
P2 |h2 |2
R2 B log(1 +
)
2
Thus for |h1 | = |h2 | = |h| and the power constraint P P1 + P2 the sum capacity
is given by
P|h|2
)
R1 + R2 B log B log(1 +
2
. April 2012 | NTS TU Darmstadt | Marius Pesavento | 181
19.
NTS
R1
B log(1 +
NTS
In the K -user broadcast case, the boundary of the capacity region can be proved
to be given by
!
Pk |hk |2
, k = 1, ... , K
Rk = log 1 +
PK
2 + ( l=k+1 Pl )|hk |2
for all possible power splits P =
PK
k=1
NTS
Fading channels
Until now, all multi-user channels have been considered without random channel
fading.
Let us now include fading in the signal model. Channel state information issue is
critical in such cases.
NTS
K
X
k=1
NTS
Slow fading
NTS
Fast fading
NTS
!)
|hk |2 P
Csum = E B log 1 +
2
nP
o
K
2
E
P
k=1 |hk |
B log 1 +
2
KP
= B log 1 + 2
2
where the property E |hk (m)| = 1 (k = 1, ... , K ) has been used in the last line.
The last expression can be identified as the sum capacity of the AWGN
multiple-access channel. Hence, without channel state information at the
transmitter, fading can only hurt.
PK
k=1
NTS
|hk |2 K
k=1
and the penalty due to fading vanishes. Basically, the effect of fading is averaged
over a large number of users.
NTS
Let us now assume that we have full (possibly also non-causal) channel state
information at both the transmitter and receiver sides.
Block-fading model:
K
X
Y (i) =
hk (i)Xk (i) + Z (i)
k=1
where hk (i) = hk,l remains constant over the lth coherence channel period of Tc
(Tc 1) symbols and is i.i.d. across different coherence periods.
NTS
For a given realization of the channel gains hk,l (k = 1, ... , K ; l = 1, ... , L), the
sum capacity is given by
!
PK
L
2
BX
k=1 Pk,l |hk,l |
log 1 +
max
2
{Pk,l } L
l=1
1X
Pk,l = P,
L
k = 1, ... , K
l=1
NTS
This leads to a variable rate scheme: in each lth sub-channel, the rates that are
dictated by the above optimization problem are used.
Optimal strategy: The sum rate in the lth sub-channel
!
PK
2
k=1 Pk,l |hk,l |
B log 1 +
2
PK
for a given total power k=1 Pk,l allocated to this sub-channel is maximized by
giving all this power to the user with the strongest channel gain. That is, each
time only one user with the best channel is allowed to transmit. Under this
strategy, the multiuser channel for each time l reduces to a point-to-point channel
with the channel gain
max |hk,l |2
k=1,...,K
NTS
k = 1, ... , K
k = 1, ... , K
NTS
Let us first consider the case when the channel state information is available only
at the receiver.
We have the following single-user bounds:
P|h|2
Rk < B E log 1 +
, k = 1, ... , K
2
where h is a random channel gain.
For any k, this upper bound on Rk can be attained by using all the transmit power
to communicate to user k (with the rate to the remaining users being zero). Thus,
as in the non-fading case, we have K extreme points of the capacity region.
NTS
Similar to the non-fading case, it can be shown that the sum rate is also bounded
by the same quantity
K
X
k=1
P|h|2
Rk < B E log 1 +
2
NTS
Let us now consider the case when the channel state information is available both
at the transmitter and receiver.
Let us focus on the sum capacity. As in the uplink case, it can be shown that the
sum capacity is achieved by transmitting only to the best user at each time. Under
this strategy, the downlink channel reduces to a point-to-point channel with the
channel gain
max |hk |2
k=1,...,K
NTS
Multiuser diversity
We have seen that in the full channel state information case, from the sum
capacity perspective, the optimal strategy both in the uplink and downlink cases
reduces the multiuser case to the single-user (point-to-point) case with the fading
of magnitude maxk |hk (i)|. Compared to a system with a single user, the multiuser
diversity gain comes from:
I
the improvement of the effective channel gain at time i from |hk (i)|2 to
maxk=1,...,K |hk |2 .
The second effect appears entirely due to the ability to dynamically schedule
resources among the users as a function of the channel state.
NTS
Remarks
The multiuser diversity gain comes from the following effect: when many
users fade independently, at any time there is a high probability that one of
them has a strong channel. By allowing only that user to transmit or, vice
versa, transmitting only to that user, the shared channel resource is used in
the most efficient manner, and the total throughput is maximized.
The larger the number of users, the higher is the multiuser diversity gain.
The amount of multiuser diversity gain depends critically on the tail of the
distribution of |hk |2 : the heavier the tail, the more likely there is a user with
the strong channel, and the larger the multiuser diversity gain.
NTS
the base station has to access the channel quality of each user:
I
in downlink, each user has to track its own channel SNR and feed back the
channel quality to the base station.
in uplink, the base station has to track the user channel quality (user SNRs).
the base station has to schedule transmissions among the users as well as to
adapt the data rate as a function of instantaneous channel quality.
NTS
NTS
All scheduling decisions are done as a function of user channel states. Hence,
the quality of channel estimation is a primary issue, and feedback from the
users to the base station is needed in the downlink case.
Both the error in channel measurement and the delay/error in feeding the
channel state back are significant bottlenecks of practical applications of the
multiuser diversity strategy.
We have observed that the use of multiuser diversity strategy requires fading
to be rich and fast. Not useful for line-of-sight scenarios or cases with little
scattering or slowly changing environments.
NTS
In the ith time-slot, the base station receives the requested/supportable rates
Rk (i) (k = 1, ... K ) from all users, and transmits to the user k with the
largest
= Rk (i)/Tk (i)
NTS
Divides the users in several classes (say, in two classes depending on whether
they are near to the base station or near the cell edge). Then, users in each
class have statistically comparable channel strengths.
NTS
We have already discussed the linear block channel codes in the Information
Theory I. Now, we will discuss cyclic codes as well as convolutional codes.
NTS
Cyclic codes
NTS
An (n, k) linear block code C is called a cyclic code if every cyclic shift of any
codeword in C is also a codeword in C .
Properties:
I Linearity: the sum of any two codewords is also a codeword;
I Cyclic property: Any cyclic shift of any codeword is also a codeword.
To develop the theory of cyclic codes, let us treat the components of the
codeword c as the coefficients of the following polynomial:
c(X ) = c0 + c1 X + + cn1 X n1
where X is an indeterminate.
The fact that all ci are binary is taken into account by using the binary arithmetic
for all polynomial coefficients when operating with polynomials.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 206
NTS
Cyclic codes
NTS
Cyclic codes
NTS
Cyclic codes
The polynomial c (i) (X ) can be recognized as the code polynomial of the codeword
c(i) obtained by applying i cyclic shifts to the codeword c.
Moreover, from the latter equation, we readily see that c (i) (X ) is the remainder
that results from dividing X i c(X ) by (X n + 1).
Hence, we may formally state the cyclic property in polynomial notation as
follows: if c(X ) is a code polynomial, then the polynomial
c (i) (X ) = X i c(X ) mod(X n + 1)
is also a code polynomial for any cyclic shift i, where mod(X n + 1) stands for
modulo-(X n + 1) multiplication.
NTS
Cyclic codes
Note that n cyclic shifts of any codeword does not change it, which means that
X n = 1, and hence X n + 1 = 0 in modulo-(X n + 1) arithmetics!
Generator polynomial: a polynomial g(X ) of minimal degree that completely
specifies the code and is a factor of X n + 1. The degree of g(X ) is equal to the
number of parity-check bits of the code, n k.
It can be shown that any cyclic code is uniquely determined by its generator
polynomial in that each code polynomial in the code can be expressed in the form
of a polynomial product as follows:
c(X ) = a(X )g (X )
where a(X ) is a polynomial of degree k 1.
NTS
Cyclic codes
NTS
Cyclic codes
NTS
NTS
k1
X
hi X i + X k
i=1
NTS
NTS
That is,
a(X ) = X + X 3 ,
b(X ) = X + X 2
NTS
for the considered (7, 4) code, we can construct the generator matrix from
generator polynomial by using
g (X ) = 1 + X + X 3
Xg (X ) = X + X 2 + X 4
X 2 g (X ) = X 2 + X 3 + X 5
X 3 g (X ) = X 3 + X 4 + X 6
as the rows of the 4 7 generator matrix
1 1 0 1
0 1 1 0
G=
0 0 1 1
0 0 0 1
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 217
0
1
0
1
0
0
1
0
0
0
0
1
NTS
1
1
1
0
0
1
1
1
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
1
Decoding cyclic codes can be made in the same way as for any other linear block
codes, e.g., using syndrome.
NTS
Popular cyclic codes are the so-called cyclic redundancy check (CRC) codes,
Bose-Chaudhuri-Hocquenghem (BCH) codes, and non-binary Reed-Solomon (RS)
codes. The are parts of different international communication standards, e.g.,
digital subscriber line (DSL) standards.
NTS
Convolutional codes
NTS
Convolutional codes
The message sequence m = [m0 , m1 , m2 , ...] enters the encoder one bit at a time.
The encoder output sequences are obtained as the convolution of the input
sequence with the encoder generator sequences. For an encoder with the memory
order M, the length of these sequences is M + 1. For example, in the case of two
impulse generator sequences,
(0)
(0)
(1)
(1)
c(1) = m g(1)
where denotes the discrete convolution and all operations are modulo-2.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 221
NTS
Convolutional codes
cl
M
X
(j)
mli gi ,
j = 0, 1
i=0
After encoding, the output sequences are multiplexed into a single sequence called
the codeword
(0) (1) (0) (1)
c = [c0 , c0 , c1 , c1 , ... ]
NTS
Convolutional codes
Defining a matrix
(0)
g0
G =
(1)
g0
(0)
g1
(0)
g0
(1)
g1
(1)
g0
..
.
(0)
g1
(0)
gM
(1)
g1
..
.
(1)
gM
(0)
gM
(1)
gM
..
where all blank areas are zeros, we can rewrite the encoding equations in matrix
form as
c = mG
This form of this equation is equivalent to that of linear block codes! Therefore,
we call G the generator matrix of the code.
In the case of semi-infinite message sequence, the matrix G is semi-infinite as well.
However, if m is finite-length, then G becomes finite-length as well.
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 223
NTS
NTS
11 01 11 11
11 01 11
G =
11 01
11
and obtain the same codeword
11 01 11 11
11 01 11
c = [10111]
11 01
11
generator matrix as
11
11 11
01 11 11
11 01 11 11
as
11
11 11
01 11 11
11 01 11 11
= [11 01 00 01 01 01 00 11]
NTS
Let us discuss the concepts of code tree and trellis using a particular example of
the R = 1/2 convolutional code with M = 2 and the impulse responses
g(0) = [111],
g(1) = [101]
Consider the input sequence m = [10011]. Similar to the example above, it can be
shown that the codeword becomes
c = [11 10 11 11 01 01 11]
To enforce the R = 1/2 property, let us truncate the codeword by dropping the
last 2M = 4 bits (the effect of truncation becomes negligible if longer messages
and codewords are used). Then, the codeword becomes [11 10 11 11 01]
19. April 2012 | NTS TU Darmstadt | Marius Pesavento | 226
NTS
Convolutional Encoder
NTS
The code tree is defined as follows: each branch of the tree represents an input
symbol (0 or 1). The corresponding output (coded) symbols are indicated on each
branch. A specific path can be traced for each message sequence. The
corresponding coded symbols on the branches following this path form the output
sequence.
NTS
Code tree
NTS
State diagram
NTS
Trellis diagram
NTS
Trellis diagram
NTS
Trellis diagram
NTS
Trellis diagram
NTS
Trellis diagram
NTS
Over L binary intervals, the total number of comparisons made by the Viterbi
algorithm is 2K 1 L, rather than 2L comparisons required by the standard
maximum-likelihood procedure (full tree search).
NTS
Let a(d) denote the number of pathes with a Hamming distance d deviating from,
and then returning to, the all-0 test path. The error probability of Pe of deviating
from the correct path is then upper bounded by
Pe <
a(d)Pd
d=dF
where Pd denotes that probability that d bits are received in error and dF denotes
the minimum free distance.
Inequality sign because pathes a not mutually exclusive.
Pe depends critically on the minimum free distance dF !
NTS
CONCLUSION
NTS