channel coding
DESCRIPTION
presentation of channel codingTRANSCRIPT
![Page 1: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/1.jpg)
Introduction to Information theorychannel capacity and models
A.J. Han VinckUniversity of Essen
October 2002
![Page 2: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/2.jpg)
content
Introduction Entropy and some related properties
Source coding Channel coding Multi-user models Constraint sequence Applications to cryptography
![Page 3: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/3.jpg)
This lecture
Some models Channel capacity
converse
![Page 4: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/4.jpg)
some channel models
Input X P(y|x) output Y
transition probabilities
memoryless:
- output at time i depends only on input at time i
- input and output alphabet finite
![Page 5: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/5.jpg)
channel capacity:
I(X;Y) = H(X) - H(X|Y) = H(Y) –H(Y|X) (Shannon 1948)
H(X) H(X|Y)
notes:capacity depends on input probabilities
because the transition probabilites are fixed
channelX Y
capacity)Y;X(Imax)x(P
![Page 6: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/6.jpg)
channel model:binary symmetric channel
Error Source
+
E
X
OutputInput
EXY
E is the binary error sequence s.t. P(1) = 1-P(0) = p
X is the binary information sequence
Y is the binary output sequence
1-p
0 0
p
1 1
1-p
![Page 7: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/7.jpg)
burst error model
Error Source
RandomRandom error channel; outputs independent P(0) = 1- P(1);
BurstBurst error channel; outputs dependent
Error SourceP(0 | state = bad ) = P(1|state = bad ) = 1/2;
P(0 | state = good ) = 1 - P(1|state = good ) = 0.999
State info: good or bad
good bad
transition probability Pgb
Pbg
PggPbb
![Page 8: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/8.jpg)
Interleaving:
Message interleaver channel interleaver -1 message
encoder decoder
bursty
„random error“
Note: interleaving brings encoding and decoding delay
Homework: compare the block and convolutional interleaving w.r.t. delay
![Page 9: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/9.jpg)
Interleaving: block
Channel models are difficult to derive:
- burst definition ?
- random and burst errors ?
for practical reasons: convert burst into random error
read in row wise
transmit
column wise
1
0
0
1
1
0
1
0
0
1
1
0
0
0
0
0
0
1
1
0
1
0
0
1
1
![Page 10: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/10.jpg)
De-Interleaving: block
read in column wisethis row contains 1 error
1
0
0
1
1
0
1
0
0
1
1
e
e
e
e
e
e
1
1
0
1
0
0
1
1
read out
row wise
![Page 11: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/11.jpg)
Interleaving: convolutional
input sequence 0
input sequence 1 delay of b elements
input sequence m-1 delay of (m-1)b elements
Example:b = 5, m = 3
in
out
![Page 12: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/12.jpg)
Class A Middleton channel model
AWGN, σ 20
…
AWGN, σ21
…
AWGN, σ22
Select channel k with probability Q(k)
I and Q same variance
I
Q
I
Q
Transition probability P(k)
0
1
0
1
![Page 13: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/13.jpg)
Example: Middleton’s class A
Pr{ σ = σ(k) } = Q(k), k = 0,1, · · ·
A ke AQ(k) :
k!
A is the impulsive index
2 2I Gand
2 21/ 2I G
2 2I G
k / A(k) : ( )
are the impulsive and Gaussian noise power
![Page 14: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/14.jpg)
Example of parameters
Middleton’s class A= 1; E = σ = 1; σI /σG = 10-1.5
k Q(k) p(k) (= transition probability )
0 0.36 0.00
1 0.37 0.16
2 0.19 0.24
3 0.06 0.28
4 0.02 0.31Average p = 0.124; Capacity (BSC) =
0.457
![Page 15: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/15.jpg)
Example of parameters
Middleton’s class A: E = 1; σ = 1; σI /σG = 10-3
Transition probability P(k)
0
1
0
1
p(k) p(k) p(k)
Q(k) Q(k) Q(k)
1
0.5
0.00.5 0.5 0.5
A = 0.1 A = 1 A = 10
1
0.5
0.0
1
0.5
0.0
![Page 16: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/16.jpg)
Example of parameters
Middleton’s class A: E = 0.01; σ = 1; σI /σG = 10-3
Transition probability P(k)
0
1
0
1
p(k) p(k) p(k)
Q(k) Q(k) Q(k)
1
0.5
0.00.5 0.5 0.5
A = 0.1 A = 1 A = 10
1
0.5
0.0
1
0.5
0.0
![Page 17: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/17.jpg)
channel capacity: the BSC
1-p
0 0
p
1 1
1-p
X Y
I(X;Y) = H(Y) – H(Y|X)
the maximum of H(Y) = 1
since Y is binary
H(Y|X) = h(p)
= P(X=0)h(p) + P(X=1)h(p)
Conclusion: the capacity for the BSC CBSC = 1- h(p)Homework: draw CBSC , what happens for p > ½
![Page 18: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/18.jpg)
channel capacity: the Z-channel
Application in optical communications
0
1
0 (light on)
1 (light off)
p
1-p
X Y
H(Y) = h(P0 +p(1- P0 ) )
H(Y|X) = (1 - P0 ) h(p)
For capacity, maximize I(X;Y) over P0
P(X=0) = P0
![Page 19: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/19.jpg)
channel capacity: the erasure channel
Application: cdma detection
0
1
0
E
1
1-e
e
e
1-e
X Y
I(X;Y) = H(X) – H(X|Y)
H(X) = h(P0 )
H(X|Y) = e h(P0)
Thus Cerasure = 1 – e
(check!, draw and compare with BSC and Z)P(X=0) = P0
![Page 20: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/20.jpg)
channel models: general diagram
x1
x2
xn
y1
y2
ym
:
:
:
:
:
:
P1|1
P2|1P1|2P2|2
Pm|n
Input alphabet X = {x1, x2, …, xn}
Output alphabet Y = {y1, y2, …, ym}
Pj|i = PY|X(yj|xi)
In general:
calculating capacity needs more theory
![Page 21: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/21.jpg)
clue:
I(X;Y)
is convex in the input probabilities
i.e. finding a maximum is simple
![Page 22: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/22.jpg)
Channel capacity
Definition:
The rate R of a code is the ratio , where
k is the number of information bits transmitted
in n channel uses
Shannon showed that: :
for R C
encoding methods exist
with decoding error probability 0
n
k
![Page 23: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/23.jpg)
System design
message estimate
channel decoder
n
Code word in
receive
There are 2k code words of length n
2k
Code book
Code book
![Page 24: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/24.jpg)
Channel capacity: sketch of proof for the BSC
Code: 2k binary codewords where p(0) = P(1) = ½
Channel errors: P(0 1) = P(1 0) = p
i.e. # error sequences 2nh(p)
Decoder: search around received sequence for codeword
with np differences
space of 2n binary sequences
![Page 25: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/25.jpg)
Channel capacity: decoding error probability
1. For t errors: |t/n-p|> Є
0 for n (law of large numbers)
2. > 1 code word in region (codewords random)
nand
)p(h1n
kRfor
02
2)12()1(P
n
)p(nhk
![Page 26: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/26.jpg)
Channel capacity: converse
For R > C the decoding error probability > 0
k/n
C
Pe
![Page 27: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/27.jpg)
Converse: For a discrete memory less channel
1 1 1 1
( ; ) ( ) ( | ) ( ) ( | ) ( ; )n n n n
n n n
i i i i i i ii i i i
I X Y H Y H Y X H Y H Y X I X Y nC
Xi Yi
m Xn Yn m‘encoder channel
channel
Source generates one
out of 2k equiprobable
messages
decoder
Let Pe = probability that m‘ m
source
![Page 28: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/28.jpg)
converse R := k/n
Pe 1 – C/R - 1/k
Hence: for large k, and R > C,
the probability of error Pe > 0
k = H(M) = I(M;Yn)+H(M|Yn)
Xn is a function of M Fano
I(Xn;Yn) +1+ k Pe
nC +1+ k Pe
1 – C n/k - 1/k Pe
![Page 29: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/29.jpg)
Appendix:
Assume:
binary sequence P(0) = 1 – P(1) = 1-p
t is the # of 1‘s in the sequence
Then n , > 0
Weak law of large numbers
Probability ( |t/n –p| > ) 0
i.e. we expect with high probability pn 1‘s
![Page 30: Channel Coding](https://reader036.vdocuments.site/reader036/viewer/2022062423/563dbb8d550346aa9aae2d63/html5/thumbnails/30.jpg)
Appendix:
Consequence:
n(p- ) < t < n(p + ) with high probability
)p(nh222
)p(n
)p(n2 2logn2log)
pn
nn2(log
t
nlog
1.
2.
3.
4.
)p(h2logn
1n2log
n
1 )p(nh22
A sequence in this set has probability
)p(nh2