Discrete Memory less Channel

Engineering world
4 minute read
0

 DISCRETE MEMORYLESS CHANNEL:

 

·        Transmission rate over a noisy channel

 

Repetition code

 

Transmission rate

 

·        Capacity of DMC

 

Capacity of a noisy channel

 

Examples

Ø   All these transition probabilities from xi to yj are gathered in a transition matrix.

 

Ø   The (i ; j) entry of the matrix is P(Y = yj /jX = xi ), which is called forward transition probability.

 

Ø   In DMC the output of the channel depends only on the input of the channel at the same instant and not on the input before or after.

 

Ø   The input of a DMC is a RV (random variable) X who selects its value from a discrete limited set X.

 

Ø   The cardinality of X is the number of the point in the used constellation.

 

Ø   In an ideal channel, the output is equal to the input.

 

Ø   In a non-ideal channel, the output can be different from the input with a given probability.

 

·        Transmission rate:

Ø   H(X) is the amount of information per symbol at the input of the channel.

 

Ø   H(Y ) is the amount of information per symbol at the output of the channel.

 

Ø   H(XjY ) is the amount of uncertainty remaining on X knowing Y .

 

Ø   The information transmission is given by:I (X; Y ) = H(X)  H(XjY ) bits/channel use

 

Ø   For an ideal channel X = Y , there is no uncertainty over X when we observe Y . So all the information is transmitted for each channel use: I (X;Y ) = H(X)

Ø   If the channel is too noisy, X and Y are independent. So the uncertainty over X remains the same knowing or not Y , i.e. no information passes through the channel: I (X; Y ) = 0.

 

·        Hard and soft decision:

 

Ø   Normally the size of constellation at the input and at the output are the same, i.e., jXj = jYj

 

Ø   In this case the receiver employs hard-decision decoding.

 

Ø   It means that the decoder makes a decision about the transmitted symbol.

 

Ø   It is possible also that jXj 6= jY j.

 

Ø   In this case the receiver employs a soft-decision.

 

ü   Channel models and channel capacity:


 

1.     The encoding process is a process that takes a k information bits at a time and maps each k-bit sequence into a unique n-bit sequence. Such an n-bit sequence is called a code word.

 

2. The code rate is defined as k/n.

 

3.     If the transmitted symbols are M-ary (for example, M levels), and at the receiver the output of the detector, which follows the demodulator, has an estimate of the transmitted data symbol with

 

(a). M levels, the same as that of the transmitted symbols, then we say the detector has made a hard decision;

 

(b). Q levels, Q being greater than M, then we say the detector has made a soft decision.

ü   Channels models:

 

1. Binary symmetric channel (BSC):

 

If (a) the channel is an additive noise channel, and (b) the modulator and demodulator/detector are included as parts of the channel. Furthermore, if the modulator employs binary waveforms, and the detector makes hard decision, then the channel has a discrete-time binary input sequence and a discrete-time binary output sequence.


Note that if the channel noise and other interferences cause statistically independent errors in the transmitted binary sequence with average probability p, the channel is called a BSC. Besides, since each output bit from the channel depends only upon the corresponding input bit, the channel is also memoryless.

2. Discrete memoryless channels (DMC):

 

A channel is the same as above, but with q-ary symbols at the output of the channel encoder, and Q-ary symbols at the output of the detector, where Q ³ q . If the channel and the modulator are memoryless, then it can be described by a set of qQ conditional probabilities

 

(= y i  | X = x j ) º P ( y i  | x j ), i = 0,1,...,- 1; j = 0,1,..., q -1

 

Such a channel is called discrete memory channel (DSC).


If the input to a DMC is a sequence of n symbols u1 , u2 ,..., un selected from the alphabet X and the corresponding output is the sequence v1 , v 2 ,..., vn of symbols from the alphabet Y, the joint conditional probability is


the probability transition matrix for the channel.

 

3. Discrete-input, continuous-output channels:

 

Suppose the output of the channel encoder has q-ary symbols as above, but the output of the detector is unquantized (Q = ¥) . The conditional probability density functions

 

( y | X = x k ),      = 0,1, 2,..., q -1

 

AWGN is the most important channel of this type.

 

= X + G


For any given sequence X i , i = 1, 2,..., n , the corresponding output is Yi , i = 1, 2,..., n

 

Yi  = X i + Gi , i = 1, 2,..., n

 

If, further, the channel is memoryless, then the joint conditional pdf of the detector‘s output is


 

4. Waveform channels:

 


If such a channel has bandwidth W with ideal frequency response C ( f ) = 1 , and if the bandwidth-limited input signal to the channel is x ( t) , and the output signal, y ( t) of the channel is corrupted by AWGN, then

( t ) = x ( t ) + n ( t)

 

The channel can be described by a complete set of orthonormal functions:



Since { ni } are uncorrelated and are Gaussian, therefore, statistically independent. So



ü   Channel Capacity:

Channel model: DMC

 

Input alphabet: X = {x0 , x1 , x 2 ,..., xq-1}

 

Output alphabet: Y = {y 0 , y1 , y 2 ,..., yq-1}

 

Suppose x j is transmitted,  yi is received, then

The mutual information (MI) provided about the event {X = x j } by the occurrence of the event


Hence, the average mutual information (AMI) provided by the output Y about the input X is


To maximize the AMI, we examine the above equation: 


(1). P ( y i) represents the jth output of the detector; 

(2). P ( y i  | x j ) represents the channel characteristic, on which we cannot do anything;

(3). P ( x j ) represents the probabilities of the input symbols, and we may do something or control them. Therefore, the channel capacity is defined by

Post a Comment

0Comments
Post a Comment (0)