Binary-input Memoryless Channels

Theory

What is a Communication Channel?

A communication channel is a medium through which communication happens. In this virtual lab, we are dealing with specifically those channels that accept binary-valued inputs. We call these channels as binary-input channels. For some binary channels, we write the possible set of inputs as the logical bits {0,1}\{0,1\}. That is, at any time instant, we can send a logical "0" through the channel, or a logical "1". Equivalently, we may also write the binary alphabet in the bipolar form, which is written as {+1,1}\{+1,-1\}. Normally, we take the logical-bit to bipolar mapping as 0+10\to +1 and 111\to -1.

We generally use the notation X\cal X to denote the input alphabet of the channel. From the point of view of the receiver, the input to the channel is unknown, and hence is modelled as a random variable with some input probability distribution. We denote this input random variable as XX. Similarly, the output of the channel, is a random variable denoted by YY. We assume that the output alphabet, the set of all values that the output can possibly take, is denoted by Y\cal Y.

Types of Channels considered in this virtual lab

The problem of designing good communication systems arises precisely due to the existence of noise in communication channels. The noise in the communication channel is generally modelled via the conditional probabilities (of the output value, given the input value). We consider some three important types of communication channels (or in other words, noise models) in this virtual lab.

  1. The Binary Erasure Channel: The noise in this channel is modelled as a bit-erasure, which denotes the transmitted bit was erased or lost. Formally, in this channel, the input alphabet is the set of logical bits, i.e., X={0,1}{\cal X}=\{0,1\} and the output alphabet is the set of logical bits along with the erasure symbol ??, i.e., Y={0,1,?}{\cal Y}=\{0,1,?\}. The erasure symbol ?? denotes that the input symbol was erased during the process of transmission through the channel. The binary erasure channel, denoted formally as BEC(ϵ)BEC(\epsilon), has the property that the bit that is transmitted is erased with probability ϵ\epsilon. Here, ϵ\epsilon denotes the erasure probability, and we assume ϵ\epsilon is a real number between 00 and 11, i.e., ϵ(0,1)\epsilon\in(0,1).

Binary Erasure Channel

Depiction of a Binary Erasure Channel. The left side denotes the possible inputs {0,1} and the right denotes the possible outputs {0,1,ϵ} . The arrows indicate possible transitions when the bit passes through the channel. The values ϵ, (1-ϵ) marked upon the respective arrows indicates the probability of such a transition.


  1. The Binary Symmetric Channel: In this channel, the input alphabet is the set of logical bits, i.e., X={0,1}{\cal X}=\{0,1\} and the output alphabet is the set of logical bits i.e., Y={0,1}{\cal Y}=\{0,1\}. The noise of this channel is characterized by bit-flips (i.e., a transmitted 00 bit is received as a 11, or vice-versa). In the binary symmetric channel denoted by BSC(p)BSC(p), we assume that bit-flip happens with some probability pp, where pp is a real number and p(0,1)p\in(0,1).

Binary Symmetric Channel

Depiction of a Binary Symmetric Channel. The left side denotes the possible inputs {0,1} and the right denotes the possible outputs {0,1}. The arrows indicate possible transitions when the bit passes through the channel. The values p, (1-p) marked upon the respective arrows indicates the probability of such a transition.


  1. The Additive White Gaussian Noise Channel (AWGN): The AWGN channel, accepts a real number as an input, and adds to it a noise random variable ZZ that is distributed independently according to a Gaussian distribution N(0,N0/2){\cal N}(0,N_0/2), with zero-mean and variance N0/2N_0/2. Thus, the input alphabet and the output alphabet are both X=R{\cal X}=\mathbb{R}. The relationship between the input XX and the output YY is then given as

Y=X+Z.Y=X+Z.


Gaussian Channel

Depiction of a Additive White Gaussian Noise Channel. The left side denotes the possible input X and the right denotes the possible output Y. The arrows indicate the input X getting added to a noise Z to give the output Y.


Conditional Distribution Associated with the Communication Channel

We can also describe the channels above using the conditional distribution of the output random variable YY given by the input random variable XX. Specifically, we have the following.

  1. Binary Erasure Channel: The conditional distribution of this channel BEC(ϵ)BEC(\epsilon) is given as follows.

pYX(yx)={1ϵif y=x,x{0,1},ϵif y=?. p_{Y|X}(y|x)= \begin{cases} 1-\epsilon&\text{if}~ y=x, \forall x \in\{0,1\},\\ \epsilon & \text{if}~y=? . \end{cases}

  1. Binary Symmetric Channel: The conditional distribution of this channel BSC(p)BSC(p) is given as follows.

pYX(yx)={1pif y=x,x{0,1},pif xy. p_{Y|X}(y|x)= \begin{cases} 1-p&\text{if}~ y=x, \forall x \in\{0,1\},\\ p & \text{if}~x\neq y . \end{cases}

  1. AWGN Channel: For this channel, we have pYX(yx)=1πN0e(yx)2N0,x,yR. p_{Y|X}(y|x)=\frac{1}{\sqrt{\pi N_0}}e^{\frac{-(y-x)^2}{N_0}}, \forall x,y \in \mathbb{R}.

The Memoryless Property of the Channels

We assume that the three channels we have considered in this virtual lab have the memoryless property and exist without feedback. To be precise, if we transmit a nn-length sequence of bits denoted by (x1,,xn)(x_1,\ldots,x_n) through any of these channels, the output is a sequence of bits (y1,,yn)(y_1,\ldots,y_n), with probability as follows.

p((y1,,yn)(x1,,xn))=p(y1x1)p(ynxn)=Πi=1np(yixi).p((y_1,\ldots,y_n)|(x_1,\ldots,x_n))=p(y_1|x_1)\cdots p(y_n|x_n)=\Pi_{i=1}^n p(y_i|x_i).

The above property is a naturally expected property. For instance, consider the channel BEC(ϵ)BEC(\epsilon). The probability of receiving (?,0,?,1)(?,0,?,1) when transmitting (1,0,0,1)(1,0,0,1) is given by p((?,0,?,1)(1,0,0,1))=p(?1)p(00)p(?0)p(11)=ϵ(1ϵ)ϵ(1ϵ)=ϵ2(1ϵ)2. \begin{aligned} p((?,0,?,1)|(1,0,0,1))&=p(?|1)p(0|0)p(?|0)p(1|1)\\ &=\epsilon\cdot(1-\epsilon)\cdot \epsilon\cdot (1-\epsilon)\\ &=\epsilon^2(1-\epsilon)^2. \end{aligned}

More generally, we can say the following. Let xF2n\bm{x}\in \mathbb{F}_2^n be an nn-length binary vector. A vector y{0,1,?}n\bm{y} \in \{0,1,?\}^n is said to be compatible with x\bm{x} if y\bm{y} and x\bm{x} agree (i.e., are equal) in all positions which are unerased (not equal to "?" symbol) in y\bm{y}. Otherwise, they are not compatible. For instance, the vectors (0,1,0,1)(0,1,0,1) and (?,1,0,?)(?,1,0,?) are compatible. However, the vectors (0,1,0,1)(0,1,0,1) and (?,1,1,?)(?,1,1,?) are not compatible, since the third bits of the two vectors are both unerased and not equal.

For any vector y{0,1,ϵ}n\bm{y}\in\{0,1,\epsilon\}^n, let we(y)w_e(\bm{y}) denote the number of erased symbols in y\bm{y}. Then for any xF2n\bm{x}\in\mathbb{F}_2^n, we have, in the memoryless BEC(ϵ)BEC(\epsilon) channel, the following to be true. p(yx)={ϵwe(y)(1ϵ)nwe(y)if x and y are compatible0otherwise.p(\bm{y}|\bm{x})=\begin{cases} \epsilon^{w_e(\bm{y})}(1-\epsilon)^{n-w_e(\bm{y})} & \text{if}~\bm{x} ~\text{and}~\bm{y}~\text{are compatible}\\ 0 & \text{otherwise}. \end{cases}

Turning our focus to the BSC(p)BSC(p), we have the following. For any x,yF2n\bm{x},\bm{y}\in\mathbb{F}_2^n, let d(x,y)d(\bm{x},\bm{y}) denote the Hamming distance (number of positions where x\bm{x} and y\bm{y} have distinct values). For example, d((1,0,1,0),(0,0,1,1))=2d((1,0,1,0),(0,0,1,1))=2 as the two vectors are distinct in the first and the fourth locations. Then, for the BSC(p)BSC(p) memoryless channel, we have the following. p(yx)=pd(x,y)(1p)nd(x,y).p(\bm{y}|\bm{x})=p^{d(\bm{x},\bm{y})}(1-p)^{n-d(\bm{x},\bm{y})}.

For the memoryless AWGN channel, we have, for any two vectors x,yRn\bm{x},\bm{y}\in\mathbb{R}^n, p(yx)=1(πN0)n/2e(yx2)N0.p(\bm{y}|\bm{x})=\frac{1}{(\pi N_0)^{n/2}}e^{-\frac{(||\bm{y}-\bm{x}||^2)}{N_0}}.