1 Shannon theory
1.3 Capacity
1.3.2 Capacity of an AWGN channel
Now we study the capacity of the analog AWGN channel, assuming that we do not use any digital modulator. Then the situation is the following (see figure 1.4):
Fig.1.4: Block diagram for the AWGN channel
ο An analog symbol ΞΎ with a given probability density function π ΞΎ(π₯) is transmitted over the channel; it is assumed that the variation of ΞΎ is equal to ππ2 and the mean πΞΎ is zero, but there are no further restriction on π ΞΎ(π₯).
ο The AWGN channel adds Ξ½, a Gaussian random variable with variance ππ2 and mean value zero (the probability density function of π is denoted as π π(π₯)).
ο The receiver gets Ξ· = ΞΎ + Ξ½, a random variable with probability density function πΞ· (π₯) = π ΞΎ(π₯) β π π(π₯) (where β stands for convolution).
For the case of an analog AWGN channel, the capacity is obtained by maximizing just β(π). But we know from section 1.1.1.2 that the maximum entropy of an analog source π is obtained when the analog source has a Gaussian probability density function. In particular we showed that, for a Gaussian source π₯
β(π₯) =1
2log2(2ππππ₯2)
where ππ₯2 is the variance of π₯. In this case, if the source ΞΎ is Gaussian with zero mean, then also π = π + π is Gaussian, being the sum of two statistically independent Gaussian random variable, and π has mean equal to the sum of the means and variance equal to the sum of variance of π and π. So if π has variance
β(π) =1
2log2[2ππ(ππ2+ ππ2)]
and this is the maximum value of β(π) (for the given and fixed ππ2).
Let us then complete the evaluation of the conditional entropy β(π|π).
πΞ·|ΞΎ(π¦|π₯) = 1
β2πππ2ππ₯π {β(π¦ β π₯)2 2ππ2 } β(π|π) = β β« πβ Ξ·|ΞΎ(π€|π’)
ββ
log2πΞ·|ΞΎ(π€|π’)ππ€ = (1
2log2[2ππππ2])
In the overall, when the input is Gaussian mutual information is πΌ(ΞΎ; Ξ·) =1
2log2[2ππ(ππ2+ ππ2)] β1
2log2[2ππππ2] =1
2log2ππ2+ ππ2 ππ2 But this is also the capacity of the AWGN channel:
πΆ =1
2log2ππ2+ ππ2 ππ2
So, each time the AWGN channel is used, it carries at most πΆ information bits, and πΆ depends on the signal to noise ratio πππ2
π2: if the noise variance reduces or the source increases, then the capacity increases.
Let us consider now not just the transmission of one analog symbol π, but a sequence of symbols, and let us limit the problem to the case of a bandlimited channel, in particular a low-pass channel with bandwidth B. Then, only a process π(π‘) with bandwidth at most equal to B can pass through the channel without being distorted, and we can represent the information content of π(π‘)using just its samples, taken at sampling frequency 2B6. Then the entropy of the Gaussian source is
β(π(π‘)) = 2π΅β(π) = π΅ log2(2ππππ2)
where ππ2 is the variance of the process; remember that, if the process is statistically and ergodic, which we will assume then ππ2 does not change with time and is equal to mean power ππ of the process.
The channel output process π(π‘) is the sum of π(π‘) and the white Gaussian noise π(π‘) having power spectral density π0/2. The receiver has an initial low pass filter followed by a sampler at frequency 2B, so that we can write that the input of the detector is a sequence of samples, generated as rate 2B samples per seconds, which are the sum of the samples of π(π‘) and noise random variables with variance
ππ2 = π0
2 2π΅ = π0π΅ The entropy of π(π‘) = π(π‘) + π(π‘), sampled at rate 2B, is
β(π(π‘)) = 2π΅β(π) = π΅ log2[2ππ(ππ2+ ππ2)]
The conditional entropy is β(π(π‘)|π(π‘)) = π΅ log2[2ππππ2] as before.
The AWGN channel capacity is then
πΆβ²= β(π(π‘)) β β(π(π‘)|π(π‘)) = π΅ log2ππ2+ ππ2 ππ2 We can substitute the values of the variances and obtain
6 According to the sampling theorem that states that if a signal π₯(π‘) has bandwidth B, it is possible to exactly evaluate π₯(π‘) from its samples, provided that the sampling frequency is larger than 2π΅π₯.
πΆβ²= π΅ log2(1 + ππ π0π΅)
where now the unit of measure of πΆβ²is bits of information per second (not just bit of information).
In brief, compare to πΆ and πΆβ²:
ο πΆ is capacity per channel use πΆ =12log2(1 +πππ
0π΅) [information bit per channel use]
ο πΆβ² is capacity measured, which use the channel 2B times per second. If we do not use the low pass filter, the capacity is zero for sure.
πΆβ² = π΅ log2(1 +πππ
0π΅) [information bit per second]
Let us see if we can relate the discrete channel capacities with the AWGN channel capacity. We can imagine that process π(π‘) is the output of a digital modulator that generates bits (real bits β1β or β0β) at rate π π bits/s, so that the power ππ can be expressed as ππ = πΈππ
π= πΈππ π, where πΈπ is the energy per bit. So we have, for the AWGN channel, πΆβ² = π΅ log2(πΈπππ π
0π΅ + 1) or πΆπ΅β² = log2(πΈπππ π
0π΅ + 1).
It is not possible to get an error probability equal to zero if the input entropy is larger than the channel capacity. At most one bit transmitted by the digital modulator carries one information bit, so that we can say that the source entropy is π»(π) = π π information bits per second, and, if we assume that we are working at the limit, i.e.
the best case, with π»(π) = πΆβ², we have πΆβ²
π΅ = π π π΅ which leads to
π π
π΅ = log2(πΈππ π π0π΅ + 1)
which provides a relationship between the signal to noise ration πΈπ/π0 and the modulation efficiency π π/π΅ (measured in bits/second per hertz). In particular, we can write
πΈπ
π0 = 2π π/π΅β 1 π π/π΅
ο If π π΅π= 1, then πΈππ
0 = 1 (0 ππ΅);
ο if π π
π΅ β 0, πΈππ
0β β;
ο if π π
π΅ β β, then πΈππ
0 β ln2 (β1.6 ππ΅):
π πlim/π΅ββ
2π π/π΅β 1
π π/π΅ = lim
π π/π΅ββ
ππ π/π΅ logπ2β 1
π π/π΅ = lim
π π/π΅ββ
1 β π π/π΅ logπ2 β 1
π π/π΅ = ln 2
= 0.693
The last limit is quite interesting: it starts that it is possible to transmit with error probability equal to zero if the signal to noise ratio is πΈπ/π0 > β1.6 ππ΅, provided that the bandwidth B is infinite. Note that πΈπ/π0 = 1.6 ππ΅ in the case in which the noise variance π0/2 is equal to 0.72πΈπ, really very high. Another interesting consideration is that we can trade energy with bandwidth: if we increase the bandwidth, we can reduce πΈπ/π0 and vice-versa.
Note that it is not possible to get error probability equal to zero if, having fixed πΈπ/π0, the spectral efficiency is higher than the value shown in the curve of Fig. 1.5;
similarly it is not possible to get error probability equal to zero, if, having fixed the spectral efficiency π π/π΅, the signal to noise ratio is lower than the value shown in the curve of Fig.1.5. In principle, any transmission system specified by a couple of values (πΈπ/π0), (π π/π΅) below the curve in Fig.1.5 can work with error probability equal to zero.
Fig.1.5: Plot Shannon channel capacity curve of spectral efficiency π π/π΅ versus πΈπ/π0 for the AWGN channel (channel capacity limit)
In summary, the Shannon channel capacity curve, meaning the theoretical tightest upper bound on the information rate of data that can be communicated at an arbitrarily low error rate using an average received signal power through an analog communication channel subject to additive white Gaussian noise of power:
πΆβ²= π΅ log2(1 + ππ π0π΅) where
ο πΆβ² is the channel capacity in information bits per second, a theoretical upper bound on the net bit rate (information rate) excluding error-correction codes;
ο π΅ is the bandwidth of the channel in hertz (passband bandwidth in case of a bandpass signal);
ο ππ is the average received signal power over the bandwidth (in case of a carrier-modulated passband transmission), measured in watts (or volts squared);
ο π0 is the average power of the noise and interference over the bandwidth, measured in watts (or volts squared);
ο ππ/(π0π΅) is the signal-to-noise ratio (SNR) of the communication signal to the noise and interference at the receiver (expressed as a linear power ratio, not as logarithmic decibels).