Channel Capacity in Information Theory
Channel capacity represents the maximum amount of information that can be reliably transmitted over a communication channel. It is determined by the maximum mutual information between the input and output, optimized over all possible input probability distributions.
A. Channel Capacity per Symbol
For a discrete memoryless channel (DMC), the capacity per symbol is:
\[ C_s = \max_{P(x_i)} I(X; Y) \quad \text{(bits per symbol)} \]
- The maximization is over all input probability distributions \(P(x_i)\)
- This depends solely on the channel’s transition probabilities
B. Channel Capacity per Second
If \(r\) symbols are transmitted per second, the capacity becomes:
\[ C = r \cdot C_s \quad \text{(bits per second)} \]
This represents the highest achievable data rate for reliable communication over the channel.
Capacities of Special Channels
1. Lossless Channel
- All source information reaches the output intact:
\[ H(X|Y) = 0, \quad I(X; Y) = H(X) \] - Maximum entropy occurs when input is uniformly distributed: \[ C_s = \max_{P(x_i)} H(X) = \log_2 m \] where \(m\) is the number of input symbols
Example:
If \(m = 4\), then
\[
C_s = \log_2 4 = 2 \text{ bits/symbol}
\]
2. Deterministic Channel
- The output is completely determined by the input:
\[ H(Y|X) = 0, \quad I(X; Y) = H(Y) \] - Capacity per symbol: \[ C_s = \max_{P(x_i)} H(Y) = \log_2 n \] where \(n\) is the number of output symbols
3. Noiseless Channel
- Both lossless and deterministic: \[ I(X; Y) = H(X) = H(Y) \]
- Capacity per symbol: \[ C_s = \log_2 m = \log_2 n \]
4. Binary Symmetric Channel (BSC)
For a binary channel with crossover probability \(p\), the capacity is:
\[ C_s = 1 + p \log_2 p + (1 - p) \log_2 (1 - p) \]
This formula accounts for the probability of errors introduced by the channel.