The capacity of a noisy channel is defined to be the maximum value of the mutual information of the input and output variables with respect to the probability distribution over the input alphabets.

Claude Shannon's noisy channel coding theorem states that an error correcting code exists such that it is possible to communicate over a noisy channel at a rate equal to or lower than the capacity with an arbitarily small probability of bit error.