The theorem shows how entropy is a lower bound to how efficiently a source can be encoded. For a memoryless source, producing source words with random variable X, then H(X) ≤ l(C), where H(X) is the entropy of X, C is any binary, uniquely decipherable encoding of X, and l(C) is the mean length of the codewords. See also Huffman coding.