In telecommunication and information theory, the **code rate** (or **information rate**^{[1]}) of a forward error correction code is the proportion of the data-stream that is useful (non-redundant). That is, if the code rate is for every bits of useful information, the coder generates a total of bits of data, of which are redundant.

If is the gross bitrate or data signalling rate (inclusive of redundant error coding), the net bitrate (the useful bit rate exclusive of error-correction codes) is .

For example: The code rate of a convolutional code will typically be , , , , , etc., corresponding to one redundant bit inserted after every single, second, third, etc., bit. The code rate of the octet oriented Reed Solomon block code denoted RS(204,188) is 188/204, meaning that redundant octets (or bytes) are added to each block of 188 octets of useful information.

A few error correction codes do not have a fixed code rate—rateless erasure codes.

Note that bit/s is a more widespread unit of measurement for the information rate, implying that it is synonymous with *net bit rate* or *useful bit rate* exclusive of error-correction codes.

## See also

- Information rate
- Source information rate (Entropy rate)
- Puncturing

## References

**^**Huffman, W. Cary, and Pless, Vera,*Fundamentals of Error-Correcting Codes*, Cambridge, 2003.

This computer science article is a stub. You can help Wikipedia by expanding it. |