Structured low-density parity-check (LDPC) code
RE049225 · 2022-09-27
Assignee
Inventors
- Michael Livshitz (Rockville, MD, US)
- Aleksandar Purkovic (Potomac, MD)
- Nina K. Burns (Clarksville, MD, US)
- Sergey Sukhobok (Rockville, MD, US)
- Muhammad Chaudhry (Gaithersburg, MD, US)
Cpc classification
H03M13/1111
ELECTRICITY
H03M13/114
ELECTRICITY
H03M13/1137
ELECTRICITY
H03M13/6368
ELECTRICITY
H03M13/116
ELECTRICITY
H03M13/1188
ELECTRICITY
H03M13/118
ELECTRICITY
H03M13/1185
ELECTRICITY
H03M13/6393
ELECTRICITY
International classification
H03M13/00
ELECTRICITY
Abstract
.[.A method for constructing a low-density parity-check (LDPC) code using a structured base parity check matrix with permutation matrix, pseudo-permutation matrix, or zero matrix as constituent sub-matrices; and expanding the structured base parity check matrix into an expanded parity check matrix. A method for constructing a LDPC code using a structured base parity check matrix H=[H.sub.d|H.sub.p], H.sub.d is the data portion, and H.sub.p is the parity portion of the parity check matrix; the parity portion of the structured base parity check matrix is such so that when expanded, an inverse of the parity portion of the expanded parity check matrix is sparse; and expanding the structured base parity check matrix into an expanded parity check matrix. A method for encoding variable sized data by using the expanded LDPC code; and applying shortening, puncturing..]. .Iadd.System and method for operating a wireless device to encode data using low-density parity-check (LDPC) encoding is discussed. One example method includes: computing a number of modulated orthogonal frequency-division multiplexing (OFDM) symbols for transmitting the data; computing a number of shortening bits; distributing the number of shortening bits over the at least one LDPC codeword; computing a number of puncturing bits for the at least one LDPC codeword; distributing the number of puncturing bits over the at least one LDPC codeword; determining a criterion using at least one of the number of shortening bits and the number of puncturing bits; if the criterion is met, increasing the number of modulated OFDM symbols and recalculating the number of puncturing bits; generating the encoded data using the number of shortening bits, the number of puncturing bits, and the at least one LDPC codeword; and transmitting the encoded data..Iaddend.
Claims
.[.1. A method of low-density parity-check (LDPC) encoding data, comprising: receiving input data from a data source; and applying the following expanded parity check matrix to the input data to generate encoded data: TABLE-US-00009 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 32 52 55 80 95 22 6 51 −1 63 31 88 20 −1 −1 −1 6 40 66 16 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.2. A method of decoding low-density parity-check (LDPC) encoded data, comprising: receiving encoded data from a data source; and applying the following expanded parity check matrix to the encoded data to generate decoded data: TABLE-US-00010 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.3. Apparatus for low-density parity-check (LDPC) encoding data, comprising: an input port operable to receive input data from a data source; and circuitry coupled to the input port and operable to apply the following expanded parity check matrix to the input data to generate encoded data: TABLE-US-00011 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.4. Apparatus for low-density parity-check (LDPC) encoding data, comprising: an input port operable to receive input data from a data source; and a matrix application element operable to apply the following expanded parity check matrix to the input data to generate encoded data: TABLE-US-00012 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.5. Apparatus for low-density parity-check (LDPC) encoding data, comprising: an input port operable to receive input data from a data source; and means for applying the following expanded parity check matrix to the input data to generate encoded data: TABLE-US-00013 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.6. Apparatus for decoding low-density parity-check (LDPC) encoded data, comprising: an input port operable to receive encoded data from a data source; and circuitry coupled to the input port and operable to apply the following expanded parity check matrix to the encoded data to generate decoded data: TABLE-US-00014 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.7. Apparatus for decoding low-density parity-check (LDPC) encoded data, comprising: an input port operable to receive encoded data from a data source; and a matrix application element operable to apply the following expanded parity check matrix to the encoded data to generate decoded data: TABLE-US-00015 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.8. Apparatus for decoding low-density parity-check (LDPC) encoded data, comprising: an input port operable to receive encoded data from a data source; and means for applying the following expanded parity check matrix to the encoded data to generate decoded data: TABLE-US-00016 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96)..].
.[.9. A telecommunications network, comprising: an LDPC encoder operable to apply the following expanded parity check matrix to input data to generate encoded data: TABLE-US-00017 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96); a transmitter operable to transmit the encoded data over a transmission medium; a receiver operable to receive the transmitted encoded data; and an LDPC decoder operable to apply said expanded parity check matrix to the encoded data to recover the input data..].
.[.10. A method of operating a telecommunications network, comprising: applying the following expanded parity check matrix to input data to generate encoded data: TABLE-US-00018 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96); transmitting the encoded data over a transmission medium; receiving the transmitted encoded data; and applying said expanded parity check matrix to the encoded data to recover the input data..].
.[.11. A transceiver, comprising: an LDPC encoder operable to apply the following expanded parity check matrix to input data to generate encoded data: TABLE-US-00019 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96); a transmitter operable to transmit the encoded data over a transmission medium; a receiver operable to receive encoded data from the transmission medium; and an LDPC decoder operable to apply said expanded parity check matrix to the received encoded data to generate decoded data..].
.[.12. A method of operating a transceiver, comprising: applying the following expanded parity check matrix to input data to generate encoded data: TABLE-US-00020 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 95 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 66 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0 wherein an expansion factor, L, is between 24 and 96, −1 represents an L×L all-zero square matrix, and any other integer, Sij, represents an L×L identity matrix circularly right shifted by a shift amount equal to floor ((L×Sij)/96); transmitting the encoded data over a transmission medium; receiving encoded data from the transmission medium; and applying said expanded parity check matrix to the received encoded data to generate decoded data..].
.Iadd.13. A method to operate a wireless device to encode data using low-density parity-check (LDPC) encoding, the method comprising: computing a number of modulated orthogonal frequency-division multiplexing (OFDM) symbols for transmitting the data; computing a number of shortening bits N.sub.shortened for at least one LDPC codeword to be used during an encoding; distributing the number of shortening bits N.sub.shortened over the at least one LDPC codeword; computing a number of puncturing bits N.sub.punctured for the at least one LDPC codeword; distributing the number of puncturing bits over the at least one LDPC codeword; determining a criterion using at least one of the number of shortening bits N.sub.shortened and the number of puncturing bits N.sub.punctured; if the criterion is met, increasing the number of modulated OFDM symbols and recalculating the number of puncturing bits N.sub.punctured; generating the encoded data using the number of shortening bits N.sub.shortened, the number of puncturing bits N.sub.punctured, and the at least one LDPC codeword, wherein shortening is performed by setting N.sub.shortened information bits to 0 within the at least one LDPC codeword, and puncturing is performed by discarding N.sub.punctured parity bits from the at least one LDPC codeword; and transmitting the encoded data..Iaddend.
.Iadd.14. The method of claim 13, wherein the criterion is determined using both the number of shortening bits N.sub.shortened and the number of puncturing bits N.sub.punctured..Iaddend.
.Iadd.15. The method of claim 13, further comprising computing a number of the at least one LDPC codeword to encode the data and a length of each of the at least one LDPC codeword..Iaddend.
.Iadd.16. The method of claim 13, wherein transmitting the encoded data comprises transmitting the encoded data using the number of modulated OFDM symbols..Iaddend.
.Iadd.17. A wireless device comprising: a low-density parity-check (LDPC) encoder circuit configured to: compute a number of modulated orthogonal frequency-division multiplexing (OFDM) symbols for transmitting data; compute a number of shortening bits N.sub.shortened for at least one LDPC codeword to be used during an encoding; distribute the number of shortening bits N.sub.shortened over the at least one LDPC codeword; compute a number of puncturing bits N.sub.punctured for the at least one LDPC codeword; distribute the number of puncturing bits over the at least one LDPC codeword; determine a criterion using at least one of the number of shortening bits N.sub.shortened and the number of puncturing bits N.sub.punctured; if the criterion is met, increase the number of modulated OFDM symbols and recalculate the number of puncturing bits N.sub.punctured; and generate the encoded data using the number of shortening bits N.sub.shortened, the number of puncturing bits N.sub.punctured, and the at least one LDPC codeword, wherein shortening is performed by setting N.sub.shortened information bits to 0 within the at least one LDPC codeword, and puncturing is performed by discarding N.sub.punctured parity bits from the at least one LDPC codeword; and a transmitter configured to transmit the encoded data..Iaddend.
.Iadd.18. The wireless device of claim 17, wherein the criterion is determined using both the number of shortening bits N.sub.shortened and the number of puncturing bits N.sub.punctured..Iaddend.
.Iadd.19. The wireless device of claim 17, wherein the LDPC encoder circuit is further configured to compute a number of the at least one LDPC codeword to encode the data and a length of each of the at least one LDPC codeword..Iaddend.
.Iadd.20. The wireless device of claim 17, wherein the transmitter is configured to transmit the encoded data using the number of modulated OFDM symbols..Iaddend.
.Iadd.21. A non-transitory computer readable medium comprising instructions that when executed by a wireless device cause the wireless device to perform steps of a method to encode data using low-density parity-check (LDPC) encoding, the method comprising: computing a number of modulated orthogonal frequency-division multiplexing (OFDM) symbols for transmitting the data; computing a number of shortening bits N.sub.shortened for at least one LDPC codeword to be used during an encoding; distributing the number of shortening bits N.sub.shortened over the at least one LDPC codeword; computing a number of puncturing bits N.sub.punctured for the at least one LDPC codeword; distributing the number of puncturing bits over the at least one LDPC codeword; determining a criterion using at least one of the number of shortening bits N.sub.shortened and the number of puncturing bits N.sub.punctured; if the criterion is met, increasing the number of modulated OFDM symbols and recalculating the number of puncturing bits N.sub.punctured; generating the encoded data using the number of shortening bits N.sub.shortened, the number of puncturing bits N.sub.punctured, and the at least one LDPC codeword, wherein shortening is performed by setting N.sub.shortened information bits to 0 within the at least one LDPC codeword, and puncturing is performed by discarding N.sub.punctured parity bits from the at least one LDPC codeword; and transmitting the encoded data..Iaddend.
.Iadd.22. The non-transitory computer readable medium of claim 21, wherein the criterion is determined using both the number of shortening bits N.sub.shortened and the number of puncturing bits N.sub.punctured..Iaddend.
.Iadd.23. The non-transitory computer readable medium of claim 21, wherein the method further comprises computing a number of the at least one LDPC codeword to encode the data and a length of each of the at least one LDPC codeword..Iaddend.
.Iadd.24. The non-transitory computer readable medium of claim 21, wherein transmitting the encoded data comprises transmitting the encoded data using the number of modulated OFDM symbols..Iaddend.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The invention and the illustrated embodiments may be better understood, and the numerous objects, advantages, and features of the present invention and illustrated embodiments will become apparent to those skilled in the art by reference to the accompanying drawings, and wherein:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
(28) Reference will now be made in detail to some specific embodiments of the invention including the best modes contemplated by the inventors for carrying out the invention. Examples of these specific embodiments are illustrated in the accompanying drawings. While the invention is described in conjunction with these specific embodiments, it will be understood that it is not intended to limit the invention to the described embodiments. On the contrary, it is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the invention as defined by the appended claims. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. The present invention may be practiced without some or all of these specific details. In other instances, well-known process operations have not been described in detail in order not to unnecessarily obscure the present invention.
(29) Efficient decoder architectures are enabled by designing the parity check matrix, which in turn defines the LDPC code, around some structural assumptions: structured LDPC codes.
(30) One example of this design is that the parity check matrix comprises sub-matrices in the form of binary permutation or pseudo-permutation matrices.
(31) The term “permutation matrix” is intended to mean square matrices with the property that each row and each column has one element equal to 1 and other elements equal to 0. Identity matrix, a square matrix with ones on the main diagonal and zeros elsewhere, is a specific example of permutation matrix. The term “pseudo-permutation matrix” is intended to include matrices that are not necessarily square matrices, and matrices may have row(s) and/or column(s) consisting of all zeros. It has been shown, that using this design, significant savings in wiring, memory, and power consumption are possible while still preserving the main portion of the coding gain. This design enables various serial, parallel, and semi-parallel hardware architectures and therefore various trade-off mechanisms.
(32) This structured code also allows the application of layered decoding, also referred to as layered belief propagation decoding, which exhibits improved convergence properties compared to a conventional sum-product algorithm (SPA) and its derivations. Each iteration of the layered decoding consists of a number of sub-iterations that equals the number of blocks of rows (or layers).
(33) LDPC code parity check matrix design also results in the reduction in encoder complexity. Classical encoding of LDPC codes is more complex than encoding of other advanced codes used in FEC, such as turbo codes. In order to ease this complexity it has become common to design systematic LDPC codes with the parity portion of the parity check matrix containing a lower triangular matrix. This allows simple recursive decoding. One simple example of a lower triangular matrix is a dual diagonal matrix as shown in
(34) Referring to
(35)
where d=[d.sub.0 . . . d.sub.k−1].sup.T is the block of data bits and p=[p.sub.0 . . . p.sub.M−1].sup.T are the parity bits. A codeword is any binary, or in general, non-binary, N-vector c that satisfies:
Hc=H.sub.dd+H.sub.pp=0
(36) Thus, a given data block d is encoded by solving binary equation H.sub.dd=H.sub.pp for the parity bits p. In principle, this involves inverting the M×M matrix H.sub.p to resolve p:
p=H.sub.p.sup.−1H.sub.dd [equation 1]
(37) H.sub.p is assumed to be invertible. If the inverse of H.sub.p, H.sub.p.sup.−1 is also low density then the direct encoding specified by the above formula can be done efficiently. However, with the dual diagonal structure of H.sub.p 32 encoding can be performed as a simple recursive algorithm:
(38)
where i.sub.n.sup.0 is the index of the column in which row 0 contains a “1”
(39)
where i.sub.n.sup.1 is the index of the column in which row 1 contains a “1”
(40)
where i.sub.n.sup.M−1 is the index of the column in which row M−1 contains a “1”.
(41) In these recursive expressions h.sub.r,c are non-zero elements (1 in this exemplary matrix) of the data portion of the parity check matrix, H.sub.d 31. The number of non-zero elements in rows 0, 1, . . . , M−1, is represented by k.sub.0, k.sub.1, . . . , k.sub.M−1, respectively.
(42) One desirable feature of LDPC codes is that they support various required code rates and block sizes. A common approach is to have a small base parity check matrix defined for each required code rate and to support various block sizes by expanding the base parity check matrix. Since it is usually required to support a range of block sizes, a common approach is to define expansion for the largest block size and then apply other algorithms which specify expansion for smaller block sizes. Below is an example of a base parity check matrix:
(43) TABLE-US-00003 11 0 10 6 3 5 −1 0 −1 −1 −1 −1 10 9 2 2 3 0 −1 0 0 −1 −1 −1 7 9 11 10 4 7 −1 −1 0 0 −1 −1 9 2 4 6 5 3 0 −1 −1 0 0 −1 3 11 2 4 2 11 −1 −1 −1 −1 0 0 2 7 1 0 10 7 1 −1 −1 −1 −1 0
(44) In this example the base parity check matrix is designed for the code rate R=½ and its dimensions are (M.sub.b×N.sub.b)=(6×12). Assume that the codeword sizes (lengths) to be supported are in the range N=[72, 144], with increments of 12, i.e. N=[72, 84, . . . , 132, 144]. In order to accommodate those block lengths the parity check matrix needs to be of the appropriate size (i.e. the number of columns match N, the block length). The number of rows is defined by the code rate: M=(1−R) N. The expansion is defined by the base parity check matrix elements and the expansion factor L, which results in the maximum block size. The conventions used in this example, for interpreting the numbers in the base parity check matrix, are as follows: −1, represents L×L all-zero square matrix, 0.sub.L, L equals 12 in this example; 0, represents L×L identity matrix, I.sub.L. integer, r (<L), represents L×L identity matrix, I.sub.L, rotated to the right (for example) a number of times corresponding to the integer.
(45) The following example shows a rotated identity matrix where the integer specifying rotation is 5:
(46) TABLE-US-00004 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0
(47) Therefore, for the largest block (codeword) size of N=144, the base parity check matrix needs to be expanded by an expansion factor of 12. That way the final expanded parity check matrix to be used for encoding and generating the codeword of size 144, is of the size (72×144). In other words, the base parity check matrix was expanded L.sub.max=12 times (from 6×12 to 72×144). For the block sizes smaller than the maximum, the base parity check matrix is expanded by a factor L<L.sub.max. In this case expansion is performed in the similar fashion except that now matrices I.sub.L and 0.sub.L, are used instead of I.sub.Lmax and 0.sub.Lmax, respectively. Integers specifying the amount of rotation of the appropriate identity matrix, I.sub.L, are derived from those corresponding to the maximum expansion by applying some algorithm. For example, such an algorithm may be a simple modulo operation:
r.sub.L=(r.sub.Lmax)modulo L
(48) An example of such a matrix is shown in
(49) The expansion may be done for example by replacing each non-zero element with a permutation matrix of the size of the expansion factor. One example of performing expansion is as follows.
(50) H.sub.p is expanded by replacing each “0” element by an L×L zero matrix, 0.sub.L×L, and each “1” element by an L×L identity matrix, I.sub.L×L, where L represent the expansion factor.
(51) H.sub.d is expanded by replacing each “0” element by an L×L zero matrix, 0.sub.L×L, and each “1” element by a circularly shifted version of an L×L identity matrix, I.sub.L×L. The shift order, s (number of circular shifts, for example, to the right) is determined for each non-zero element of the base parity check matrix.
(52) It should be apparent to a person skilled in the art that these expansions can be implemented without the need to significantly change the base hardware wiring.
(53)
(54) The simple recursive algorithm described earlier can still be applied in a slightly modified form to the expanded parity check matrix. If h.sub.i,j represent elements of the H.sub.d portion of the expanded parity check matrix, then parity bits can be determined as follows:
p.sub.0=h.sub.0,0d.sub.0+h.sub.0,1d.sub.1+h.sub.0,2d.sub.2+ . . . +h.sub.0,11d.sub.11
p.sub.1=h.sub.1,0d.sub.0+h.sub.1,1d.sub.1+h.sub.1,2d.sub.2+ . . . +h.sub.1,11d.sub.11
p.sub.2=h.sub.2,0d.sub.0+h.sub.2,1d.sub.1+h.sub.2,2d.sub.2+ . . . +h.sub.2,11d.sub.11
p.sub.3=p.sub.0+h.sub.3,0d.sub.0+h.sub.3,1d.sub.1+h.sub.3,2d.sub.2+ . . . +h.sub.3,11d.sub.11
p.sub.4=p.sub.1+h.sub.4,0d.sub.0+h.sub.4,1d.sub.1+h.sub.4,2d.sub.2+ . . . +h.sub.4,11d.sub.11
p.sub.5=p.sub.2+h.sub.5,0d.sub.0+h.sub.5,1d.sub.1+h.sub.5,2d.sub.2+ . . . +h.sub.5,11d.sub.11
p.sub.6=p.sub.3+h.sub.6,0d.sub.0+h.sub.6,1d.sub.1+h.sub.6,2d.sub.2+ . . . +h.sub.6,11d.sub.11
p.sub.7=p.sub.4+h.sub.7,0d.sub.0+h.sub.7,1d.sub.1+h.sub.7,2d.sub.2+ . . . +h.sub.7,11d.sub.11
p.sub.8=p.sub.5+h.sub.8,0d.sub.0+h.sub.8,1d.sub.1+h.sub.8,2d.sub.2+ . . . +h.sub.8,11d.sub.11
p.sub.9=p.sub.6+h.sub.9,0d.sub.0+h.sub.9,1d.sub.1+h.sub.9,2d.sub.2+ . . . +h.sub.9,11d.sub.11
p.sub.10=p.sub.7+h.sub.10,0d.sub.0+h.sub.10,1d.sub.1+h.sub.10,2d.sub.2+ . . . +h.sub.10,11d.sub.11
p.sub.11=p.sub.8+h.sub.11,0d.sub.0+h.sub.11,1d.sub.1+h.sub.11,2d.sub.2+ . . . +h.sub.11,11d.sub.11
(55) However, when the expansion factor becomes large, then the number of columns with only one non-zero element, i.e. 1 in the example here, in the H.sub.p becomes large as well. This may have a negative effect on the performance of the code.
(56) One remedy for this situation is to use a slightly modified dual diagonal H.sub.p matrix. This is illustrated with reference to
(57) The parity check equations now become:
h.sub.0,0d.sub.0h.sub.0,1d.sub.1+ . . . +h.sub.0,11d.sub.11+p.sub.0+p.sub.3=0 [Equation 2]
h.sub.1,0d.sub.0h.sub.1,1d.sub.1+ . . . +h.sub.1,11d.sub.11+p.sub.1+p.sub.4=0 [Equation 3]
h.sub.2,0d.sub.0h.sub.2,1d.sub.1+ . . . +h.sub.2,11d.sub.11+p.sub.2+p.sub.5=0 [Equation 4]
h.sub.3,0d.sub.0h.sub.3,1d.sub.1+ . . . +h.sub.3,1d.sub.11+p.sub.0+p.sub.3+p.sub.6=0 [Equation 5]
h.sub.4,0d.sub.0h.sub.4,1d.sub.1+ . . . +h.sub.4,11d.sub.11+p.sub.1+p.sub.4+p.sub.7=0 [Equation 6]
h.sub.5,0d.sub.0h.sub.5,1d.sub.1+ . . . +h.sub.5,11d.sub.11+p.sub.2+p.sub.5+p.sub.8=0 [Equation 7]
h.sub.6,0d.sub.0h.sub.6,1d.sub.1+ . . . +h.sub.6,11d.sub.11+p.sub.6+p.sub.9=0 [Equation 8]
h.sub.7,0d.sub.0h.sub.7,1d.sub.1+ . . . +h.sub.7,11d.sub.11+p.sub.7+p.sub.10=0 [Equation 9]
h.sub.8,0d.sub.0h.sub.8,1d.sub.1+ . . . +h.sub.8,11d.sub.1+p.sub.8+p.sub.11=0 [Equation 10]
h.sub.9,0d.sub.0h.sub.9,1d.sub.1+ . . . +h.sub.9,11d.sub.11+p.sub.0+p.sub.9=0 [Equation 11]
h.sub.10,0d.sub.0h.sub.10,1d.sub.1+ . . . +h.sub.10,11d.sub.11+p.sub.1+p.sub.10=0 [Equation 12]
h.sub.11,0d.sub.0h.sub.11,1d.sub.1+ . . . +h.sub.11,11d.sub.11+p.sub.2+p.sub.11=0 [Equation 13]
(58) Now by summing up equations 2, 5, 8, and 11, the following expression is obtained:
(h.sub.0,0+h.sub.3,0+h.sub.6,0+h.sub.9,0)d.sub.0+(h.sub.0,1+h.sub.3,1+h.sub.6,1+h.sub.9,0)d.sub.1 . . . +(h.sub.0,11+h.sub.3,11+h.sub.6,11+h.sub.9,11)d.sub.11p.sub.0+p.sub.3+p.sub.0+p.sub.3+p.sub.6+p.sub.6+p.sub.9+p.sub.0+p.sub.9=0
(59) Since only p.sub.0 appears an odd number of times in the equation above, all other parity check bits cancel except for p.sub.0, and thus:
p.sub.0=(h.sub.0,0+h.sub.3,0+h.sub.6,0+h.sub.9,0)d.sub.0+(h.sub.0,1+h.sub.3,1+h.sub.6,1+h.sub.9,1)d.sub.1+. . . +(h.sub.0,11+h.sub.3,11+h.sub.6,11+h.sub.9,11)d.sub.11
(60) Likewise:
p.sub.1=(h.sub.1,0+h.sub.4,0+h.sub.7,0+h.sub.10,0)d.sub.0+(h.sub.1,1+h.sub.4,1+h.sub.7,1+h.sub.10,1)d.sub.0+(h.sub.1,11+h.sub.4,11+h.sub.7,11+h.sub.10,11)d.sub.11
p.sub.2=(h.sub.2,0+h.sub.5,0+h.sub.8,0+h.sub.11,0)d.sub.0+(h.sub.2,1+h.sub.5,1+h.sub.87,1+h.sub.11,1)d.sub.1+. . . +(h.sub.2,11+h.sub.5,11+h.sub.8,11+h.sub.11,11)d.sub.11
(61) After determining p.sub.0, p.sub.1, p.sub.2 the other parity check bits are obtained recursively:
p.sub.3=h.sub.0,0d.sub.0+h.sub.0,1d.sub.1+. . . +h.sub.0,11d.sub.11+p.sub.0
p.sub.4=h.sub.1,0d.sub.0+h.sub.1,1d.sub.1+. . . +h.sub.1,11d.sub.11+p.sub.1
p.sub.5=h.sub.2,0d.sub.0+h.sub.2,1d.sub.1+. . . +h.sub.2,11d.sub.11+p.sub.2
p.sub.6=h.sub.3,0d.sub.0+h.sub.3,1d.sub.1+. . . +h.sub.3,11d.sub.11+p.sub.0+p.sub.3
p.sub.7=h.sub.4,0d.sub.0+h.sub.4,1d.sub.1+. . . +h.sub.4,11d.sub.11+p.sub.1+p.sub.4
p.sub.8=h.sub.5,0d.sub.0+h.sub.5,1d.sub.1+. . . +h.sub.5,11d.sub.11+p.sub.2+p.sub.5
p.sub.9=h.sub.6,0d.sub.0+h.sub.6,1d.sub.1+. . . +h.sub.6,11d.sub.11+p.sub.6
p.sub.10=h.sub.7,0d.sub.0+h.sub.7,1d.sub.1+. . . +h.sub.7,11d.sub.11+p.sub.7
p.sub.11=h.sub.8,0d.sub.0+h.sub.8,1d.sub.1+. . . +h.sub.8,11d.sub.11+p.sub.8 [Equation 14]
(62) The present invention provides method and system enabling high throughput, low latency implementation of LDPC codes, and preserving the simple encoding feature at the same time.
(63) In accordance with one embodiment of the present invention, a general form is shown in
(64) The data portion (H.sub.d) may also be placed on the right side of the parity (H.sub.p) portion of the parity check matrix. In the most general case, columns from H.sub.d and H.sub.p may be interchanged.
(65) Parity check matrices constructed according to the embodiments of the present invention supports both regular and irregular types of the parity check matrix. Not only the whole matrix may be irregular (non-constant weight of its rows and columns) but also that its constituents H.sub.d and H.sub.p may be irregular, if such a partition is desired.
(66) If the base parity check matrix is designed with some additional constraints, then base parity check matrices for different code rates may also be derived from one original base parity check matrix in one of two ways: a. Row combining: higher rate base parity check matrices are derived from an original lower rate base parity check matrix by combining rows of the base parity check matrix. Multiple strategies can be applied in order to make the resultant higher rate base matrix maintain the properties of the original matrix, i.e. the weight of each of the column in a block of rows is at most one. One way of doing row combining will be to combine (add together) rows that belong to the same block of rows. This guarantees the preservation of column weight properties, with decrease of the block row size. Another way of row combining will be to combine the rows that belong to different blocks of rows, where they don't have overlapping elements. b. Row splitting: lower rate base parity check matrices are derived from an original higher rate base parity check matrix by splitting rows of the base parity check matrix. The resultant lower rate parity check matrix shall maintain the properties of the original matrix, i.e. the weight of each of the column in a block of rows is at most one.
(67) Row-combining or row-splitting, with the specific constraints defined above, allow efficient coding of a new set of expanded derived base parity check matrices. In these cases the number of layers may be as low as the minimum number of block rows (layers) in the original base parity check matrix.
(68)
H.sub.p,present_invention(m)=T (H.sub.p,existing,m),
(69) Where T is the transform describing the base parity check matrix expansion process and m is the size of the permutation matrices. For m=1, H.sub.p of the present invention defines the form of the prior art H.sub.p (dual diagonal with the odd-weight column), i.e.
(70) H.sub.p,present_invention(l)=T (H.sub.p,existing,l)=H.sub.p,existing
(71) A further pair of parity portions with sub-matrices 905, 906 illustrate cases where these first and last columns, respectively, have only one sub-matrix each.
(72) The two parity portions with sub-matrices 907, 908 in
(73) One of the characteristics of the base parity check matrix expansion of the present invention is that the expanded base parity check matrix inherits structural features from the base parity check matrix. In other words, the number of blocks (rows or columns) that can be processed in parallel (or serial, or in combination) in the expanded parity check matrix equals the number of blocks in the base parity check matrix.
(74) Referring to
(75) The base parity check matrix 100 of
(76) It can be seen that expanded parity check matrix 110 has inherited structural properties of its base parity check matrix 100 from
(77) The sub-matrices of the present invention are not limited to permutation sub-matrices, pseudo-permutation sub-matrices or zero sub-matrices. In other words, the embodiments of the present invention are not restricted to the degree distribution (distribution of column weights) of the parity check matrix, allowing the matrix to be expanded to accommodate various information packet sizes and can be designed for various code rates. This generalization is illustrated through following examples.
(78)
(79)
(80) In the context of parallel row processing, layered belief propagation decoding is next briefly described with reference to
(81) A high level architectural block diagram is shown in
(82) In order to support a more general approach in accordance with an embodiment of the present invention, the architecture of
(83) By exercising careful design of the parity check matrix, the additional inter-layer storage 155 in
(84) Iterative parallel decoding process is best described as read-modify-write operation. The read operation is performed by a set of permuters, which deliver information from memory modules to corresponding processing units. Parity check matrices, designed with the structured regularity described earlier, allow efficient hardware implementations (e.g., fixed routing, use of simple barrel shifters) for both read and write networks. Memory modules are organized so as to provide extrinsic information efficiently to processing units.
(85) Processing units implement block (layered) decoding (updating iterative information for a block of rows) by using any known iterative algorithms (e.g. Sum Product, Min-Sum, Bahl-Cocke-Jelinek-Raviv (BCJR)).
(86) Inverse permuters are part of the write network that performs the write operation back to memory modules.
(87) Such parallel decoding is directly applicable when the parity check matrix is constructed based on permutation, pseudo-permutation or zero sub-matrices.
(88) To encode using sub-matrices other than permutation, pseudo-permutation or zero sub-matrices, one embodiment of the present invention uses special sub-matrices. A sub-matrix can also be constructed by concatenation of smaller permutation or pseudo-permutation matrices. An example of this concatenation is illustrated in
(89) Parallel decoding is applicable with the previously described modification to the methodology; that is, when the parity check matrix includes sub-matrices built by concatenation of smaller permutation matrices.
(90)
(91) It can be seen that for the decoding layer 171 a first processing unit receives information in the first row 179 from bit 1 (according to S.sub.21), bit 6 (S.sub.22), bit 9 (S.sub.23), bit 13 (S.sup.1.sub.24), bit 15 (S.sup.2.sub.24), bit 21 (S.sub.28), and bit 24 (S.sub.29). Other processing units are loaded in a similar way.
(92) For layered belief propagation type decoding algorithms, the processing unit inputs extrinsic information accumulated, by all other layers, excluding the layer currently being processed. Thus, the prior art implementation described using
(93) This is illustrated in
(94) For simplicity,
(95) Improvement in throughput, and reduction in latency in accordance to an embodiment of the present invention is further illustrated by the following example.
(96) The LDPC codes can be decoded using several methods. In general, iterative decoding is applied. The most common is the sum-product algorithm (SPA) method. Each iteration in SPA comprises two steps: a. horizontal step, during which all row variables are updated at the same time based on the column variables; and b. vertical step, during which all column variables are updated at the same time based on row variables.
(97) It has been shown that better performance, in terms of the speed of convergence, can be achieved with layered decoding. In layered decoding only row variables are updated for a block of rows, one block row at a time. The fastest approach is to process all the rows within a block of rows simultaneously.
(98) The following is a comparison of the achievable throughput (bit rate) of two LDPC codes: one based on the existing method for expanding matrix, as described in
T=(K×F)/(C×I),
where K is number of info bits, F is clock frequency, C is number of cycles per iteration, and I is the number of iterations. Assuming that K, F, and I are fixed and, for example, equal: K=320 bits, F=100 MHz, and I=10, the only difference between the existing method and the present invention is derived from C, the factor which is basically a measure of the level of allowed parallelism. It can be seen, by comparing
C.sub.existing=16 and C.sub.present_invention=4.
(99) Using these numbers in the formula gives:
T.sub.max,existing=200 Mbps
T.sub.max,present_invention=800 Mbps
(100) As expected, the maximum throughput is 4 times greater. All the desirable features of the code design in terms of efficient encoding are preserved. For example, without degradation in performance, the encoding algorithm as described earlier with respect to
(101) Furthermore, when a scaleable solution is desired, the size of the expanded LDPC parity check matrix is designed to support the maximum block size. The existing solutions do not scale well with respect to the throughput for various block sizes. For example, using the existing method for layered decoding, processing of short and long blocks takes the same amount of time. This is caused by the fact that for shorter blocks, not all processing units are used, resulting proportionally lower achieved throughput.
(102) The following example is based on the same example as before by comparing matrices as described earlier in
(103)
(104) The following table compares the computed results.
(105) TABLE-US-00005 Number of Codeword processing Throughput size C units (Mbps) Existing (FIG. 5) 320 16 20 200 1280 16 80 800 Embodiment of 320 4 80 800 present invention (FIG. 17) 1280 16 80 800
(106) It can be seen from the table that the embodiment of the present invention provides constant throughput independent on the codeword size, whereas in the case of the existing method the throughput for the smaller blocks drops considerably. The reason is that while the embodiment of the present invention fully utilizes all available processing resources irrespective of block size, the existing method utilizes all processing units only in the case of the largest block, and a fraction of the total resources for other cases.
(107) The example here illustrating the throughput improvement for shorter blocks, leads also to the conclusion that reduced latency is also achieved with the embodiment of the present invention. When large blocks of data are broken into smaller pieces, the encoded data is split among multiple codewords. If one places a shorter codeword at the end of series of longer codewords, then the total latency depends primarily on the decoding time of the last codeword. According to the table above, short blocks require proportionally less time to be decoded (as compared to the longer codewords), thereby allowing reduced latency to be achieved by encoding the data in suitably short blocks.
(108) In addition to the full hardware utilization illustrated above, embodiments of the present invention allow hardware scaling, so that short blocks can use proportionately less hardware resources if an application requires it.
(109) Furthermore, utilization of more efficient processing units and memory blocks is enabled. Memory can be organized to process a number of variables in parallel. The memory can therefore, be partitioned in parallel.
(110) The present invention provides new LPDC base parity matrices, and expanded matrices based on the new base parity matrices, and method for use thereof.
(111) The locations of non-zero matrices for rate R in an exemplary matrix are chosen, so that: a) parity part ((1−R)*24 rightmost columns) of the matrix is designed to allow simple encoding algorithms; b) weights of all columns in the data portion of base parity check matrix is uniform; c) weights of all rows in the data portion of a base parity check matrix is uniform; d) the parity part of the matrix allows simple encoding algorithms. For example, the encoding algorithm based on equation 1, or equation 14.
(112) An example of R=¾ base parity check matrix design using criteria a) to d) is:
(113) TABLE-US-00006 1 1 1 1 0 0 0 1 1 0 1 0 1 1 1 1 0 1 1 1 0 0 0 0 1 1 1 1 0 1 1 0 1 0 0 1 0 1 1 1 1 0 0 1 1 0 0 0 1 0 1 0 1 1 1 1 0 1 0 0 1 0 1 1 1 1 0 0 1 1 0 0 1 1 0 1 1 1 1 1 0 1 1 1 0 0 0 0 1 1 1 0 0 1 1 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 1 1 0 1 1 1 1 0 0 0 1 1 1 1 1 1 0 0 1 1 1 0 0 0 0 1
(114) The rate R=¾ matrix definition built based on such base parity check matrix covers expansion factors in the range L between 24 and L.sub.max=96 in increments of 4. Right circular shifts of the corresponding L×L identity matrix s′.sub.ij, are determined as follows:
(115)
where s.sub.ij is specified in the matrix definition below:
(116) TABLE-US-00007 6 38 3 93 −1 −1 −1 30 70 −1 86 −1 37 38 4 11 −1 46 48 0 −1 −1 −1 −1 62 94 19 84 −1 92 78 −1 15 −1 −1 92 −1 45 24 32 30 −1 −1 0 0 −1 −1 −1 71 −1 55 −1 12 66 45 79 −1 78 −1 −1 10 −1 22 55 70 82 −1 −1 0 0 −1 −1 38 61 −1 66 9 73 47 64 −1 39 61 43 −1 −1 −1 −1 95 32 0 −1 −1 0 0 −1 −1 −1 −1 −1 32 52 55 80 95 22 6 51 24 90 44 20 −1 −1 −1 −1 −1 −1 0 0 −1 63 31 88 20 −1 −1 −1 6 40 56 16 71 53 −1 −1 27 26 48 −1 −1 −1 −1 0
(117) The present invention further enables flexible rate adjustments by the use of shortening, or puncturing, or a combination thereof. Block length flexibility is also enabled through expansion, shortening, or puncturing, or combinations thereof.
(118) Any of these operations can be applied to the base or expanded parity check matrices.
(119) Referring to
(120) The data packet 201 of length L is divided into segments 208. These segments are in turn encoded using an LDPC code (N, K). The information block K 202 may be optionally pruned to K′ 204; and the parity check bits M may be pruned to M′ 205. The term “pruning” is intended to mean applying code shortening by sending less information bits than possible with a given code, (K′<K). The term “puncturing” is intended to mean removing some of the parity bits and/or data bits prior to sending the encoded bits to the modulator block and subsequently over the channel. Pruned codewords may be concatenated 206 in order to accommodate the encoded data packet, and the resulting stream 207 is padded with bits 209 to match the boundaries 210 of modulated symbols before being sent to the modulator. The amount of shortening and puncturing may be different for the constituent pruned codewords. The objectives here are: (a) Keep the performance in terms of coding gain as high as possible. This objective translates into the following needs: Select the largest suitable codeword from the available set of codewords. For the LDPC codes and other block codes, the longer the codeword the more coding gain can be achieved, although at certain codeword size the point of diminishing return is reached. Adjust properly the amount of shortening and puncturing, as this directly affects the performance, and the efficiency of the transmission. (b) Use as few of the modulated symbols as possible. This in turn means that it is desirable to utilize transmit power economically. This is especially important for battery operated hand-held wireless devices by keeping the air time at minimum. (c) Keep the overall complexity at a reasonable level. This usually translates into a requirement to operate with a relatively small set of codewords in different sizes. In addition, it is desirable to have a code designed in such a way that various codeword lengths can be implemented efficiently. Finally, the actual procedure defining concatenation rules should be simple.
(121) From objective (a) above it follows that in order to use a small number of codewords, an efficient shortening and puncturing operation needs to be applied. However, those operations have to be implemented in a way that would neither compromise the coding gain advantage of LDPC codes, nor lower the overall transmit efficiency unnecessarily. This is particularly important when using the special class of LDPC parity check matrices that enable simple encoding operation, for example, as the one describe in the previous embodiments of the present invention. These special matrices employ either a lower triangular, a dual-diagonal, or a modified dual-diagonal in the parity portion of the parity check matrix corresponding. An example of a dual-diagonal matrix is described earlier in
(122) Work to achieve efficient puncturing has been done using the “rate compatible” approach. One or more LDPC parity check matrix is designed for the low code rate application. By applying the appropriate puncturing of the parity portion, the same matrix can be used for a range of code rates which are higher than the original code rate as the data portion in relation to the codeword increases. These methods predominantly target applications where adaptive coding (e.g. hybrid automatic repeat request, H-ARQ) and/or unequal bit protection is desired.
(123) Puncturing may also be combined with code extension to mitigate the problems associated with “puncturing only” cases. The main problem that researchers are trying to solve here is to preserve an optimum degree distribution through the process of modifying the original parity check matrix.
(124) However, these methods do not directly address the problem described earlier: apply shortening and puncturing in such a way that the code rate is approximately the same as the original one, and the coding gain is preserved.
(125) One method attempting to solve this problem specifies shortening and puncturing such that the code rate of the original code is preserved. The following notation is used:
(126) N.sub.punctured—Number of punctured bits,
(127) N.sub.shortened—Number of shortened bits.
(128) Shortening to puncturing ratio, q, is defined as: q=N.sub.shortened/N.sub.punctured.
(129) In order to preserve the same code rate, q has to satisfy the following equation:
q.sub.rate_preserved=R/(1−R)
(130) Two approaches are prescribed for choosing which bits to shorten and which to puncture to reach a shortening and a puncturing pattern.
(131) Two approaches for shortening and puncturing of the expanded matrices are described in Dale Hocevar and Anuj Batra, “Shortening and Puncturing Scheme to Simplify LDPC Decoder Implementation,” Jan. 11, 2005, a contribution to the informal IEEE 802.16e LDPC ad-hoc group, the entirely of the document is incorporated herein by reference. These matrices are generated from a set of base parity check matrices, one base parity check matrix per code rate. The choice depends on the code rate, i.e. on the particular parity check matrix design.
(132) The method may preserve the column weight distribution, but may severely disturb the row weight distribution of the original matrix. This, in turn, causes degradation when common iterative decoding algorithms are used. This adverse effect strongly depends on the structure of the expanded matrix.
(133) This suggests that this approach fails to prescribe general rules for performing shortening and puncturing, and has an unnecessary restriction for a general case such as the one described in
(134) In general, the amount of puncturing needs to be limited. Extensive puncturing beyond certain limits paralyzes the soft decision decoder. Prior art methods, none of which specify a puncturing limit or alternatively offer some other way for mitigating the problem, may potentially compromise the performance significantly.
(135) In accordance with another embodiment of the present invention, above described shortcomings may be addressed by: (a) specifying general rules for shortening and puncturing patterns; (b) providing mechanism for q >q.sub.rate_preserved; (c) establishing a limit on the amount of puncturing; and (d) providing an algorithmic method for finding the optimal solution within the range of given system parameters.
(136) This embodiment of the present invention may be beneficially applied to both the transmitter and the receiver. Although developed for wireless systems, embodiments of the invention can be applied to any other communication system which involves encoding of variable size data packets by a fixed error correcting block code.
(137) The advantage of this invention can be summarized as providing an optimal solution to the above described problem given the range of the system parameters such as the performance, power consumption, and complexity. It comprises the following steps: 1. Based on the data packet size determine the minimum number of required modulated symbols; 2. Select the codeword length from the set of available codeword lengths; 3. In an iterative loop determine required amount of shortening and puncturing and corresponding estimated performance and add additional modulated symbol(s), if necessary; 4. Distribute amount of shortening and puncturing across all constituent codewords efficiently; and 5. Append padding bits in the last modulated symbol if necessary.
Referring to
(138) At step 213, the minimum number of modulated symbols N.sub.sym_min is calculated. Next at step 214, the codeword size N is selected, and the number of codewords to be concatenated N.sub.cwords is computed. At step 216 the required shortening and puncturing are computed, and performance estimated. If the performance criterion are met 217, the number of bits required to pad the last modulated symbol is computed 218 and the process ends 219. Where the performance criterion are not met 217, an extra modulated symbol is added 215 and the step 214 is reentered.
(139) Both the encoder and the decoder may be presented with the same input parameters in order to be able to apply the same procedure and consequently use the same codeword size, as well as other relevant derived parameters, such as the amount of shortening and puncturing for each of the codewords, number of codewords, etc.
(140) In some cases only the transmitter (encoder) has all the parameters available, and the receiver (decoder) is presented with some derived version of the encoding procedure parameters. For example, in some applications it is desirable to reduce the initial negotiation time between the transmitter and the receiver. In such cases the transmitter initially informs the receiver of the number of modulated symbols it is going to use for transmitting the encoded bits rather than the actual data packet size. The transmitter performs the encoding procedure differently taking into consideration the receiver's abilities (e.g. using some form of higher layer protocol for negotiation). Some of the requirements are relaxed in order to counteract deficiencies of the information at the receiver side. For example, the use of additional modulated symbols to enhance performance may always be in place, may be bypassed altogether, or may be assumed for the certain ranges of payload sizes, e.g. indirectly specified by the number of modulated symbols.
(141) One example of such an encoding procedure is an OFDM based transceiver, which may be used in IEEE 802.11n. In this case the reference to the number of bits per modulated symbol translates into the number of bits per OFDM symbol. In this example, the AggregationFlag parameter specified in 801.11n is used to differentiate between the case when both the encoder and the decoder are aware of actual data packet size (AggregationFlag=0) and the case when the packet size is indirectly specified by the number of required OFDM symbols (AggregationFlag=1).
(142) An exemplary algorithm in accordance with one embodiment of the present invention is with following parameters are now described:
(143) Algorithm Parameters NN.sub.max=2304, NN.sub.min=576, NN.sub.inc=576: maximum, minimum and increment of codeword lengths, effectively resulting 4 codeword lengths: 576, 1152, 1728, 2304; p.sub.max: maximum puncturing percentage, which is defined as: number of punctured bits/total number of parity bits (%).
(144) Algorithm Input R: target code rate; N.sub.CBPS: number of data bits in OFDM symbol; AggregationFlag: Boolean signaling whether PSDU is an aggregate of MPDUs (AggregationFlag=1), HT_LENGTH: number of payload octets (AggregationFlag=0), or number of OFDM symbols (AggregationFlag=1)
(145) Algorithm Output: NN: code length to use; N.sub.CodeWords: number of codewords to use; KK.sub.S,KK.sub.S_Last: number of information bits to send in first codeword(s), and in last codeword; N.sub.p, N.sub.p_Last: number of bits to puncture in first codeword (s), and in last codeword; N.sub.OFDM: number of OFDM symbols used; N.sub.PaddingBits: number of bits the last OFDM symbol is padded;
(146) Algorithm Procedure
(147) TABLE-US-00008 if(AggregationFlag == 0) { N.sub.InfoBits=8×HT_LENGTH; // in non-aggregation case HT_LENGTH is the number of payload octets N.sub.OFDM=ceil(N.sub.InfoBits/(N.sub.CBPS × R)); // minimum number of OFDM symbols } else { N.sub.OFDM = HT_LENGTH; // in aggregation case HT_LENGTH is the number of OFDM symbols N.sub.InfoBits = N.sub.OFDM × N.sub.CBPS × R; // number of info bits includes padding;MAC will use its own delineation //method to recover an aggregate payload } N.sub.CodeWords = ceil(N.sub.CBPS× N.sub.OFDM/ NN.sub.max); // number of codewords is based on maximum codeword length NN = ceil(N.sub.CBPS× N.sub.OFDM/(N.sub.CodeWords×NN.sub.inc))× N.sub.inc; // codeword length will be the larger of the closest one //to N.sub.CBPS × N.sub.OFDM/N.sub.CodeWords KK=NN×R; // number of information bits in codeword chosen MM=NN−KK; // number of parity bits in codeword chosen N.sub.ParityBits_requested =N.sub.CodeWords× MM; // total number of parity bits allocated in N.sub.OFDM symbols N.sub.ParityBits =min(N.sub.OFDM× N.sub.CRPS− N.sub.InfoBits,N.sub.ParityBits_requested); // in non-aggregation case allow adding extra OFDM symbol(s) to limit //puncturing if(AggregationFlag==0) { while(100× (N.sub.ParityBits_requested −N.sub.ParityBits)/ N.sub.ParityBits_requested >P.sub.max) { N.sub.OFDM = N.sub.OFDM+1; // extra OFDM symbol(s) are used to carry parity N.sub.ParityBits = min(N.sub.ParityBits + N.sub.CBPS,N.sub.ParityBits_requested); } } // Finding number of information bits to be sent per codeword(s), //KK.sub.S, KK.sub.S_Lasr and number of bits the codeword(s) will be punctured N.sub.P , //and N.sub.P_Last, Making sure that last codeword may only be shortened // more then others, and punctured less then others. KK.sub.S =ceil(N.sub.InfoBits/ N.sub.CodeWords); KK.sub.S_Last = N.sub.InfoBits ~ KK.sub.S × ( N.sub.CodeWords −1); MM.sub.P =min(MM, floor(N.sub.ParityBits/.sub.CodeWords); MM.sub.P_Last = min(MM, N.sub.ParityBits − MM.sub.P × (N.sub.CodeWords − 1)); N.sub.P =MM − MM.sub.P; N.sub.P_Last =MM− MM.sub.P_Last; // Finally, calculating number of padding bits in last OFDM symbol N.sub.PaddingBits = N.sub.OFDM × N.sub.CBPS − N.sub.InfoBits −N.sub.ParityBits;
(148) Each of those features will be now described in more detail.
(149) (a) General Rules for Shortening and Puncturing
(150) Much effort has been spent to come up with designs of LDPC parity check matrices such that the derived codes provide optimum performance. Examples include: T. J. Richardson et al., “Design of Capacity-Approaching Irregular Length Low-Density Parity-Check Codes,” IEEE Transactions on Information Theory, vol. 47, February 2001 and S. Y. Chung, et al., “Analysis of Sum-Product Decoding of Low-Density Parity-Check Codes Using a Gaussian Approximation,” IEEE Transactions on Information Theory, vol. 47, February 2001, both of which are incorporated herein by reference, are examples. These papers show that, in order to provide optimum performance, a particular variable nodes degree distribution should be applied. Degree distribution refers here to the distribution of the column weights in a parity check matrix. This distribution, in general, depends on the code rate and the size of the parity check matrix, or codeword. It is desirable that the puncturing and shortening pattern, as well as the number of punctured/shortened bits, are specified in such a way that the variable nodes degree distribution is preserved as much as possible. However, since shortening and puncturing are qualitatively different operations, different rules apply to them, as will now be explained.
(151) (b) Rules for Shortening
(152) Shortening of a code is defined as sending less information bits than possible with a given code, K′<K. The encoding is performed by: taking K′ bits from the information source, presetting the rest (K-K′) of the information bit positions in the codeword to a predefined value, usually 0, computing M parity bits by using the full M×N parity check matrix, and finally forming the codeword to be transmitted by concatenating K′ information bits and M parity bits. One way to determine which bits to shorten in the data portion of the parity check matrix, H.sub.d (31 in
(153) 3 3 3 8 3 3 3 8 3 3 3 8
(154) When discarding columns, the aim is to ensure that the ration of ‘3’s to ‘8’s remains close to optimal, say 1:3 in this case. Obviously it cannot be 1:3 when one to three columns are removed. In such circumstances, the removal of 2 columns might result in e.g.:
(155) 3 3 8 3 3 8 3 3 3 8
(156) giving a ratio of ˜1:3.3 and the removal of a third column—one with weight ‘8’—might result in:
(157) 3 3 3 3 8 3 3 3 8
(158) thus preserving a ratio of 1:3.5, which is closer to 1:3 than would be the case where the removal of the third column with weight ‘3’, which results in:
(159) 8 3 3 3 8 3 3 3 8
(160) giving a ratio of 1:2.
(161) It is also important to preserve approximately constant row weight throughout the shortening process.
(162) An alternative to the above-described approach is to prearrange columns of the part of the parity check matrix, such that the shortening can be applied to consecutive columns in H.sub.d. Although perhaps suboptimal, this method keeps the degree distribution of H.sub.d close to the optimum. However, the simplicity of the shortening pattern, namely taking out the consecutive columns of H.sub.d, gives a significant advantage by reducing complexity. Furthermore, assuming the original matrix satisfies this condition, approximately constant row weight is guaranteed. An example of this concept is illustrated in
(163) After rearranging the columns of the H.sub.d part of the original matrix, the new matrix takes on the form 221 shown in
(164) In the case of a regular column parity check matrix, or more generally, approximately regular, or regular and approximately regular only in the data part of the matrix, H.sub.d, the method described in the previous paragraph is still preferred compared to the existing random or periodic/random approach. The method described here ensures approximately constant row weight, which is another advantage from the performance and the implementation complexity standpoint.
(165) (c) Puncturing
(166) Puncturing of a code is defined as removing parity bits from the codeword. In a wider sense, puncturing may be defined as removing some of the bits, either parity bits or data bits or both, from the codeword prior to sending the encoded bits to the modulator block and subsequently over the channel. The operation of puncturing, increases the effective code rate. Puncturing is equivalent to a total erasure of the bits by the channel. The soft iterative decoder assumes a completely neutral value corresponding to those erased bits. In case that the soft information used by the decoder is the log-likelihood ratio, this neutral value is zero.
(167) Puncturing of LDPC codes can be given an additional, somewhat different, interpretation. An LDPC code can be presented in the form of the bipartite graph of
(168) Each variable node 231 is connected 234 by edges, for example 233, to all the check nodes 232 in which that particular bit participates. Similarly, each check node (corresponding to a parity check equation) is connected by a set of edges 237 to all variable nodes corresponding to bits participating in that particular parity check equation. If a bit is punctured, for example node 235, then all the check nodes connected to it, those connected by thicker lines 236, are negatively affected. Therefore, if a bit chosen for puncturing participates in many parity check equations, the performance degradation may be very high. On the other hand, since the only way that the missing information (corresponding to the punctured bits) can be recovered is from the messages coming from check nodes those punctured bits participate in, the more of those the more successful recovery may be. Faced with contradictory requirements, the optimum solution can be found somewhere in the middle. These general rules can be stated as following: Bits selected for puncturing should be chosen such that each one is connected to as few check nodes as possible. This can be equivalently stated as follows: bits selected for puncturing should not be the ones corresponding to the heavy-weight, or strong columns, i.e. columns containing large number of non-zero elements, 1's in this particular case. Bits selected for puncturing should be chosen such that they all participate in as many parity check equations as possible.
(169) Some of these trade-offs can be observed from
(170)
(171) In
(172) It can be seen from the
(173) The matrix in
(174) As discussed previously, in the case where the preservation of the exact code rate is not mandatory, the shortening-to-puncturing ratio can be chosen such that it guarantees preservation of the performance level of the original code.
(175) Normalizing the shortening-to-puncturing ratio, q, as follows:
q.sub.normalized=(N.sub.shortened/N.sub.punctured)/[R/(1−R)],
(176) means that q becomes independent of the code rate, R. Therefore, q.sub.normalized=1, corresponds to the rate preserving case of combined shortening and puncturing. However, if the goal is to preserve performance, this normalized ratio must be greater than one: q.sub.normalized>1. It was found through much experimentation that one: q.sub.normalized in the range of 1.2-1.5 complies with the performance preserving requirements.
(177) In the case of a column regular parity check matrix, or more generally, approximately regular, or regular and approximately regular only in the data part of the matrix, H.sub.d the method described above in accordance with one embodiment of the present invention is still preferred compared to the existing random or periodic/random approach since the present invention ensures approximately constant row weight, which provides another advantage from both the performance and the implementation complexity standpoints.
(178) A large percentage of punctured bits paralyzes the iterative soft decision decoder. In the case of LDPC codes this is true even if puncturing is combined with some other operation such as shortening or extending the code. One could conclude this by studying the matrix 250 of
P.sub.puncture=100×(N.sub.puncture/M),
(179) then it can be seen that the matrix 250 from
(180) Some of the embodiments of the present invention may include the following characteristics: Shortening, or combined shortening and puncturing is applied in order to provide a large range of codeword sizes from a single parity check matrix. The effective code rate of the code defined by the parity check matrix modified by shortening and puncturing is equal to or less than the original code rate. Shortening is performed so that the column weight distribution of the modified matrix is optimal for the new matrix size and code rate. Another solution is to keep the column weight distribution of the new matrix only approximately optimum. Shortening is performed so that the approximately uniform row weight is preserved. Puncturing is performed so that each of the bits selected for puncturing is connected to as few check nodes as possible. Puncturing is performed so that the bits selected for puncturing all participate in as many parity check equations as possible. Puncturing is performed so that the approximately uniform row weight is preserved. A suboptimal but computationally efficient method is to first rearrange the columns of the data portion of the parity check matrix, H.sub.d, by applying the preceding rules assuming that shortening is applied to a group of consecutive bits of the data portion of the parity check matrix and puncturing is applied to another group of consecutive bits of the data portion of the parity check matrix as illustrated by the example matrix 250 shown in
(181) The system, apparatus, and method as described above are preferably combined with one or more matrices shown in the
(182) The matrices in
(183) A first group of matrices (
(184) A further matrix (
(185) The rate R=¾ matrices (
(186) The rate R=⅚ matrix (
(187) The two rate R=⅚ matrices (
s′=floor{s. (L/96)},
(188) where s is the right circular shift corresponding to the maximum codeword size (for L=Lmax=96), and it is specified in the matrix definitions.
(189) The invention can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations thereof. Apparatus of the invention can be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a programmable processor; and method actions can be performed by a programmable processor executing a program of instructions to perform functions of the invention by operating on input data and generating output. The invention can be implemented advantageously in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. Each computer program can be implemented in a high-level procedural or object oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language. Suitable processors include, by way of example, both general and special purpose microprocessors. Generally, a processor will receive instructions and data from a read-only memory and/or a random access memory. Generally, a computer will include one or more mass storage devices for storing data files. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits). Further, a computer data signal representing the software code which may be embedded in a carrier wave may be transmitted via a communication network. Such a computer readable memory and a computer data signal are also within the scope of the present invention, as well as the hardware, software and the combination thereof.
(190) While particular embodiments of the present invention have been shown and described, changes and modifications may be made to such embodiments without departing from the true scope of the invention.