MULTI-MODE UNROLLED POLAR DECODERS
20170230059 · 2017-08-10
Inventors
- PASCAL GIARD (Ste-Marthe-sur-le-Lac, CA)
- Gabi Sarkis (San Diego, CA)
- Warren Gross (Montreal, CA)
- Claude THIBEAULT (Brossard, CA)
Cpc classification
H03M13/134
ELECTRICITY
H03M13/6516
ELECTRICITY
International classification
H04L1/00
ELECTRICITY
Abstract
There is described a multi-mode unrolled decoder. The decoder comprises a master code input configured to receive a polar encoded master code of length N carrying k information bits and N−k frozen bits, decoding resources comprising processing elements and memory elements connected in an unrolled architecture and defining an operation path between the master code input and an output, for decoding a polar encoded code word, at least one constituent code input configured to receive a polar encoded constituent code of length N/p carrying j information bits and N/p−j frozen bits, where p is a power of 2, and at least one input multiplexer provided in the operation path to selectively transmit N/p bits of one of the master code and the constituent code to a subset of the decoding resources.
Claims
1. A decoder for polar encoded code words comprising: a master code input configured to receive a polar encoded master code of length N carrying k information bits and N−k frozen bits; decoding resources comprising processing elements and memory elements connected in an unrolled architecture and defining an operation path between the master code input and an output, for decoding a polar encoded code word; at least one constituent code input configured to receive a polar encoded constituent code of length N/p carrying j information bits and N/p−j frozen bits, where p is a power of 2; and at least one input multiplexer provided in the operation path to selectively transmit N/p bits of one of the master code and the constituent code to a subset of the decoding resources.
2. The decoder of claim 1, further comprising at least one output multiplexer provided in the operation path downstream from the at least one input multiplexer to selectively transmit an estimated code word corresponding to the N/p bits of one of the master code and the constituent code towards the output.
3. The decoder of claim 1, wherein the master code is composed of q constituent codes, and the at least one input multiplexer comprises m input multiplexers, where m<q.
4. The decoder of claim 3, further comprising s output multiplexers provided in the operation path downstream from the m input multiplexers to selectively transmit an estimated code word corresponding to the N/p bits of one of the master code and at least one of the q constituent codes towards the output.
5. The decoder of claim 4, wherein s<m.
6. The decoder of claim 1, further comprising a controller configured for generating multiplexer select signals to route data through the decoder.
7. The decoder of claim 1, wherein the unrolled architecture is partially pipelined.
8. The decoder of claim 1, wherein the decoder is configured to apply successive-cancellation decoding to the polar encoded code words.
9. The decoder of claim 1, wherein the decoder is configured to apply list-based decoding to the polar encoded code words.
10. A method for decoding polar encoded codes using an unrolled polar code decoder, the method comprising: receiving, at a first intermediate node of an operation path defined between an input and an output of an unrolled decoder formed of decoding resources for decoding a master code of length N, a first polar encoded constituent code of length N/p, where p is a power of 2; transmitting the first polar encoded constituent code to a first subset of the decoding resources though a first input multiplexer; estimating a first code word for the first polar encoded constituent code through the first subset of decoding resources; and outputting, at the output, the estimated first code word corresponding to the first polar encoded constituent code.
11. The method of claim 10, further comprising transmitting the estimated first code word towards the output through an output multiplexer.
12. The method of claim 10, further comprising: receiving, at a second intermediate node of the operation path, a second polar encoded constituent code of length N/p, where p is a power of 2; transmitting the second polar encoded constituent code to a second subset of the decoding resources though a second input multiplexer; estimating a second code word for the second polar encoded constituent code through the second subset of decoding resources; and outputting, at the output, the estimated second code word corresponding to the second polar encoded constituent code.
13. The method of claim 12, wherein the first polar encoded constituent code and the second polar encoded constituent code are of different lengths.
14. The method of claim 12, wherein the second subset of decoding resources is a subset of the first subset of decoding resources.
15. The method of claim 12, further comprising transmitting the estimated first code word and the estimated second code word towards the output through at least one output multiplexer.
16. The method of claim 15, wherein the first polar encoded constituent code and the second polar encoded constituent code are of a same length, and wherein the estimated first code word and the estimated second code word are transmitted towards the output through a same one of the at least one output multiplexer.
17. The method of claim 12, wherein the first polar encoded constituent code and the second polar encoded constituent code are sibling constituent codes that share a parent node in a decoder tree.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0013] Further features of the present disclosure will become apparent from the following detailed description, taken in combination with the appended drawings, in which:
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022] It will be noted that throughout the appended drawings, like features are identified by like reference numerals.
DETAILED DESCRIPTION
[0023] The present disclosure is directed to polar codes and more specifically to fully-unrolled polar code decoders supporting multiple polar code lengths and data rates, referred to herein as a multi-mode unrolled decoder.
[0024] An (N, k) polar code has length N, carries k information bits and is of rate R=k/N. The other N−k bits, so called frozen bits, are set to a predetermined value, usually zero but possibly another value, during the encoding process. Non-systematic polar encoding can be represented as x=uF.sub.N where u is a row vector containing both information and frozen bits, and F.sub.N is a generator matrix. This F.sub.N can be defined recursively so that F.sub.N=F.sub.2.sup. log.sup.
and is the Kronecker product. Determining the optimal locations for the information bits depends on the channel type and condition.
[0025] Encoding schemes for polar codes can be either non-systematic or systematic. Systematic polar codes offer better bit-error-rate (BER) than their non-systematic counterparts, while maintaining the same frame-error-rate (FER). The multi-mode unrolled decoder supports systematic and non-systematic codes. Both encoding types use F.sub.N and as F.sub.N is built recursively, so are polar codes i.e. a code of length N is the concatenation of two codes of length N/2.
[0026] Referring to
[0027] Polar codes can also be represented as decoder trees, such as depicted in
[0028] By definition, and like the master code, a constituent code of length N/2 is in turn the concatenation of two polar codes of length N/4, and so on until the leaf nodes are reached. As such, the decoding of a polar code of length N can be seen as the decoding of two constituent codes of length N/2, or of four constituent codes of length N/4, etc. For example, and as shown in the graph representation of
[0029] It should be noted that sibling constituent codes with a same parent node share a special relationship. Let us consider the polar code (constituent code) of length N=8 taking root in v as illustrated in
[0030] Going down a left edge of each tree/sub-tree within the SC decoder tree 200a, α, is calculated with the min-sum approximation given by Equation (1) for
where α.sub.v is the input to the node and N.sub.v the width of α.sub.v. Now traversing the right edge of each tree/sub-tree within the SC decoder tree 200a, α.sub.r is calculated using Equation (2) for
where β.sub.l is the bit estimate from the LHS child.
α.sub.l[i]=sgn(α.sub.v[i].Math.α.sub.v[i+N.sub.v/2])min(|α.sub.v[i],|α.sub.v[i+N.sub.v/2]|) (1)
[0031] Once a leaf node is reached, the bit estimate is set to zero when it corresponds to a frozen bit location. Otherwise, it is calculated by threshold detection on α.sub.v. Going back up a RHS edge the bit estimates from both children are combined to generate the node's bit-estimate vector β.sub.v[i] as given by Equation (3), where ⊕ is the modulo-2 addition (XOR).
[0032] In a Simplified SC (SSC) algorithm, decoder tree nodes are split into three categories: Rate-0, Rate-1, and Rate-R nodes. Rate-0 nodes are subtrees whose leaf nodes all correspond to frozen bits. We do not need to use a decoding algorithm on such a subtree as the exact decision, by definition, is always the all-zero vector. Rate-1 nodes are subtrees where all leaf nodes carry information bits, none are frozen. The maximum-likelihood decoding rule for these nodes is to take a hard decision on input log-likelihood ratios (LLRs):
[0033] for 0≦i<N.sub.v. With a fixed-point representation, this operation amounts to copying the most significant bit of the input LLRs.
[0034] Rate-R nodes, where 0<R<1, are subtrees such that leaf nodes are a mix of information and frozen bits. Instead of always using the SC or SSC algorithm, some Rate-R nodes corresponding to specific frozen-bit locations can be decoded using algorithms with lower complexity and latency.
[0035] The F and G operations are among the functions used in the conventional SC decoding algorithm and are calculated using (1) and (2), respectively. G0R is a special case of the G operation where the left child is a frozen node i.e. β.sub.l is known a priori to be the all-zero vector of length N.sub.v/2.
[0036] As defined by equation (3), the Combine operation generates the bit estimate vector. A C0R operation is a special case of the Combine operation where the LHS constituent code, β.sub.i, is a Rate-0 node.
[0037] In a Repetition node, all leaf nodes are frozen bits, with the exception of the node that corresponds to the most RHS leaf in a tree. At encoding time, the only information bit gets repeated over the N.sub.v outputs. The information bit can be estimated by using threshold detection over the sum of the input LLRs α.sub.v:
[0038] Where β.sub.v gets replicated N.sub.v times to create the bit-estimate vector.
[0039] A single-parity-check (SPC) node is a node such that all leaf nodes are information bits with the exception of the node at the least significant position (LHS leaf in a tree). To decode an SPC code, we start by calculating the parity of the input LLRs:
[0040] The estimated bit vector is then generated by reusing the calculated β.sub.v above unless the parity constraint is not satisfied, i.e. is different than zero. In that case, the estimated bit corresponding to the input with the smallest LLR magnitude is flipped:
β.sub.v[i]=β.sub.v[i]⊕1, where i=arg min(|α.sub.v[j])
[0041]
[0042] The multi-mode unrolled decoder borrows from the Fast-SCC algorithm in that it uses specialized nodes and operations described above to reduce the decoding latency. However, the family of architectures used differ from those known from Fast-SCC decoders.
[0043] In an unrolled decoder, each and every operation required is instantiated so that data can flow through the decoder with minimal control. A dedicated unrolled decoder is only applicable to an (N, k) polar code of length N that carries k information bits and is of rate R=k/N. The multi-mode unrolled decoder can be used to decode polar encoded code words of length N as well as polar encoded code words of length<N. The multi-mode unrolled decoder comprises decoding resources, such as processing elements and memory elements, which are connected together in an unrolled architecture. The decoding resources define an operation path between an input, referred to as a master code input, and an output. The master code input is configured to receive a polar encoded master code of length N carrying k information bits and N−k frozen bits. The multi-mode unrolled decoder comprises at least one constituent code input configured to receive a polar encoded constituent code of length N/p carrying j information bits and N/p−j frozen bits, where p is a power of 2. The multi-mode unrolled decoder comprises at least one input multiplexer provided in the operation path to selectively transmit N/p bits of the master code or the constituent code to a subset of the decoding resources.
[0044]
[0045] An unrolled decoder for a polar code of length N is composed of a subset of decoding resources for two polar codes of length N/2, which are each composed of a subset of decoding resources for two polar codes of length N/4, and so on. Thus, control and routing logic is provided in order to directly feed and read from the subset of decoding resources for constituent codes of length smaller than N. The end result is a multi-mode decoder supporting frames of various lengths and code rates.
[0046] In some embodiments, at least one output multiplexer is provided in the operation path downstream from the at least one input multiplexer. The output multiplexer selectively transmits an estimated code word corresponding to the N/p bits of the master code or the constituent code towards the output. While the output multiplexer is not strictly required, not having it would increase both time complexity and decoder operation complexity. In some embodiments, two or more constituent codes may share an output multiplexer. For example, this may be done when at least two constituent codes of the same length are used.
[0047] Consider the decoder tree 200b shown
[0048] Although not illustrated in
[0049] Referring to
[0050] At 504, the constituent code is transmitted to a subset of the decoding resources through an input multiplexer, for example m.sub.1, m.sub.2, or m.sub.3. At 506, the data provided to the input multiplexer travels through the various memory and processing elements of the subset of the decoding resources to produce an estimated code word. Optionally at 508, the estimated code word is transmitted towards the output through an output multiplexer. At 510, the estimated code word for the constituent code is output from the multi-mode unrolled decoder.
[0051] In some embodiments, the method comprises performing the steps for a first constituent code and performing the steps for a second constituent code, for example using two different multiplexers such as m.sub.1 and m.sub.2. The constituent codes may be input on a same or different clock cycles. Alternatively, or in combination therewith, the two constituent codes may be input in the same multiplexer, on subsequent clock cycles. The clock cycles may be consecutive or spaced apart. The two constituent codes may be of a same or different length, and they may transmitted towards the output using a same or different output multiplexer.
[0052] It will be understood that the method may be used to decode polar codes of length N as well as any number of polar codes of length N/p, where p is a power of 2, provided the suitable input and control logic is present in the multi-mode unrolled decoder. For example, when applied to the embodiment of
[0053] Note that the SC-based multi-mode unrolled decoder may also be applied to a list-based decoder. In a list-based multi-mode unrolled decoder, a decoder for a polar code smaller than the master code is part of a larger decoder used to accommodate for a list size of L. Routing and control logic may therefore also be added to the list-based decoder in order to allow for decoding of code words smaller than length N.
[0054] In some embodiments, a master code may be assembled using two optimized constituent codes in order to increase the number of optimized polar codes available. Doing so, the number of information bits, or the code rate, of the second largest supported codes can be selected. As an example, a master code of length 2048 is constructed by concatenating two constituent codes of length 1024. The LHS and RHS constituent codes are chosen to have a rate of 1/2 and of 5/6, respectively. As a result, the assembled master code has rate of 2/3. The location of the frozen bits in the master code is dictated by its constituent codes. Note that the constituent code with the lowest rate is put on the left, and the one with the highest rate on the right, in order to minimize the coding loss associated with a non-optimized polar code.
[0055]
[0056] From
[0057] The location of the frozen bits in non-optimized constituent codes is dictated by their parent code, i.e. the master code (polar code) of length N. In other words, if the master code of length N has been assembled from two optimized (constituent) polar codes of length N/2, the shorter optimized codes of length N/2 determine the location of the frozen bits in their respective constituent codes of length<N/2. Otherwise, the master code dictates the frozen bit locations for all constituent codes.
[0058] Assuming that the decoding algorithm takes advantage of the a priori knowledge of these locations, the code rate and frozen bit locations of the constituent codes cannot be changed at execution time. However, there are many constituent codes to choose from and code shortening can be used to create more, e.g. in order to obtain a specific number of information bits or code rate.
[0059] Because of the polarization phenomenon, given any two sibling constituent codes, the code rate of the LHS is always lower than that of the RHS for a properly constructed polar code. That property can be beneficially exploited as, in many wireless applications, it is desirable to offer a variety of codes of both high and low rates. It should be noted that not all constituent codes within a master code are of practical use e.g. codes of very high rate offer negligible coding gain over an uncoded communication. For example, among the four constituent codes of length 4 included in the (16,12) polar code illustrated in
[0060]
[0061] If a decoding algorithm taking advantage of the a priori knowledge of the frozen bit locations is used in the unrolled decoder, such as Fast-SSC, the latency will vary even among constituent codes of the same length. However, the coded throughput will not. The coded throughput of a polar code of length N will be double that of a constituent code of N/2, which in turn, is double that of a constituent code of length N/4, and so on.
[0062] In an unrolled decoder, the throughput is defined by the code length N.sub.υ, the clock frequency in Hz f and the initiation interval I. The coded and information throughput are given by Equations (4) and (5) respectively. For a master code N.sub.υ=N.
[0063] In wireless communication standards where multiple codes lengths and rates are supported, the peak information throughput may be achieved with the longest code that has both the greatest latency and highest code rate. Another possible scenario is to use a low-rate master code, e.g. R=1/3, that is more powerful in terms of error-correction performance. The resulting multi-mode unrolled decoder would reach its peak information throughput with the longest constituent code of length N/2 that has the highest code rate, a code with a significantly lower decoding latency than that of the master code.
[0064] Results for two implementations of the multi-mode unrolled decoder are presented below. These examples are built around (1024, 853) and (2048, 1365) master codes. In the following, the former is referred to as the decoder supporting a maximum code length N.sub.MAX=1024 and the latter as the decoder with N.sub.MAX=2048. A total of ten polar codes were selected for the decoder supporting codes of lengths up to 2048. The other decoder with N.sub.MAX=1024 has eight modes corresponding to a subset of the ten polar codes supported by the bigger decoder.
[0065] For the decoder with N.sub.MAX=1024, the Repetition and SPC nodes were constrained to a maximum size N.sub.v of 8 and 4 respectively. For the decoder with N.sub.MAX=2048 it was found to be more beneficial to lower the execution frequency and increase the maximum sizes of the Repetition and SPC nodes to 16 and 8, respectively. Additionally, the decoder with N.sub.MAX=2048 also uses RepSPC nodes to reduce latency.
[0066] In the examples, the multi-mode unrolled decoders are built with sufficient memory to accommodate storing an extra frame at the input, and to preserve an estimated code word at the output. As a result, the next frame can be loaded while a frame is being decoded. Similarly, an estimated code word can be read while the next frame is being decoded. We define decoding latency to include the time required to load channel LLRs, decode a frame and offload the estimated code word.
[0067] The quantization used was determined by running fixed point simulations with bit-true models of the decoders. A smaller number of bits is used to store the channel LLRs compared to that of the other LLRs used in the decoder. All LLRs use 2's complement representation and share the same number of fractional bits. We denote quantization as Q.sub.i, Q.sub.c, Q.sub.f, where Q.sub.c is the total number of bits to store a channel LLR, Q.sub.i is the total the number of bits used to store internal LLRs and Q.sub.f is the number of fractional bits in both. Q.sub.i and Q.sub.c both include the sign bit.
[0068] ASIC synthesis results are for the 65 nm CMOS GP technology from TSMC and are obtained with Cadence RTL Compiler. Unless indicated otherwise, all results are for the worst-case library at a supply voltage of 0.72 V with an operating temperature of 125° C. Power consumption estimations are also obtained from Cadence RTL Compiler, switching activity is derived from simulation vectors. Only registers were used for memory due to the lack of access to an SRAM compiler.
[0069] Table I shows the results for various initiation intervals. Besides the effect on throughput, increasing the initiation interval causes a significant reduction in memory requirements without significantly affecting combinational logic. Since area is largely dominated by registers, increasing the initiation interval has great effect on the total area. For example, using I=50 results in an area that is more than 10 times smaller, at the cost of a throughput that is 50 times lower. Table I also shows that reducing the area has a direct effect on the estimated power consumption, which significantly drops as I.
TABLE-US-00001 TABLE I Tot. Area Log. Area Mem. Area T/P Power Energy I (mm.sup.2) (mm.sup.2) (mm.sup.2) (Gbps) (mW) (pJ/bit) 1 12.369 0.60 11.75 512.0 3,830 7.5 4 4.921 0.64 4.24 128.0 1,060 8.3 50 1.232 0.65 0.56 10.2 107 10.5 167 0.998 0.63 0.34 3.1 62 20.0
[0070] Increasing the initiation interval I offers a diminishing return as it gets closer to the maximum, 167 for the example (1024, 512) code. Also, as I is increased, the energy efficiency is reduced.
[0071] Tables II and III show the effect of the code length on area, decoding latency, coded throughput, power consumption, and on energy efficiency for polar codes of short to moderate lengths. Table II contains results for the fully-unrolled deeply-pipelined architecture (I=1) and the code rate R is fixed to 1/2 for all polar codes. Table III contains results for the fully-unrolled partially-pipelined architecture where the maximum initiation interval (I.sub.max) is used and the code rate R is 5/6.
TABLE-US-00002 TABLE II Tot. Log. Mem. Area Area Area Latency T/P Power Energy N (mm.sup.2) (mm.sup.2) (mm.sup.2) (ns) (Gbps) (mW) (pJ/bit) 128 0.349 0.05 0.29 152 64 105 1.6 256 1.121 0.12 0.99 268 128 342 2.7 512 3.413 0.27 3.14 408 256 1,050 4.0 1024 12.369 0.60 11.75 728 512 3,830 7.5 2048 43.541 1.32 42.16 1,304 1,024 13,526 13.2
[0072] As shown in Table II, with a deeply-pipelined architecture, logic area usage almost grows as N log.sub.2 N, whereas memory area is closer to being quadratic in code length N. The logic area required for a deeply-pipelined unrolled decoder implemented in 65 nm ASIC technology can be approximated with an accuracy greater than 98% using C.Math.N log.sub.2 N, where the constant C is set to 1/17,000.
[0073] Curve fitting shows that the memory area is quadratic with code length N. Let the memory area be defined by a+bN+cN.sup.2, setting a=0.249, b=2.466×10.sup.−3 and c=8.912×10.sup.−6 results in a standard error of 0.1839.
[0074] As shown in Table II, throughput exceeding 1 Tbps and 500 Gbps can be achieved with a deeply-pipelined decoder for polar codes of length 2048 and 1024, respectively. As the memory area grows quadratically with the code length, the amount of energy required to decode a bit increases with the code length.
TABLE-US-00003 TABLE III Mem. Tot. Area Area Latency T/P Power Energy N I (mm.sup.2) (mm.sup.2) (μs) (Gbps) (mW) (pJ/bit) 1024 206 0.793 0.28 0.646 2.5 51 20.5 2048 338 1.763 0.61 0.888 3.0 108 35.6 4096 665 4.248 1.44 1.732 3.1 251 81.5
[0075] For a partially-pipelined architecture with I.sub.max, both the memory and total area scale linearly with N. The power consumption is shown to almost scale linearly as well. The results of Table III also show that it was possible to synthesize ASIC decoders for larger code lengths than what was possible with a deeply-pipelined architecture.
[0076] The effect of using different code rates for a polar code of length N=1024 is shown in Table IV. We note that the higher rate codes do not have noticeably lower latency compared to the rate-1/2 code. This is due to limiting the width of SPC nodes to N.sub.SPC=4. The result is that long SPC codes are implemented as trees whose leftmost child is a width-4 SPC node and the others are all rate-1 nodes. Thus, for each additional stage (log.sub.2 N.sub.v−log.sub.2 N.sub.SPC) of an SPC code of length N.sub.v>N.sub.SPC, four nodes with a total latency of 3 clock cycles are required: F, G followed by I, and Combine. This brings the total latency of decoding a long SPC code to 3(log.sub.2 N.sub.v−log.sub.2 N.sub.SPC)+1 clock cycles.
TABLE-US-00004 TABLE IV Tot. Area Mem. Area Latency Power Energy R (mm.sup.2) (mm.sup.2) (CCs) (ns) (mW) (pJ/bit) 1/2 12.369 11.75 364 727 3,830 7.5 2/3 13.049 12.45 326 651 4,041 6.2 3/4 15.676 15.05 373 745 4,865 6.5 5/6 14.657 14.05 323 645 4,549 7.1
[0077] From Table IV, it can be seen that varying the rate does not affect the logic area that remains almost constant at approximately 0.61 mm.sup.2. Memory, in the form of registers, dominates the decoder area. Therefore, the estimated power consumption scales according to the memory area.
[0078] To decode a frame, an SC decoder needs to load a frame, visit all Σ.sub.i=1.sup.log.sup.
[0079]
[0080] It can be seen from
[0081] Table V shows the latency and information throughput for both decoders with N.sub.MAXε{1024, 2048}. To reduce the area and latency while retaining the same throughput, the initiation interval I can be increased along with the clock frequency.
[0082] If both decoders have initiation intervals of 20, Table V assumes clock frequencies of 500 MHz and 250 MHz for the decoders with M.sub.max=1024 and N.sub.max=2048, respectively. While their master codes differ, both decoders feature a peak information throughput in the vicinity of 20 Gbps. For the decoder with the smallest N.sub.max, the seven other polar codes have an information throughput in the multi-gigabit per second range with the exception of the shortest and lowest-rate constituent code. That (128, 39) constituent code still has an information throughput close to 1 Gbps. The decoder with N.sub.max=2048 offers multi-gigabit throughput for most of the supported polar codes. The minimum information throughput is also with the (128, 39) polar code at approximately 500 Mbps.
TABLE-US-00005 TABLE V Info. T/P Latency Latency (Gbps) (CCs) (ns) Code Rate N.sub.max = (N, k) (k/N) 1024 2048 1024 2048 1024 2048 (2048, 1365) 2/3 — 17.1 — 503 — 2,012 (1024, 853) 5/6 21.3 10.7 323 236 646 944 (1024, 512) 1/2 — 6.4 — 265 — 1,060 (512, 490) 19/20 12.3 6.2 95 75 190 300 (512, 363) 7/10 9.1 4.5 226 159 452 636 (256, 228) 9/10 5.7 2.6 86 61 172 244 (256, 135) 1/2 3.4 1.7 138 96 276 384 (128, 108) 5/6 2.7 1.4 54 40 108 160 (128, 96) 3/4 2.4 1.2 82 52 164 208 (128, 39) 1/3 0.98 0.49 54 42 108 168
[0083] In terms of latency, the decoder with N.sub.max=1024 requires 646 ns to decode its longest supported code. The latency for all the other codes supported by that decoder is under 500 ns. Even with its additional dedicated node and relaxed maximum size constraint on the Repetition and SPC nodes, the decoder with N.sub.max=2048 has greater latency overall because of its lower clock frequency. For example, its latency is of 2.01 us, 944 ns and 1.06 us for the (2048, 1365), (1024, 853) and (1024, 512) polar codes, respectively.
[0084] Using the same nodes and constraints as for N.sub.max=1024, the N.sub.max=2048 decoder would allow for greater clock frequencies. While 689 clocks cycles would be required to decode the longest polar code instead of 503, a clock of 500 MHz would be achievable, effectively reducing the latency from 2.01 μs to 1.38 μs and doubling the throughput. However, this reduction comes at the cost of much greater area and an estimated power consumption close to 1 W.
[0085] Table VI shows the synthesis results along with power consumption estimations for the two implementations of the multi-mode unrolled decoder. The first two columns are for the decoder with N.sub.max=1024, based on the (1024, 853) master code. It was synthesized for clock frequencies of 500 MHz and 650 MHz, respectively, with initiation intervals I of 20 and 26. The third and fourth columns are for the decoders with N.sub.max=2048, built from the assembled (2048, 1365) polar code. These decoders have an initiation interval I of 20 or 28, with lower clock frequencies of 250 MHz and 350 MHz, respectively. For comparison with the prior art, the same table also includes results for a dedicated partially-pipelined decoder for a (1024, 512) polar code.
TABLE-US-00006 TABLE VI Multi-mode Dedicated Algorithm Fast-SSC Fast-SSC Technology 65 nm 65 nm N.sub.max 1024 2048 1024 Code (1024, 853) (2048, 1365) (1024, 512) Init. Interval (I) 20 26 20 28 20 Supply (V) 0.72 1.0 0.72 1.0 1.0 Oper. temp. (° C.) 125 25 125 25 25 Area (mm.sup.2) 1.71 1.44 4.29 3.58 1.68 Area @65 nm 1.71 1.44 4.29 3.58 1.68 (mm.sup.2) Frequency (MHz) 500 650 250 350 500 Latency (μs) 0.65 0.50 2.01 1.44 0.73 Coded T/P (Gbps) 25.6 25.6 25.6 25.6 25.6 Sust. Coded T/P 25.6 25.6 25.6 25.6 25.6 (Gbps) Area Eff. 15.42 17.75 5.97 7.16 15.27 (Gbps/mm.sup.2) Power (mW) 226 546 379 740 386 Energy (pJ/bit) 8.8 27.3 14.8 28.9 15.1
[0086] For consistency, only the largest polar code supported by each of the multi-mode unrolled decoders is used. While the area of the decoder with N.sub.MAX=2048 is over twice that of the one with N.sub.MAX=1024, the master code has twice the length and that decoder implementation supports two more modes. Both of the multi-mode decoders have a coded throughput that is an order of magnitude greater than the dedicated decoder.
[0087] Implementation of the embodiments described above may be done in various ways. For example, the techniques, blocks, steps and means may be implemented in hardware, software, or a combination thereof. For a hardware implementation, the processing elements may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above and/or a combination thereof.
[0088] Furthermore, embodiments may be implemented by hardware, software, scripting languages, firmware, middleware, microcode, hardware description languages and/or any combination thereof. When implemented in software, firmware, middleware, scripting language and/or microcode, the program code or code segments to perform the necessary tasks may be stored in a machine readable medium, such as a storage medium. A code segment or machine-executable instruction may represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a script, a class, or any combination of instructions, data structures and/or program statements. A code segment may be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters and/or memory content. Information, arguments, parameters, data, etc. may be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, token passing, network transmission, etc.
[0089] For a firmware and/or software implementation, the methodologies may be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein. Any machine-readable medium tangibly embodying instructions may be used in implementing the methodologies described herein.
[0090] The above description is meant to be exemplary only, and one skilled in the relevant arts will recognize that changes may be made to the embodiments described without departing from the scope of the invention disclosed. For example, the blocks and/or operations in the flowcharts and drawings described herein are for purposes of example only. There may be many variations to these blocks and/or operations without departing from the teachings of the present disclosure. For instance, the blocks may be performed in a differing order, or blocks may be added, deleted, or modified. The structure illustrated is thus provided for efficiency of teaching. The present disclosure may be embodied in other specific forms without departing from the subject matter of the claims. Also, one skilled in the relevant arts will appreciate that while the systems, methods and computer readable mediums disclosed and shown herein may comprise a specific number of elements/components, the systems, methods and computer readable mediums may be modified to include additional or fewer of such elements/components. The present disclosure is also intended to cover and embrace all suitable changes in technology. Modifications which fall within the scope of the present invention will be apparent to those skilled in the art, in light of a review of this disclosure, and such modifications are intended to fall within the appended claims.