H03M7/4037

UPDATING PREFIX CODES FOR PSEUDO-DYNAMIC DATA COMPRESSION

In a data storage system, a prior set S of prefix codes for pseudo-dynamic compression as well as data compressed utilizing prior set S are stored. While data compressed utilizing prior set S are stored in the data storage system, the number of prefix codes utilized by the data storage system for pseudo-dynamic compression are augmented. Augmenting the number of codes includes determining a new set S′ of prefix codes for pseudo-dynamic compression from a training data set selected from a workload of the data storage system and storing the new set S′ in the data storage system with the prior set S.

DEVICE AND METHOD FOR REDUCING FEEDBACK OVERHEAD ASSOCIATED WITH BITMAP REPORTING

A method for reducing feedback overhead associated with bitmap reporting may be implemented between a user equipment and a base station. The method includes activating a coding scheme for reporting a bitmap in association with a prefix coding scheme, encoding a plurality of bit groups using the prefix coding scheme, generating a plurality of codeword sets for the plurality of bit groups, and reporting the codeword sets generated.

METHOD AND APPARATUS FOR SPECULATIVE DECOMPRESSION

An apparatus and method for performing parallel decoding of prefix codes such as Huffman codes. For example, one embodiment of an apparatus comprises: a first decompression module to perform a non-speculative decompression of a first portion of a prefix code payload comprising a first plurality of symbols; and a second decompression module to perform speculative decompression of a second portion of the prefix code payload comprising a second plurality of symbols concurrently with the non-speculative decompression performed by the first compression module.

SHARED DECOMPRESSION ENGINE
20170235752 · 2017-08-17 · ·

A method for sharing a hardware decompression engine, including performing a compression type check on a first data stream to determine a compression type of the first data stream, wherein the first data stream is compressed using one selected from a group consisting of a first compression type and a second compression type; wherein, when the first data stream is compressed with the second compression type: receiving the second compression type at a selector; converting the first data stream compressed with the second compression type into a second data stream of the first compression type; inputting the converted second data stream into the selector; and decompressing the converted second data stream using the hardware decompression engine capable of decompressing a data stream compressed using the first compression type. In other aspects, a system for sharing a hardware decompression engine and a computing system are provided.

Ultra-high speed optical transport employing LDPC-coded modulation with non-uniform signaling
09735921 · 2017-08-15 · ·

A low-density parity-check (LDPC) coded bit-interleaved coded modulation with iterative decoding (BICM-ID) scheme with nonuniform signaling which is effected by mapping simple variable-length prefix codes onto the constellation. By employing Huffman procedure(s), prefix codes can be designed to approach optimal performance. Experimental evaluations of the schemes demonstrate that the nonuniform scheme performs better than 8-QAM by at least 8.8 dB.

Multi-symbol, multi-format, parallel symbol decoder for hardware decompression engines

In some data compression algorithms and/or standards, the compressed data comprises variable length symbols. A set of parallel decoders speculatively decode/decompress a window (i.e., sub-block) of data. Each of the decoders attempts to decode/decompress a symbol that starts at a different location in the compressed data block. Once the decoders have finished decoding a symbol (or determined that a valid symbol does not begin at the beginning of the window assigned to that decoder), a symbol strider selects the decoder outputs corresponding to valid symbols. The symbol strider successively selects decoder outputs based on the size of the previous symbols that were found to be valid. When the next valid symbol begins outside the current window, its location is stored to indicate the location of the next valid symbol in a subsequent window.

REAL-TIME HISTORY-BASED BYTE STREAM COMPRESSION
20220231699 · 2022-07-21 ·

Described embodiments provide systems and methods for stream-based compression. An encoder of a first device receives an input stream of bytes including a first byte preceded by one or more second bytes. The encoder may determine to identify a prefix code for the first byte. The encoder may select a prefix code table using the one or more second bytes. The encoder may identify, from the selected prefix code table, the prefix code of the first byte. The encoder may generate an output stream of bytes by replacing the first byte in the input stream with the prefix code of the first byte. The encoder may transmit the output stream from the encoder of the first device to a decoder of a second device. The output stream may have a fewer number of bits than the input stream.

Parallelized decoding of variable-length prefix codes

Methods and systems are provided for decoding variable-length codes in a parallel process. A stream of variable-length code words is divided into fixed length words. A plurality of parallel sets of decoder circuits each receive, in parallel, a current fixed length word and a prior fixed length word. Each decoder circuit has a respective fixed leftover bit-count. Each decoder circuit generates a respective output that may include a decoded symbol and a new leftover bit-count. Each respective output is determined based on the respective current fixed length word, the respective prior fixed length word, and the respective fixed leftover bit-count. A set of selected decoder circuit outputs is generated for each set of the parallel sets of decoder circuits based on a set of first leftover bit-counts. One output from each set of selected decoder circuit outputs is selected as a final output based on a second prior leftover bit-count.

METHOD AND APPARATUS FOR EFFICIENT DEFLATE DECOMPRESSION USING CONTENT-ADDRESSABLE DATA STRUCTURES
20220200623 · 2022-06-23 ·

Apparatus and method for efficient compression block decoding using content-addressable structure for header processing. For example, one embodiment of an apparatus comprises: a header parser to extract a sequence of tokens and corresponding length values from a header of a compression block, the tokens and corresponding length values associated with a type of compression used to compress a payload of the compression block; and a content-addressable data structure builder to construct a content-addressable data structure based on the tokens and length values, the content-addressable data structure builder to write an entry in the content-addressable data structure comprising a length value and a count value, the count value indicating a number of times the length value was previously written to an entry in the content-addressable data structure.

Lossless reduction of data by using a prime data sieve and performing multidimensional search and content-associative retrieval on data that has been losslessly reduced using a prime data sieve
11363296 · 2022-06-14 · ·

Input data can be losslessly reduced by using a data structure that organizes prime data elements based on their contents. Alternatively, the data structure can organize prime data elements based on the contents of a name that is derived from the prime data elements. Specifically, video data can be losslessly reduced by (1) using the data structure to identify a set of prime data elements, and (2) using the set of prime data elements to losslessly reduce intra-frames. The input data can be dynamically partitioned based on the memory usage of components of the data structure. Parcels can be created based on the partitions to facilitate archiving and movement of the data. The losslessly reduced data can be stored using a set of distilled files and a set of prime data element files.