H03M7/3071

Sample array coding for low-delay

The entropy coding of a current part of a predetermined entropy slice is based on, not only, the respective probability estimations of the predetermined entropy slice as adapted using the previously coded part of the predetermined entropy slice, but also probability estimations as used in the entropy coding of a spatially neighboring, in entropy slice order preceding entropy slice at a neighboring part thereof. Thereby, the probability estimations used in entropy coding are adapted to the actual symbol statistics more closely, thereby lowering the coding efficiency decrease normally caused by lower-delay concepts. Temporal interrelationships are exploited additionally or alternatively.

Methods and systems for combined lossless and lossy coding
11706410 · 2023-07-18 · ·

A decoder includes circuitry configured to receive a bitstream identify, in the bitstream, a current frame, wherein the current frame includes a first region and a third region, detect, in the bitstream, an indication that the first region is encoded according to a lossless encoding protocol, and decode the current frame, wherein decoding the current frame further comprises decoding the first region using a lossless decoding protocol corresponding to the lossless encoding protocol.

Method and device for transmitting/receiving signal in wireless communication system

A radio frequency (RF) unit, a digital unit, and methods of transmitting and receiving data in a wireless communication system are provided. The digital unit may include: a transceiver configured to receive compressed data from an RF unit, a processor configured to divide a frequency domain and a time domain into a plurality of blocks, set a compression parameter to be applied to each of the plurality of blocks, and expand the received data in units of the blocks based on the set compression parameter, and a memory storing the compression parameter.

TECHNOLOGY FOR EARLY ABORT OF COMPRESSION ACCELERATION

An integrated circuit includes a compression accelerator to process a request from software to compress source data into an output file. The compression accelerator includes early-abort circuitry to provide for early abort of compression operations. In particular, the compression accelerator uses a predetermined sample size to compute an estimated size for a portion of the output file. The sample size specifies how much of the source data is to be analyzed before computing the estimated size. The compression accelerator also determines whether the estimated size reflects an acceptable amount of compression, based on a predetermined early-abort threshold. The compression accelerator aborts the request if the estimated size does not reflect the acceptable amount of compression. The compression accelerator may complete the request if the estimated size reflects the acceptable amount of compression. Other embodiments are described and claimed.

TEXT COMPRESSION WITH PREDICTED CONTINUATIONS

A method for text compression comprises recognizing a prefix string of one or more text characters preceding a target string of a plurality of text characters to be compressed. The prefix string is provided to a natural language generation (NLG) model configured to output one or more predicted continuations each having an associated rank. If the one or more predicted continuations include a matching predicted continuation relative to the next one or more text characters of the target string, the next one or more text characters are compressed as an NLG-type compressed representation. If no predicted continuations match the next one or more text characters of the target string, a longest matching entry in a compression dictionary is identified. The next one or more text characters of the target string are compressed as a dictionary-type compressed representation that includes the dictionary index value of the longest matching entry.

TECHNIQUES TO ENABLE STATEFUL DECOMPRESSION ON HARDWARE DECOMPRESSION ACCELERATION ENGINES
20220405142 · 2022-12-22 ·

A hardware decompression acceleration engine including: an input buffer for receiving to-be-decompressed data from a software layer of a host computer; a decompression processing unit coupled to the input buffer for decompressing the to-be-decompressed data, the decompression processing unit further receiving first and second flags from the software layer of the host computer, wherein the first flag is indicative of a location of the to-be-decompressed data in a to-be-decompressed data block and the second flag is indicative of a presence of an intermediate state; and an output buffer for storing decompressed data from the decompression processing unit.

DETERMINING COMPRESSION LEVELS TO APPLY FOR DIFFERENT LOGICAL CHUNKS OF COLLECTED SYSTEM STATE INFORMATION

An apparatus comprises a processing device configured to collect system state information from host devices, to split the collected system state information into logical chunks, and to determine, based at least in part on a plurality of factors, a compression level to be applied to each of the logical chunks. The plurality of factors comprise a first factor characterizing a time at which the collected system state information is needed at a destination device and at least a second factor characterizing resources available for at least one of performing compression of the collected system state information and transmitting the collected system state information over at least one network to the destination device. The processing device is further configured to apply the determined compression level to each of the logical chunks to generate compressed logical chunks, and to transmit the compressed logical chunks to the destination device.

CONTENT-BASED DYNAMIC HYBRID DATA COMPRESSION
20220382717 · 2022-12-01 ·

An information handling system includes a processor configured to process a training data file to determine an optimal data compression algorithm. The processor may also perform a compression ratio analysis that includes compressing the training data file using data compression algorithms, calculating a compression ratio associated with each of the data compression algorithms, determining an optimal compression ratio from the compression ratio associated with the each data compression algorithm; and determining a desirable data compression algorithm associated with the training data file based on the optimal compression ratio. The processor may also perform a probability analysis that includes generating a symbol transition matrix based on the desirable data compression algorithm, extracting statistical feature data based on the symbol transition matrix, and generating probability matrices based on the statistical feature data to determine the optimal data compression algorithm for each segment of a working data file.

Bandwidth compression for neural network systems

Techniques and systems are provided for compressing data in a neural network. For example, output data can be obtained from a node of the neural network. Re-arranged output data having a re-arranged scanning pattern can be generated. The re-arranged output data can be generated by re-arranging the output data into the re-arranged scanning pattern. One or more residual values can be determined for the re-arranged output data by applying a prediction mode to the re-arranged output data. The one or more residual values can then be compressed using a coding mode.

Compressed versions of image data based on relationships of data

Methods of image compression are described. A stream of color image data is filtered with a prediction routine using a pixel neighborhood. The filtered stream of color image data is sorted with a block sorting routing. A version of the color image data is compressed based on the sorted and filtered stream of color image data.