G06F16/1744

POINT CLOUD DATA ENCAPSULATION METHOD AND POINT CLOUD DATA TRANSMISSION METHOD
20230048715 · 2023-02-16 · ·

Provided are a point cloud data encapsulation method and a point cloud data transmission method. The point cloud data encapsulation method includes: compressing a point cloud media file; and encapsulating the compressed point cloud media file and setting a classification record according to a data type, where the classification record is used for indicating a data type of each portion of point cloud data in the point cloud media file. The problem of low transmission efficiency of the media content and waste of storage space in the point cloud data transmission is solved, the flexibility of the media content combination is increased, management control of a multimedia service is increased, users are given more rights to independently select and acquire the media content, diversified application scenarios are provided, and better multimedia content provision and transmission solutions are provided for network operators.

Policy enforcement and performance monitoring at sub-LUN granularity
11579910 · 2023-02-14 · ·

Techniques are provided for enforcing policies at a sub-logical unit number (LUN) granularity, such as at a virtual disk or virtual machine granularity. A block range of a virtual disk of a virtual machine stored within a LUN is identified. A quality of service policy object is assigned to the block range to create a quality of service workload object. A target block range targeted by an operation is identified. A quality of service policy of the quality of service policy object is enforced upon the operation using the quality of service workload object based upon the target block range being within the block range of the virtual disk.

Technologies for providing shared memory for accelerator sleds

Technologies for providing shared memory for accelerator sleds includes an accelerator sled to receive, with a memory controller, a memory access request from an accelerator device to access a region of memory. The request is to identify the region of memory with a logical address. Additionally, the accelerator sled is to determine from a map of logical addresses and associated physical address, the physical address associated with the region of memory. In addition, the accelerator sled is to route the memory access request to a memory device associated with the determined physical address.

Real-time archiving method and system based on hybrid cloud
11556497 · 2023-01-17 · ·

Provided are a data archiving method and apparatus capable of providing a remote near-line data archiving function by receiving remote function invoking from a target system in which data is stored, providing the target system with a first function for archiving, in a storage system, at least some of the data stored in the target system over a network in response to the remote function invoking, and providing the target system with a second function for the query of the data archived in the storage system over the network.

Systems and methods for managing single instancing data

Described in detail herein are systems and methods for managing single instancing data. Using a single instance database and other constructs (e.g. sparse files), data density on archival media (e.g. magnetic tape) is improved, and the number of files per storage operation is reduced. According to one aspect of a method for managing single instancing data, for each storage operation, a chunk folder is created on a storage device that stores single instancing data. The chunk folder contains three files: 1) a file that contains data objects that have been single instanced; 2) a file that contains data objects that have not been eligible for single instancing; and 3) a metadata file used to track the location of data objects within the other files. A second storage operation subsequent to a first storage operation contains references to data objects in the chunk folder created by the first storage operation instead of the data objects themselves.

METHOD AND SYSTEM FOR COMPRESSING GENOME SEQUENCES USING GRAPHIC PROCESSING UNITS
20180011870 · 2018-01-11 ·

The present invention provides a method for compressing genome sequences readers using GPU processing unit. The method comprising the steps of: identifying position of each given genome reader characters string in the sequence of a reference genome, determining alignment of each reader string within the reference genome, comparing each reader characters string to corresponding reference genome sequence based on determined alignment, filtering characters in each reader by GPU processor by eliminating similar characters and extracting only characters differences in association to their position in the genome sequence and recording filtered data of each reader in association to its alignment in genome reference at the genome compressed database.

HIGH-DENSITY COMPRESSION METHOD AND COMPUTING SYSTEM

Certain implementations of the disclosed technology may include methods and computing systems for performing high-density data compression, particularly on numerical data that demonstrates various patterns, and patterns of patters. According to an example implementation, a method is provided. The method may include extracting a data sample from a data set, compressing the data sample using a first compression filter configuration, and calculating a compression ratio associated with the first compression filter configuration. The method may also include compressing the data sample using a second compression filter configuration and calculating a compression ratio associated with the second compression filter configuration. A particular compression filter configuration to utilize in compressing the entire data set may be selected based on a comparison of the compression ratio associated with the first compression filter configuration and a compression ratio associated with the second compression filter configuration.

Storing compression units in relational tables

A database server stores compressed units in data blocks of a database. A table (or data from a plurality of rows thereof) is first compressed into a “compression unit” using any of a wide variety of compression techniques. The compression unit is then stored in one or more data block rows across one or more data blocks. As a result, a single data block row may comprise compressed data for a plurality of table rows, as encoded within the compression unit. Storage of compression units in data blocks maintains compatibility with existing data block-based databases, thus allowing the use of compression units in preexisting databases without modification to the underlying format of the database. The compression units may, for example, co-exist with uncompressed tables. Various techniques allow a database server to optimize access to data in the compression unit, so that the compression is virtually transparent to the user.

Reduction of data stored on a block processing storage system
11567901 · 2023-01-31 · ·

Techniques and systems for reducing data stored on a block processing storage system are described. A losslessly reduced representation of a data block can include references to one or more prime data element blocks, and optionally a description of a reconstitution program which, when applied to the one or more prime data element blocks, results in the data block.

Scaling delta table optimize command

The interface is to receive an indication to execute an optimize command. The processor is to receive a file name; determine whether adding a file of the file name to a current bin causes the current bin to exceed a threshold; associate the file with the current bin in response to determining that adding the file does not cause the current bin to exceed the bin threshold; in response to determining that adding the file to the current bin causes the current bin to exceed the bin threshold: associate the file with a next bin, indicate that the current bin is closed, and add the current bin to a batch of bins; determine whether a measure of the batch of bins exceeds a batch threshold; and in response to determining that the measure exceeds the batch threshold, provide the batch of bins for processing.