Patent classifications
G06F16/137
DISTRIBUTED DATA SET ENCRYPTION AND DECRYPTION
An apparatus includes a processor component of a first node device caused to receive data block encryption data and an indication of size of an encrypted data block distributed to the first node device for decryption, and in response to the data set being of encrypted data: receive an indication of the quantity of sub-blocks within the encrypted data block, and a hashed identifier for each data sub-block; use the data block encryption data to decrypt the encrypted data block to regenerate data set portions from the data sub-blocks; analyze the hashed identifier of each data sub-block to determine whether all data set portions are distributed to the first node device for processing; and in response to a determination that at least one data set portion is to be distributed to a second node device for processing, transmit the at least one data set portion to the second node device.
PREVENTING A TRANSMISSION OF AN INCORRECT COPY OF A RECORD OF DATA TO A DISTRIBUTED LEDGER SYSTEM
An incorrect copy of a record of data can be prevented from being transmitted to a distributed ledger system. A first file can be received and can include information, in audio or video form, with a description of a subject matter of the record of data and with an authorization to transmit the copy to the distributed ledger system. The first file can be sent to a device. A second file can be received from the device and can include information that confirms that the description of the subject matter, included in the first file, is correct, and that confirms that an entity, which controlled production of the first file, has permission to authorize causing the copy to be transmitted to the distributed ledger system. The correct copy can be caused, based on a receipt of the first and the second files, to be transmitted to the distributed ledger system.
Remote access service inspector
A method, system, and computer program product for providing protected remote access from a remote access client to a remote access server over a computer network through a plurality of inspections. A remote access configuration file is created for the remote access client. A digital hash of the configuration file is then generated. The digital hash is compared with a configuration file stored at a predefined web location. If the comparison results in a match between the digital hash and the stored configuration file, a digital hash comparison is performed between an encrypted remote access configuration file and an encrypted configuration file stored at the predefined web location. If the plurality of inspections are passed, the remote access client is released from a quarantine state and a virtual private network (VPN) connection to the remote access server is established.
Method, device and computer program product for information processing
According to embodiments of the present disclosure, a method, device and computer program product for information processing are proposed. The method comprises: obtaining identification information of a shard of metadata at a first node of a blockchain-based metadata management system; determining, based on similarities of the identification information of the shard and identification information of candidate nodes of the metadata management system, a second node for positioning the shard from the candidate nodes; and enabling the second node to process the identification information of the shard, to manage storage of the shard in the metadata management system. Therefore, the present solution can improve efficiency, security and robustness of the metadata management system.
DETERMINING SHARED NODES BETWEEN SNAPSHOTS USING PROBABILISTIC DATA STRUCTURES
The present disclosure is related to methods, systems, and machine-readable media for determining shared nodes between snapshots using probabilistic data structures. A unique identifier can be assigned to each node of a first tree data structure corresponding to a first snapshot of a virtual computing instance (VCI). A first probabilistic data structure representing the first tree data structure can be created that includes hashes of the identifiers assigned to the nodes of the first tree data structure. A unique identifier can be assigned to each node of a second tree data structure corresponding to a second snapshot of the VCI. A second probabilistic data structure representing the second tree data structure can be created that includes hashes of the identifiers assigned to the nodes of the second tree data structure. A particular node of the second tree data structure can be determined to be shared by the first tree data structure responsive to a determination that the first probabilistic data structure includes a hash of an identifier assigned to the particular node.
Information source agent systems and methods for distributed data storage and management using content signatures
Information source agent systems and methods for distributed content storage and management using content signatures that use file identicality properties are provided. A data management system is provided that includes a content engine for managing the storage of file content, a content signature generator that generates a unique content signature for a file processed by the content engine, a content signature comparator that compares content signatures and a content signature repository that stores content signatures. Information source agents are provided that include content signature generators and content signature comparators. Methods are provided for the efficient management of files using content signatures that take advantage of file identicality properties. Content signature application modules and registries exist within information source clients and centralized servers to support the content signature methods.
CROSS-SILO DATA STORAGE AND DEDUPLICATION
In some aspects, a computing system may generate a content-defined tree. A content-defined tree may be a tree of cryptographic hashes where each leaf is a hash of a chunk (e.g., data chunk) of a data object, and each parent node (e.g., interior node) is the hash of a concatenation of the hashes of the parent's children nodes. To create parent nodes for the leaf nodes, a computing system may group leaf nodes together based on a rolling hash (e.g., a rolling hash of the hashes of the leaf nodes) satisfying a condition. Each parent node may include a hash that represents the concatenation of the hashes of the leaf nodes that fall under the corresponding parent node.
Sharing unmanaged content using a content management system
In some implementations, a subscriber to an online content management system can share content items that are external to the subscriber's content library. A computing device can include managed content items associated with the subscriber's content library. The computing device can include unmanaged content items that are stored externally to the subscriber's content library. The subscriber can provide input associated with an unmanaged content item to generate a link (e.g., URL, file path, location reference, etc.) for the unmanaged content item. When generating the link, the unmanaged content item can be uploaded to the online content management system and stored separately from the subscriber's content library. The generated link can be shared with recipient user (e.g., subscriber or non-subscriber). The generated link can be used by the recipient to access the unmanaged content item uploaded to and stored by the online content management system.
Efficient filename storage and retrieval
The disclosed technology relates to a system configured to detect a modification to a node in a tree data structure. The node is associated with a content item managed by a content management service as well as a filename. The system may append the filename and a separator to a filename array, determine a location of the filename in the filename array, and store the location of the filename in the node.
SYSTEMS AND METHODS FOR PERFORMANT DATA MATCHING
The present disclosure is directed to systems and methods for performant data matching. Entities maintain large amounts of data and desire to reconcile duplicative records. One way to solve this problem is through data matching. However, standard data matching at the record level can be laborious and inefficient. To remedy these inefficiencies in data matching, the present disclosure describes a system where the token records are tokenized a second time into token sets based on the token records satisfying at least one token set rule. A token set rule may be based on the common presence of multiple tokens in a token record. If multiple token records have the required tokens from the set rule, then those token records can be hashed and rolled-up into the token set (i.e., tokenized a second time into the token set). The token set allows for more efficient data matching.