Patent classifications
G06F16/183
Deduplicated data distribution techniques
In connection with a data distribution architecture, client-side “deduplication” techniques may be utilized for data transfers occurring among various file system nodes. In some examples, these deduplication techniques involve fingerprinting file system elements that are being shared and transferred, and dividing each file into separate units referred to as “blocks” or “chunks.” These separate units may be used for independently rebuilding a file from local and remote collections, storage locations, or sources. The deduplication techniques may be applied to data transfers to prevent unnecessary data transfers, and to reduce the amount of bandwidth, processing power, and memory used to synchronize and transfer data among the file system nodes. The described deduplication concepts may also be applied for purposes of efficient file replication, data transfers, and file system events occurring within and among networks and file system nodes.
Data mesh parallel file system replication
Embodiments relate to providing a multi-cloud, multi-region, parallel file system cluster service with replication between file system storage nodes. In some embodiments, a first file system storage node of a file system storage cluster receives a request from a client device to write data to a first file system stored on the first file system storage node. In response to the request to write the data to the first file system, a plurality of servers of the first file system storage node writes, in parallel, the data to the first file system and sends instructions to a second file system storage node of the file system storage cluster for writing the data to a second file system stored on the second file system storage node.
Streaming database recovery using cluster live mounts
A database can be instantly recovered by a cluster mapped to the database. Nodes of the cluster are mapped over channels to directories of the database. Scripts are generated from one or more templates that specify the order and values to be executed to perform a database job, such as database recovery. To recover the database, a template is executed that generates and populates scripts, which are processed on the host of the database to recover the database in a nearly instant manner without transferring data files.
Blockchain-based dynamic storage provisioner
Methods and systems for dynamically provisioning storage on a blockchain are provided. In one embodiment, a method is provided that includes receiving a request that includes a data unit for storage. The data unit may be buffered in a data stack that stores one or more data units. The data unit may be stored in the data stack until (i) a predetermined amount of time has passed and/or (ii) a size of the data stack exceeds a predetermined threshold. The data units stored in the data stack may then be encrypted and included in one or more storage transactions. The storage transactions may also include encryption keys used to encrypt the data units. The storage transactions may then be transmitted to nodes for storage on a blockchain.
Method For Applying Local Volume File System Policies(encryption, compression and etc) on Cloud(HTTP(S) Server) Files
The user can apply the local volume file system policies like encryption/decryption, file, backup, antivirus, file compression/decompression, file monitoring and etc to the cloud (HTTPS server) files:
The cloud (HTTPS server) files are not vulnerable to man in the middle attack when they are on the way to the cloud (HTTPS server) from the local computer and vice versa, since the files are secured by local volume file system policies (encryption/decryption) in the local computer before they are stored in the cloud (HTTPS server).
The user need not to rely on the cloud (HTTPS server) for the security of the cloud (HTTPS server) files, since the: files are secured by local volume file system policies (encryption/decryption) in the local computer before they are stored in the cloud (HTTPS Server).
File system warnings application programing interface (API)
The present technology pertains to a organization directory hosted by a synchronized content management system. The corporate directory can provide access to user accounts for all members of the organization to all content items in the organization directory on the respective file systems of the members' client devices. Members can reach any content item at the same path as other members relative to the organization directory root on their respective client device. In some embodiments novel access permissions are granted to maintain path consistency.
Technologies for integrating cloud content items across platforms
An example method can include storing, on a CSM, a first content item and representations of second and third content items, the second content item having content/features enabled by a cloud service and designed for access through a native online application and the third content item having content/features supported by a local application and having additional features designed for access through a cloud service and native online application; when the first content item is invoked, presenting the content/features of the first content item; in response to a request to access the representation of the second or third content item, sending, to a cloud service, a request for the additional features of the third content item or the content/features of the second content item; and based on metadata received from a cloud service, providing the additional features/content of the third content item or the content/features of the second content item.
DEDUPLICATED DATA DISTRIBUTION TECHNIQUES
In connection with a data distribution architecture, client-side “deduplication” techniques may be utilized for data transfers occurring among various file system nodes. In some examples, these deduplication techniques involve fingerprinting file system elements that are being shared and transferred, and dividing each file into separate units referred to as “blocks” or “chunks.” These separate units may be used for independently rebuilding a file from local and remote collections, storage locations, or sources. The deduplication techniques may be applied to data transfers to prevent unnecessary data transfers, and to reduce the amount of bandwidth, processing power, and memory used to synchronize and transfer data among the file system nodes. The described deduplication concepts may also be applied for purposes of efficient file replication, data transfers, and file system events occurring within and among networks and file system nodes.
Electronic file management
Described herein are methods and system for electronic file management having a central server that periodically scans files accessible to multiple computers to identify every file stored onto multiple electronic data repositories. The central server then executes a predetermined protocol to generate a unique identifier for each identified file. The central server then generates an interconnected nodal data structure computer model where each node represents an identified file and where the nodes are linked based on their respective files having similar unique identifiers. The central server periodically scans the electronic data repositories to identify related data. When a related file is identified, the central server modifies the nodal data structure accordingly. When a user requests access to a file, the central server displays all related data to the requested file.
Access pattern driven data placement in cloud storage
A system and method for storing data in a distributed network having a plurality of datacenters distributed over a plurality of geographic regions. The method may involve receiving data, including metadata, uploaded to a first datacenter of the distributed network, receiving access information about previous data that was previously stored in the plurality of datacenters of the distributed network, predicting one or more of the plurality of geographic regions from which the uploaded data will be accessed based on the metadata and the access information, and instructing the uploaded data to be transferred from the first datacenter to one or more second datacenters located at each of the one or more predicted geographic regions.