Patent classifications
G06F11/14
Update of programmable for computing nodes
For computing nodes having a first programmable and comprising a first node and a second node, an update of the first node from the first programmable to a second programmable across an external network is initiated. In response to the update being interrupted, the first programmable is automatically reinstated on the first node by retrieving the first programmable from the second node across an internal network. The second node is automatically updated to the second programmable by retrieving the second programmable from the first node across the internal network in response to completion of the update of the first node to the second programmable.
Reporting control information errors
Methods, systems, and devices for reporting control information errors are described. A state of a memory array may be monitored during operation. After detecting an error (e.g., in received control information), the memory device may enter a first state (e.g., a locked state) and may indicate to a host device that an error was detected, the state of the memory array before the error was detected, and/or at least a portion of a control signal carrying the received control information. The host device may diagnose a cause of the error based on receiving the indication of the error and/or the copy of the control signal. After identifying and/or resolving the cause of the error, the host device may transmit one or more commands (e.g., unlocking the memory device and returning the memory array to the original state) based on receiving the original state from the memory device.
Method and system for expanding storage capacity using cloud storage systems
In some embodiments, a computer system may buffer, in a local storage buffer, data to be stored at the cloud and upload the data to the cloud storage from the local storage buffer. Upon receipt of an acknowledgement, from the cloud storage, that the data was successfully uploaded, the computer system may delete the data from the local storage buffer. The computer system may provide a ghost file representation of the uploaded data through the computer system, through which the data uploaded to the cloud storage may be accessed. In some embodiments, the computer system may access multiple cloud storage accounts, each corresponding to a file folder on the computer system, allowing the user to easily access different cloud storage accounts. The computer system may recover data that was deleted from either local persistent storage or the cloud storage.
Data masking in a microservice architecture
A method includes retrieving an object from storage and copying the object, generating a list that identifies one or more byte ranges, of the copy of the object, to be masked, providing the list to a masker controller microservice that examines a recipe corresponding to the copy of the object, where the recipe references a slice of the copy of the object, and the slice includes one or more data segments, masking, by the masker controller microservice, a segment of the slice that is in one of the byte ranges, to create a masked segment, and replacing, in the slice, the segment with the masked segment, to create a masked slice and creating a masked object recipe that contains a reference to the masked slice, creating a masked object that includes the masked slice, and that references any unmasked segments of the slice, and deduplicating the masked object.
Enhanced retry count for uplink multi-user transmission
This disclosure describes systems, methods, and devices related to an enhanced retry count for an uplink (UL) multi-user (MU) transmission. A device may identify a trigger frame received from a first device on a wireless communication channel. The device may determine a quality of service counter associated with an access category. The device may cause to send a frame to the first device based at least in part on the trigger frame. The device may determine an error condition associated with the frame. The device may refrain from incrementing the quality of service counter based on the error condition.
Method of ensuring confidentiality and integrity of stored data and metadata in an untrusted environment
A system and method for storing and recovering a computer file. The method includes calculating fingerprint data of the file, separating the file into a plurality of data sub-files each having the same size and a single data sub-file having a smaller size than the other data sub-files, and attaching file metadata to the single data sub-file or as a metadata sub-file. The method also includes padding the single data sub-file including the metadata so that it is the same size as the plurality of data sub-files or the metadata sub-file so that it is the same size as the plurality of data sub-files, adding a header to each data sub-file that includes information about the sub-file, assigning a unique filename to each data sub-file, encrypting each data sub-file, and storing each data sub-file as separate files under their unique filename.
Garbage collection for a deduplicated cloud tier using functions
Systems and methods for performing data protection operations including garbage collection operations and copy forward operations. For deduplicated data stored in a cloud-based storage or in a cloud tier that stores containers containing dead and live segments or dead and live regions such as compression regions, the dead compression regions are deleted by copying the live compression regions into new containers and then deleting the old containers. The copy forward is based on a recipe from a data protection system and is performed using a serverless approach.
Automatic storage target recommendations using a storage classifier
Embodiments for a storage classifier that provides recommendations to a backup server for storage targets among a plurality of disparate target storage types. The storage classifier receives metadata (name, type, size), and the Service Level Agreement with information such as: retention time, Recovery Point Objective, and Recovery Time Objective) from the backup software. The backup software itself receives policy recommendations from a data label rules engine based on certain file attributes. The storage classifier receives an initial recommendation for the storage type and location (e.g., on-premises deduplication storage or public-cloud object storage, etc.) from a data classifier. Based on these inputs, the storage classifier provides recommended specific storage targets to the backup software on a file-by-file basis for data stored in a backup operation.
Method and apparatus for stress management in a searchable data service
Method and apparatus for stress management in a searchable data service. The searchable data service may provide a searchable index to a backend data store, and an interface to build and query the searchable index, that enables client applications to search for and retrieve locators for stored entities in the backend data store. Embodiments of the searchable data service may implement a distributed stress management mechanism that may provide functionality including, but not limited to, the automated monitoring of critical resources, analysis of resource usage, and decisions on and performance of actions to keep resource usage within comfort zones. In one embodiment, in response to usage of a particular resource being detected as out of the comfort zone on a node, an action may be performed to transfer at least part of the resource usage for the local resource to another node that provides a similar resource.
FILE PROCESSING DEVICE, FILE PROCESSING METHOD, AND PROGRAM
The present technology relates to a file processing device, a file processing method, and a program capable of reducing the number of images that the user loses. Restoration data for restoring a high efficiency image file format (HEIF) file storing a still image is written into the HEIF file. The HEIF file is restored using the restoration data. The present technology can be applied to processing of a HEIF file.