Patent classifications
G06F12/00
Enhanced network attached storage (NAS) services interfacing to cloud storage
An illustrative storage management appliance is interposed between client computing devices and one or more cloud storage resources. The appliance uses cloud storage resources in conjunction with a network attached storage device configured within the appliance to provide to the client computing devices seemingly unlimited network attached storage on respective network shares. The storage management appliance monitors data objects on the network shares and when a data object meets one or more criteria for archiving, the storage management appliance archives the data object to a cloud storage resource and replaces it with a stub and preview image on the network share. When access to the stub and/or preview image is detected, the storage management appliance restores the data object from the cloud storage resource. The criteria for archiving flexibly allow individual data objects to be archived to cloud storage without archiving frequently-accessed “neighboring” data objects on the same network share.
Transparent self-replicating page tables in computing systems
An example method of managing memory in a computer system implementing non-uniform memory access (NUMA) by a plurality of sockets each having a processor component and a memory component is described. The method includes replicating page tables for an application executing on a first socket of the plurality of sockets across each of the plurality of sockets; associating metadata for pages of the memory storing the replicated page tables in each of the plurality of sockets; and updating the replicated page tables using the metadata to locate the pages of the memory that store the replicated page tables.
Managing client devices associated with storage nodes in a scale-out storage system
Client devices associated with scale-out storage nodes can be managed based on scale-out storage nodes having backup power supplies. For example, a management node of a scale-out storage system can determine, from among a plurality of storage nodes of the scale-out system, that a first storage node is uncoupled to a backup power supply and that a second storage node is coupled to the backup power supply. The management node can receive device characteristics describing a type of workload and a configuration for a client device associated with the first storage node. The management node can determine the client device satisfies a migration policy based on the device characteristics. The management node can migrate the client device to the second storage node based on the client device satisfying the migration policy.
System and method for approximating replication completion time
One embodiment provides a computer implemented method of estimating replication completion time. The method includes creating a historical dataset of prior replication data; determining a set of replication parameters to consider; inputting the historical dataset and the set of replication parameters to a replication completion time estimator module; generating a replication completion time prediction based on the historical dataset and the set of replication parameters; and generating a confidence prediction corresponding to the replication completion time prediction.
Method, electronic device and computer program product for managing cache
Techniques for cache management involve accessing, when a first data block to be accessed is missing in a first cache, the first data block from a storage device storing the first data block; selecting, when the first cache is full and based on a plurality of parameters associated with a plurality of eviction policies, an eviction policy for evicting a data block in the first cache from the plurality of eviction policies, the plurality of parameters indicating corresponding possibilities that the plurality of eviction policies are selected; evicting a second data block in the first cache to a second cache based on the selected eviction policy, the second cache being configured to record the data block evicted from the first cache; and caching the accessed first data block in the first cache. Such techniques can improve the cache hit rate, thereby improving the access performance of a system.
Using per memory bank load caches for reducing power use in a system on a chip
A VPU and associated components include a min/max collector, automatic store predication functionality, a SIMD data path organization that allows for inter-lane sharing, a transposed load/store with stride parameter functionality, a load with permute and zero insertion functionality, hardware, logic, and memory layout functionality to allow for two point and two by two point lookups, and per memory bank load caching capabilities. In addition, decoupled accelerators are used to offload VPU processing tasks to increase throughput and performance, and a hardware sequencer is included in a DMA system to reduce programming complexity of the VPU and the DMA system. The DMA and VPU executes a VPU configuration mode that allows the VPU and DMA to operate without a processing controller for performing dynamic region based data movement operations.
Memory device with dynamic cache management
A memory system includes a memory array having a plurality of memory cells; and a controller coupled to the memory array, the controller configured to: designate a storage mode for a target set of memory cells based on valid data in a source block, wherein the target set of memory cells are configured with a capacity to store up to a maximum number of bits per cell, and the storage mode is for dynamically configuring the target set of memory cells in as cache memory that stores a number of bits less per cell than the corresponding maximum capacity.
Memory device with dynamic cache management
A memory system includes a memory array having a plurality of memory cells; and a controller coupled to the memory array, the controller configured to: designate a storage mode for a target set of memory cells based on valid data in a source block, wherein the target set of memory cells are configured with a capacity to store up to a maximum number of bits per cell, and the storage mode is for dynamically configuring the target set of memory cells in as cache memory that stores a number of bits less per cell than the corresponding maximum capacity.
Method and system for multi-pronged backup using real-time attributes
A method and system for backup processes that includes identifying a target volume and identifying a number of available threads to back up the target volume. The elements in the target volume are distributed among the available threads based on a currently pending size of data in the threads. The elements are stored from each thread into a backup container, and merged from each of the backup containers into a backup volume.
IMAGE FILE DISTRIBUTION APPARATUS, IMAGE FILE RECOVERY APPARATUS, IMAGE FILE DISTRIBUTION METHOD, IMAGE FILE RECOVERY METHOD, IMAGE FILE DISTRIBUTION PROGRAM, IMAGE FILE RECOVERY PROGRAM, AND RECORDING MEDIUM STORING PROGRAM
Provided are an image file distribution apparatus, an image file recovery apparatus, an image file distribution method, an image file recovery method, an image file distribution program, an image file recovery program, and a recording medium storing the program which can prevent a relatively large increase in the amount of data of an image file even when an (k, n) secret sharing scheme with high security is used. For example, distributed tag information is obtained from tag information of the image file by a (k, n)-threshold secret sharing scheme. For example, distributed image data is obtained from image data by a (k, L, n)-threshold ramp secret sharing scheme. For example, the distributed tag information and the distributed image data are combined to obtain combined data. Since the amount of data in the tag information is small, the use of the (k, n) secret sharing scheme does not cause a large increase in the amount of data. Since the (k, L, n)-threshold ramp secret sharing scheme does not cause a large increase in the amount of data, an increase in the total amount of data in the image data is relatively small.