Patent classifications
G06F3/0646
Provisioning of clustered containerized applications
Clustered containerized applications are implemented with scalable provisioning. Methods include receiving a data storage request to store one or more data values in a storage volume implemented across a storage node cluster, the storage node cluster including a plurality of storage nodes including one or more storage devices having storage space allocated for storing data associated with the storage volume. Methods may further include identifying a cluster hierarchy associated with the storage node cluster, the cluster hierarchy identifying storage characteristics of the plurality of storage nodes, the cluster hierarchy also identifying physical location information for the plurality of storage nodes, the physical location information indicating node-to-node proximity on a network graph. Methods may also include selecting a storage node on which to store the data, the selecting being based, at least in part, on the identified storage characteristics and one or more data distribution parameters associated with the storage volume.
PRESENTATION OF DIRECT ACCESSED STORAGE UNDER A LOGICAL DRIVE MODEL
In accordance with embodiments disclosed herein, there are provided methods, systems, mechanisms, techniques, and apparatuses for presentation of direct accessed storage under a logical drive model; for implementing a distributed architecture for cooperative NVM Data protection; data mirroring for consistent SSD latency; for boosting a controller's performance and RAS with DIF support via concurrent RAID processing; for implementing arbitration and resource schemes of a doorbell mechanism, including doorbell arbitration for fairness and prevention of attack congestion; and for implementing multiple interrupt generation using a messaging unit and NTB in a controller through use of an interrupt coalescing scheme.
Secure, distributed raid storage systems and methods
Systems and methods for securely and remotely storing data in a remote, distributed redundant array of independent drives (RAID) is provided. RAID storage is accomplished through a series of mapped drives, non-routable Internet protocol (IP) addresses, and routable IP addresses. In addition, authorization to access a RAID controller, network address translation (NAT) system, and domain name system (DNS) system may all be separated, increasing security and allowing storage to be securely distributed among a variety of dispersed storage locations.
Compression device and control method
According to one embodiment, a compression device includes a first storage unit, a second storage unit, a calculation unit, and a comparison unit. The first storage unit stores addresses associated with hash values, respectively. The second storage unit includes storage areas specified by the addresses, respectively. The calculation unit determines a hash function to be used for first data in accordance with at least a part of the first data, and calculates a hash value using the hash function and at least a part of second data included in the first data. The comparison unit acquires third data from a storage area in the second storage unit specified by a first address, and compares the second data with the third data. The first address is stored in the first storage unit and is associated with the hash value.
Prevent counter wrap during update-side grace-period-request processing in tree-SRCU implementations
In an SRCU environment, per-processor data structures each maintain a list of SRCU callbacks enqueued by SRCU updaters. An SRCU management data structure maintains a current-grace-period counter that tracks a current SRCU grace period, and a future-grace-period counter that tracks a farthest-in-the-future SRCU grace period needed by the SRCU callbacks enqueued by the SRCU updaters. A combining tree is used to mediate a plurality of grace-period-start requests concurrently vying for an opportunity to update the future-grace-period record on behalf of SRCU callbacks. The current-grace-period counter is prevented from wrapping during some or all of the grace-period-start request processing. In an embodiment, the counter wrapping is prevented by performing some or all of the grace-period start-request processing within an SRCU read-side critical section.
Partition and access switching in distributed storage systems
Devices, computer-readable media, and methods for reducing the number of hops that internal messages must traverse in data center switching architectures are disclosed. In one example, a data center includes a first rack housing a first server, a first computational process associated to a first storage drive hosted on the first server and residing within a first level of the distributed storage system, a second rack housing a second server, a second computational process associated to a second storage drive hosted on the second server and residing within the first level of the distributed storage system, and a first switch communicatively coupled to the first level to receive messages directly from the first computational process and the second computational process.
MEMORY SYSTEM, OPERATING METHOD THEREOF AND COMPUTING SYSTEM FOR CLASSIFYING DATA ACCORDING TO READ AND WRITE COUNTS AND STORING THE CLASSIFIED DATA IN A PLURALITY OF TYPES OF MEMORY DEVICES
A memory system includes a plurality of memory devices; and a controller. The controller includes a counter configured to determine a R/W ratio, a ratio of a number of read operations to a number of write operations of target data; a selector configured to compare the R/W ratio with a first threshold corresponding to the first memory device, and select the first memory device for storing the target data when the R/W ratio is greater than or equal to the first threshold; and a processor configured to store the target data in the first memory device.
MEMORY-FABRIC-BASED PROCESSOR CONTEXT SWITCHING SYSTEM
A memory-fabric-based processor context switching system includes server devices coupled to a memory fabric. A first processing system in a first server device receives a request to move a process it is executing and, in response, copies first processing system context values to its first local memory system in the first server device, and generates a first data mover instruction that causes a first data mover device in the first server device to transmit the first processing system context values from the first local memory system to the memory fabric. A second processing system in a second server device generates a second data mover instruction that causes a second data mover device in the second server device to retrieve the first processing system context values from the memory fabric and provide the first processing system context values in a second local memory system included in the second server device.
Point-in-time snap copy on asynchronous consistency group management
In one embodiment, a local copy target is also a local primary of an incomplete consistency group of an ongoing asynchronous mirror relationship. Completion of the consistency group is facilitated notwithstanding that the local copy operation was initiated after the consistency group was initiated. In one aspect, asynchronous data mirroring logic, prior to the overwriting of existing data of the target, reads the existing data of the target for purposes of mirroring the read data to a remote secondary target of the consistency group. As a result, existing data of the local copy target which is also a local primary source of the consistency group, may be safely overwritten. Other features and aspects may be realized, depending upon the particular application.
INTRA-DEVICE NOTATIONAL DATA MOVEMENT SYSTEM
An intra-device notational data movement system has a chassis including processing system(s) that are configured to provide a first thread and a second thread. A data mover subsystem in the chassis is coupled to the processing system(s). In a communication transmitted by the first thread, the data mover subsystem identifies a request to transfer data to the second thread that is stored in a first portion of a memory system that is associated with the first thread in a memory fabric management database. The data mover subsystem then modifies notational reference information in the memory fabric management database to disassociate the first portion of the memory system and the first thread and associate the first portion of the memory system with the second thread, which allows the second thread to reference the data using request/respond operations.