Patent classifications
G06F3/0665
OPTIMIZED ADDITION AND REMOVAL OF COMPUTE RESOURCES IN A DISTRIBUTED STORAGE PLATFORM BY IMPLEMENTING MAPPING CHANGES IN A SHARED STORAGE SUBSYSTEM
Computer-implemented methods for optimized compute resource addition and removal in a distributed storage platform. In a case of a newly added compute resource being connected to a storage subsystem shared by compute resources in the distributed storage platform, the distributed storage platform formulates a redistribution plan to redistribute a subset of a global address space of the storage subsystem to a newly added logical volume in the storage subsystem. In a case of a removed compute resource being disconnected from the storage subsystem, the distributed storage platform formulates a redistribution plan to redistribute respective logical blocks in a logical volume for the removed compute resource to respective remaining logical volumes for respective remaining compute resources in the distributed storage platform. The distributed storage platform executes the redistribution plan to reassign data block ownerships on one or more physical memory devices in the storage subsystem.
REDUCING WRITE AMPLIFICATION AND OVER-PROVISIONING USING FLASH TRANSLATION LAYER SYNCHRONIZATION
A host Flash Translation Layer (FTL) synchronizes host FTL operations with the drive FTL operations to reduce write amplification and over-provisioning. Embodiments of FTL synchronization map, at the host FTL software (SW) stack level, logical bands in which data is managed, referred to as host bands, to the physical bands on a drive where data is stored. The host FTL tracks validity levels of data managed in host bands to determine validity levels of data stored in corresponding physical bands, and optimizes defragmentation operations (such as garbage collection processes and trim operations) applied by the host FTL SW stack to the physical bands based on the tracked validity levels.
Virtualized file server smart data ingestion
In one embodiment, a system for managing a virtualization environment includes a set of host machines, each of which includes a hypervisor, virtual machines, and a virtual machine controller, and a data migration system configured to identify one or more existing storage items stored at one or more existing File Server Virtual Machines (FSVMs) of an existing virtualized file server (VFS). For each of the existing storage items, the data migration system is configured to identify a new FSVMs of a new VFS based on the existing FSVM, send a representation of the storage item from the existing FSVM to the new FSVM, such that representations of storage items are sent between different pairs of FSVMs in parallel, and store a new storage item at the new FSVM, such that the new storage item is based on the representation of the existing storage item received by the new FSVM.
Storage system and data processing method
The functions of a mainframe environment are expanded by leveraging the functions of an open environment. A second storage of an open system externally connected to a first storage of a mainframe system comprises a second main volume of an open environment generated in association with a main logical device of the second storage, and a second sub volume of an open environment generated in association with a sub logical device of the second storage; the first storage comprises a first main volume of a mainframe environment generated in association with the main logical device of the second storage, and a first sub volume of a mainframe environment generated in association with the sub logical device of the second storage; when the first storage receives a data processing request from a host, the first storage reflects the processing request in the second storage and completes the processing; and when the first storage receives an execution request of a prescribed function, the first storage causes the second storage to execute the function.
Method, device and computer program product for storage
Techniques involve determining the number of disks in a Redundant Array of Independent Disks (RAID) storage system, and the storage system comprises multiple types of mixed disks. The techniques further involve determining a target number of RAID sets to be allocated in the storage system according to the number of disks and a predetermined threshold, and then allocating the multiple types of disks to the RAID sets according to the target number and types of disks. Such techniques propose a way to allocate mixed disks to the corresponding RAID sets. When the RAID set grouping is performed to mixed disks of the storage system, the same type of disks are allocated in the same RAID set as much as possible, thereby improving the performance of the storage system.
Container Orchestrator-Aware Storage System
An illustrative method includes a storage management system ingesting a data item into a unified storage system via a storage system interface among a plurality of storage system interfaces associated with the unified storage system, determining, in response to the ingesting of the data item into the unified storage system, an operation based on the data item, and providing a notification of the operation to an orchestration system configured to manage an execution of the operation by a computing system associated with the unified storage system. In certain embodiments, the unified storage system may be implemented as compute-aware storage system such as a container orchestrator-aware storage system.
Virtualized file server
In one embodiment, a system for managing communication connections in a virtualization environment includes a plurality of host machines implementing a virtualization environment, wherein each of the host machines includes a hypervisor, at least one user virtual machine (user VM), and a distributed file server that includes file server virtual machines (FSVMs) and associated local storage devices. Each FSVM and associated local storage device are local to a corresponding one of the host machines, and the FSVMs conduct I/O transactions with their associated local storage devices based on I/O requests received from the user VMs. Each of the user VMs on each host machine sends each of its respective I/O requests to an FSVM that is selected by one or more of the FSVMs for each I/O request based on a lookup table that maps a storage item referenced by the I/O request to the selected one of the FSVMs.
Partial pass-through data transfer system
Methods and systems for performing a partial pass-through transfer are described. In an aspect, a method includes: receiving, from a first computing system, pass-through transfer definition data to be associated with a first logical storage area, the pass-through transfer definition data including a trigger condition for a pass-through transfer and an apportionment value for the pass-through transfer; storing a representation of the pass-through transfer definition data in association with the first logical storage area; detecting a first data transfer to the first logical storage area, the first data transfer representing a transfer of a resource; determining that the first data transfer satisfies the trigger condition; and in response to determining that the first data transfer satisfies the trigger condition: identifying a portion of the resource based on the apportionment value; and initiating a second data transfer.
Mapping LUNs in a storage memory
A method for mapping LUNs (logical unit numbers) in storage memory, performed by a storage system, is provided. The method includes determining a set of LUNs in the storage memory and generating a mapping from a logical address space to all of the LUNs in the set, based on the determining, so that each logical address in the logical address space maps to one LUN in the set. The method includes accessing one or more of the LUNs in accordance with the mapping.
Converting raid data between persistent storage types
Converting RAID data between persistent storage types, including: for each portion of a RAID shard of a RAID stripe: writing, to a respective plurality of source solid state drives, the portion of the RAID shard; detecting that all portions of the RAID shard have been successfully written; copying, from one of the plurality of source solid state drives to a respective target solid state drive among a plurality of target solid state drives from one of the plurality of source solid state drives, the RAID shard, where the RAID shard is copied from a source solid state drive that is different from where each other RAID shard of the RAID stripe is copied from.