Patent classifications
H04L67/1097
Automatically anonymizing data in a distributed storage system
Data can be automatically anonymized in a distributed storage system. For example, a system can receive a notification indicating that an object is stored in a non-persistent bucket of a distributed storage system. The system can read the object from the non-persistent bucket in response to receiving the notification. The system can generate an anonymized version of the object by performing one or more anonymization operations with respect to the object. The system can store the anonymized version of the object in a persistent bucket of the distributed storage system. The system can then transmit a command for causing the object to be removed from the non-persistent bucket.
Hard zoning of virtual local area networks in a fibre channel fabric
A network where FC and Ethernet storage traffic share the underlying network. The network extends FC SAN storage specific attributes to Ethernet storage devices. The network is preferably formed of FC switches, so each edge switch acts as an FCoE FCF, with internal communications done using FC. IP packets are encapsulated in FC packets for transport. Preferably, either each outward facing switch port can be configured as an Ethernet or FC port, so devices can be connected as desired. FCoE devices connected to the network are in particular virtual LANs (VLANs). The name server database is extended to include VLAN information for the device and the zoning database has automatic FCOE_VLAN zones added to provide a mechanism for enhanced soft and hard zoning. Zoning is performed with the conventional zoning restrictions enhanced by including the factor that any FCoE devices must be in the same VLAN.
Hard zoning of virtual local area networks in a fibre channel fabric
A network where FC and Ethernet storage traffic share the underlying network. The network extends FC SAN storage specific attributes to Ethernet storage devices. The network is preferably formed of FC switches, so each edge switch acts as an FCoE FCF, with internal communications done using FC. IP packets are encapsulated in FC packets for transport. Preferably, either each outward facing switch port can be configured as an Ethernet or FC port, so devices can be connected as desired. FCoE devices connected to the network are in particular virtual LANs (VLANs). The name server database is extended to include VLAN information for the device and the zoning database has automatic FCOE_VLAN zones added to provide a mechanism for enhanced soft and hard zoning. Zoning is performed with the conventional zoning restrictions enhanced by including the factor that any FCoE devices must be in the same VLAN.
Feature Resource Self-Tuning and Rebalancing
An apparatus comprises at least one processing device that includes a processor coupled to a memory. The processing device is configured to identify a plurality of resource objects associated with a processing device, to group correlated resource objects according to processing device utilization of the resource objects, to assign a first weight to a first resource object grouping, wherein the first weight is associated with a performance impact of the first resource object grouping on the processing device, and to release at least some of the first resource object grouping to provide additional resources to a second resource object grouping, the additional resources resulting from the releasing, wherein the first object grouping is selected for the releasing based on a comparison between the first weight and a second weight associated with the second resource object grouping, wherein the releasing is performed to improve performance of the processing device.
Virtualized file server
In one embodiment, a system for managing communication connections in a virtualization environment includes a plurality of host machines implementing a virtualization environment, wherein each of the host machines includes a hypervisor, at least one user virtual machine (user VM), and a distributed file server that includes file server virtual machines (FSVMs) and associated local storage devices. Each FSVM and associated local storage device are local to a corresponding one of the host machines, and the FSVMs conduct I/O transactions with their associated local storage devices based on I/O requests received from the user VMs. Each of the user VMs on each host machine sends each of its respective I/O requests to an FSVM that is selected by one or more of the FSVMs for each I/O request based on a lookup table that maps a storage item referenced by the I/O request to the selected one of the FSVMs.
Virtualized file server
In one embodiment, a system for managing communication connections in a virtualization environment includes a plurality of host machines implementing a virtualization environment, wherein each of the host machines includes a hypervisor, at least one user virtual machine (user VM), and a distributed file server that includes file server virtual machines (FSVMs) and associated local storage devices. Each FSVM and associated local storage device are local to a corresponding one of the host machines, and the FSVMs conduct I/O transactions with their associated local storage devices based on I/O requests received from the user VMs. Each of the user VMs on each host machine sends each of its respective I/O requests to an FSVM that is selected by one or more of the FSVMs for each I/O request based on a lookup table that maps a storage item referenced by the I/O request to the selected one of the FSVMs.
Sharing prepopulated container image caches among container execution environments
Techniques are described for sharing prepopulated container image caches among container execution environments to improve the performance of container launches. The container images used to prepopulate such a cache at a computing device supporting one or more container execution environments can include various container images that are used as the basis for a wide range of user-created containers such as, for example, container images representing popular operating system distributions, database servers, web-application frameworks, and so forth. Existing systems typically obtain these container images as needed at runtime when launching containers (for example, from a container registry or other external source), often incurring significant overhead in the container launch process. The use of a prepopulated container image cache can significantly improve the performance of container launches by making such commonly used container images available to container execution environments running at a computing device ahead of time.
Methods and systems of industrial processes with self organizing data collectors and neural networks
Systems and methods for data collection for an industrial heating process are disclosed. The system according to one embodiment can include a plurality of data collectors, including a swarm of self-organized data collector members, wherein the swarm of self-organized data collector members organize to enhance data collection based on at least one of capabilities and conditions of the data collector members of the swarm, and wherein the plurality of data collectors is coupled to a plurality of input channels for acquiring collected data relating to the industrial heating process, and a data acquisition and analysis circuit for receiving the collected data via the plurality of input channels and structured to analyze the received collected data using a neural network to monitor a plurality of conditions relating to the industrial heating process.
Methods and systems of industrial processes with self organizing data collectors and neural networks
Systems and methods for data collection for an industrial heating process are disclosed. The system according to one embodiment can include a plurality of data collectors, including a swarm of self-organized data collector members, wherein the swarm of self-organized data collector members organize to enhance data collection based on at least one of capabilities and conditions of the data collector members of the swarm, and wherein the plurality of data collectors is coupled to a plurality of input channels for acquiring collected data relating to the industrial heating process, and a data acquisition and analysis circuit for receiving the collected data via the plurality of input channels and structured to analyze the received collected data using a neural network to monitor a plurality of conditions relating to the industrial heating process.
PMEM cache RDMA security
Techniques are described for providing one or more clients with direct access to cached data blocks within a persistent memory cache on a storage server. In an embodiment, a storage server maintains a persistent memory cache comprising a plurality of cache lines, each of which represent an allocation unit of block-based storage. The storage server maintains an RDMA table that include a plurality of table entries, each of which maps a respective client to one or more cache lines and a remote access key. An RDMA access request to access a particular cache line is received from a storage server client. The storage server identifies access credentials for the client and determines whether the client has permission to perform the RDMA access on the particular cache line. Upon determining that the client has permissions, the cache line is accessed from the persistent memory cache and sent to the storage server client.