Patent classifications
G06F11/2097
COLLISION HANDLING DURING AN ASYNCHRONOUS REPLICATION
Methods and systems for collision handling during an asynchronous replication are provided. A system includes a cache memory system comprising a number of cache memory pages. A collision detector detects when a host is attempting to overwrite a cache memory page that has not been completely replicated. A revision page tagger copies the cache memory page to a free page and tags the copied page as protected.
Methods and systems of an all purpose broadband network with publish subscribe broker network
An example system includes a server communicatively connected to a cellular base transceiver station having an RF coverage area and configured for RF communication with a first entity that is a transceiver device in the RF coverage area, wherein the server comprises a first publish-subscribe broker that is part of a publish-subscribe broker network that comprises one or more publish-subscribe brokers, wherein a second entity connected to any of the one or more publish-subscribe brokers in the publish-subscribe broker network accepts communications from the transceiver device if the second entity subscribes to data packets published by the transceiver device, and wherein the data packets published by the transceiver device are routed through the publish-subscribe broker to which the second entity is connected.
METHODS, DEVICES AND SYSTEMS FOR REAL-TIME CHECKING OF DATA CONSISTENCY IN A DISTRIBUTED HETEROGENOUS STORAGE SYSTEM
A computer-implemented method may comprise executing, by a first plurality of replicated state machines, a sequence of ordered agreements to make mutations to a data stored in a first data storage of a first type and executing, by a second plurality of replicated state machines, the sequence of ordered agreements to make mutations to the data stored in a second data storage of a second type. First metadata of the mutated data stored in the first data storage may then be received and stored, as may second metadata of the mutated data stored in the second data storage. A comparison of the stored first and second metadata may then be carried out when the data stored in the first data storage that corresponds to the first metadata and the data stored in the second data storage that corresponds to the second metadata have been determined to have settled according to the predetermined one of the sequence of ordered agreements. A selected action may then be carried out depending upon a result of the comparison.
DISTRIBUTED DATABASE SYSTEM AND DATA DISASTER BACKUP DRILLING METHOD
A distributed database system, a data disaster backup exercise method and a non-transitory computer-readable storage medium are disclosed. The distributed database system may include a local computer room (110) and an offsite computer room (120), where the local computer room (110) includes a local management node (111) and a local database cluster (112), the offsite computer room (120) includes an offsite management node (121), an offsite exercise database cluster (123) and an offsite synchronization database cluster (122); where the local database cluster (112) and the offsite synchronization database cluster (122) are both connected with the local management node (111); the offsite exercise database cluster (123) is configured for: establishing a first connection with the offsite management node (121); and receiving a test service sent by a service layer.
Maintaining A Synchronous Replication Relationship Between Two Or More Storage Systems
Maintaining a synchronous replication relationship between two or more storage systems, including: receiving, by at least one of a plurality of storage systems across which a dataset will be synchronously replicated, timing information for at least one of the plurality of storage systems; and establishing, based on the timing information, a synchronous replication lease describing a period of time during which the synchronous replication relationship is valid, wherein a request to modify the dataset may only be acknowledged after a copy of the dataset has been modified on each of the storage systems.
REPLICATION FOR CYBER RECOVERY FOR MULTIPLE TIER DATA
Replication of a filesystem or a mount point or share may replicate all data that it consists of irrespective of where the data is stored. Replication protects data irrespective of location. One method is to replicate the filesystem namespace as is while skipping the data outside of the appliance/machine so that replication cost and time are reasonable. The data outside of the machine, like cloud/tape data is protected differently. One example method includes a data protection operation configured to replication a namespace associated with multiple data tiers. During replication, data from one of the tiers is skipped while all of the namespace metadata is replicated. The recovery restores the namespace metadata and the data that was replicated from the other tier. This may be performed in connection with cyber security, for example when replicating multi-tier data to a vault.
System and method for providing high availability data
A data processing system may store a data set at a plurality of data centers. The data centers and hosts within the data centers may, for example, be organized according to a multi-tiered ring arrangement. A hashing arrangement may be used to implement the ring arrangement to select the data centers and hosts where the writing and reading of the data sets occurs. Version histories may also be written and read a the hosts and may be used to evaluate causal relationships between the data sets after the reading occurs.
Redundant storage gateways
Methods, apparatus, and computer-accessible storage media for providing redundant storage gateways. A client may create a storage gateway group and add storage gateways to the group. The client may assign one or more volumes on a remote data store to each the storage gateways in the group. Volume data for each storage gateway in the group may be replicated to at least one other storage gateway in the group. If one of the gateways in the group becomes unavailable, one or more other gateways in the group may take over volumes previously assigned to the unavailable gateway, using the replicated data in the group to seamlessly resume gateway operations for the respective volumes. Client processes that previously communicated with the unavailable gateway may be manually or automatically directed to the gateway(s) that are taking over the unavailable gateway's volumes.
Preparing containerized applications for backup using a backup services container and a backup services container-orchestration pod
A “backup services container” comprises “backup toolkits,” which include scripts for accessing containerized applications plus enabling utilities/environments for executing the scripts. The backup services container is added to Kubernetes pods comprising containerized applications without changing other pod containers. For maximum value and advantage, the backup services container is “over-equipped” with toolkits. The backup services container selects and applies a suitable backup toolkit to a containerized application to ready it for a pending backup. Interoperability with a proprietary data storage management system provides features that are not possible with third-party backup systems. Some embodiments include one or more components of the proprietary data storage management within the illustrative backup services container. Some embodiments include one or more components of the proprietary data storage management system in a backup services pod configured in a Kubernetes node. All configurations and embodiments are suitable for cloud and/or non-cloud computing environments.
Dynamic, distributed, and scalable single endpoint solution for a service in cloud platform
A first forwarding VM may execute in a first availability zone and have a first IP address. Similarly, a second forwarding VM may execute in a second availability zone and have a second IP address. The first and second IP addresses may be recorded with a cloud DNS web service of a cloud provider such that both receive requests from applications directed to a particular DNS name acting as a single endpoint. A service cluster may include a master VM node and a standby VM node. An IPtable in each forwarding VM may forward a request having a port value to a cluster port value associated with the master VM node. Upon a failure of the master VM node, the current standby VM node may be promoted to execute in master mode and the IPtables may be updated to now forward requests having the port value to a cluster port value associated with the newly promoted master VM node (which was previously the standby VM node).