Patent classifications
G06F11/2038
STATE MANAGEMENT METHODS, METHODS FOR SWITCHING BETWEEN MASTER APPLICATION SERVER AND BACKUP APPLICATION SERVER, AND ELECTRONIC DEVICES
The present disclosure provides a state management method, a method for switching between a master application server and a backup application server, and an electronic device. In present disclosure, the management server updates the recorded backup application server state in time by querying for the connection state of the hot-backup connection between the master application server and the backup application server, and when the master application server is in failure, instead of immediately controlling the master application server and the backup application server to perform switching between the master and backup application servers, the management server controls the master application server and the backup application server to perform master-backup switching between the application servers according to the recorded backup application server state.
Intelligent and automatic load balancing of workloads on replication appliances based on appliance load scores
Various systems and methods are provided in which a replication process is initiated between a primary site and a recovery site, each having plurality of gateway appliances. Replication loads are evaluated for each given gateway appliance of the plurality of gateway appliances. If a determination is made that at least one gateway appliance of the plurality of gateway appliances is not overloaded, the plurality of gateway appliances are sorted based on replication loads respectively associated with each gateway appliance, and a determination is made as to whether a relative difference in replication loads between a gateway appliance having a highest replication load and a gateway appliance having a lowest replication load exceeds a difference threshold to determine whether the replication workloads between the gateway appliances should be rebalanced.
Front End Traffic Handling In Modular Switched Fabric Based Data Storage Systems
Systems, methods, apparatuses, and software for data storage systems are provided herein. In one example, a data storage system is provided that includes storage drives each comprising a PCIe interface, and configured to store data and retrieve the data stored on associated storage media responsive to data transactions received over a switched PCIe fabric. The data storage system includes processors configured to each manage only an associated subset of the storage drives over the switched PCIe fabric. A first processor is configured to identify first data packets received over a network interface associated with the first processor within a network buffer of the first processor as comprising a storage operation associated with at least one of the plurality of storage drives managed by a second processor, and responsively transfer the first data packets into a network buffer of the second processor.
Methods and systems of an all purpose broadband network with publish subscribe broker network
An example system includes a server communicatively connected to a cellular base transceiver station having an RF coverage area and configured for RF communication with a first entity that is a transceiver device in the RF coverage area, wherein the server comprises a first publish-subscribe broker that is part of a publish-subscribe broker network that comprises one or more publish-subscribe brokers, wherein a second entity connected to any of the one or more publish-subscribe brokers in the publish-subscribe broker network accepts communications from the transceiver device if the second entity subscribes to data packets published by the transceiver device, and wherein the data packets published by the transceiver device are routed through the publish-subscribe broker to which the second entity is connected.
TECHNIQUES FOR DEPLOYING WORKLOADS ON NODES IN A CLOUD-COMPUTING ENVIRONMENT
Described are examples for deploying workloads in a cloud-computing environment. In an aspect, based on a desired number of workloads of a process to be executed in a cloud-computing environment and based on one or more failure probabilities, an actual number of workloads of the process to execute in the cloud-computing environment to provide a level of service can be determined and deployed. In another aspect, a standby workload can be executed as a second instance of the process without at least a portion of the separate configuration used by the multiple workloads, and based on detecting termination of one of multiple workloads, the standby workload can be configured to execute based on the separate configuration of the separate instance of the process corresponding to the one of the multiple workloads.
DETECTION OF AND PROTECTION FROM MALWARE AND STEGANOGRAPHY
A method for real-time detection of and protection from steganography in a kernel mode comprises detecting transmission of a file via a firewall, an operating system, or an e-mail system. A size of the file is determined. From a file system, a stored filesize of the file is retrieved. The determined size of the file is compared to the stored filesize of the file. Responsive to the determined size of the file being larger than the stored filesize of the file, steganography detection analytics are executed on the file. Responsive to the steganography detection analytics indicating presence of steganography in the file, a steganography remediation action is executed, and information is transmitted describing the steganography to a client device.
Preparing containerized applications for backup using a backup services container and a backup services container-orchestration pod
A “backup services container” comprises “backup toolkits,” which include scripts for accessing containerized applications plus enabling utilities/environments for executing the scripts. The backup services container is added to Kubernetes pods comprising containerized applications without changing other pod containers. For maximum value and advantage, the backup services container is “over-equipped” with toolkits. The backup services container selects and applies a suitable backup toolkit to a containerized application to ready it for a pending backup. Interoperability with a proprietary data storage management system provides features that are not possible with third-party backup systems. Some embodiments include one or more components of the proprietary data storage management within the illustrative backup services container. Some embodiments include one or more components of the proprietary data storage management system in a backup services pod configured in a Kubernetes node. All configurations and embodiments are suitable for cloud and/or non-cloud computing environments.
Systems and methods for transitioning from legacy computer systems
A method may include receiving a communication from a user device, determining whether to forward the communication to a first computer system or a second computer system and forwarding the communication to the first computer system based on the determining. The method may also include generating, by the first computer system, a first response to the communication, determining whether an error occurred when processing the communication at the first computer system and forwarding the communication to the second computer system, in response to determining that an error occurred. The method may further include generating, by the second computer system, a second response to the communication and comparing the first response from the first computer system to the second response from the second computer system.
Dynamic, distributed, and scalable single endpoint solution for a service in cloud platform
A first forwarding VM may execute in a first availability zone and have a first IP address. Similarly, a second forwarding VM may execute in a second availability zone and have a second IP address. The first and second IP addresses may be recorded with a cloud DNS web service of a cloud provider such that both receive requests from applications directed to a particular DNS name acting as a single endpoint. A service cluster may include a master VM node and a standby VM node. An IPtable in each forwarding VM may forward a request having a port value to a cluster port value associated with the master VM node. Upon a failure of the master VM node, the current standby VM node may be promoted to execute in master mode and the IPtables may be updated to now forward requests having the port value to a cluster port value associated with the newly promoted master VM node (which was previously the standby VM node).
Access consistency in high-availability databases
Techniques are disclosed relating to maintaining a high availability (HA) database. In some embodiments, a computer system receives, from a plurality of host computers, a plurality of requests to access data stored in a database implemented using a plurality of clusters. In some embodiments, the computer system responds to the plurality of requests by accessing data stored in an active cluster. The computer system may then determine, based on the responding, health information for ones of the plurality of clusters, wherein the health information is generated based on real-time traffic for the database. In some embodiments, the computer system determines, based on the health information, whether to switch from accessing the active cluster to accessing a backup cluster. In some embodiments, the computer system stores, in respective clusters of the database, a changeover decision generated based on the determining.