Patent classifications
G06F11/1443
Managing machine failure
A method, computer program product, and computer system are provided. A message storage area of an adjunct processor (AP) crypto adapter is filled with a plurality of command request messages sufficient to maximize utilization and performance of the AP crypto adapter. In response to detecting an error during execution of one of the plurality of command request messages, generating an AP crypto adapter command reply message. The AP crypto adapter command reply message includes the error. In response to the error being a non-recoverable failure, determining a state of the command request message, wherein the state of the command request message comprises an in-process state or a request-pending state. The AP crypto adapter command reply message is formatted, wherein the formatted AP crypto adapter command reply message is stored in a message queue in the AP crypto adapter pending completion of machine failure recovery. The AP crypto adapter is recovered.
VIDEO CONFERENCE BRIDGE SETTING SHARING, PUSHING, AND RATIONALIZATION
A conference system is provided with enhanced settings capabilities. A controller can poll for settings at each endpoint in a conference system and be able via the video stream to selectively display and compare settings among the endpoints. One location can push its settings to one or more locations to overcome failures or degradation in the conference. The settings between different controllers may be rationalized via a common denominator method or tabular method to build a knowledge of how to configure conferences and to automate responses to problems.
Failover and recovery for replicated data instances
Replicated instances in a database environment provide for automatic failover and recovery. A monitoring component can periodically communicate with a primary and a secondary replica for an instance, with each capable of residing in a separate data zone or geographic location to provide a level of reliability and availability. A database running on the primary instance can have information synchronously replicated to the secondary replica at a block level, such that the primary and secondary replicas are in sync. In the event that the monitoring component is not able to communicate with one of the replicas, the monitoring component can attempt to determine whether those replicas can communicate with each other, as well as whether the replicas have the same data generation version. Depending on the state information, the monitoring component can automatically perform a recovery operation, such as to failover to the secondary replica or perform secondary replica recovery.
Systems and methods for failure recovery in at-most-once and exactly-once streaming data processing
This patent document describes failure recovery technologies for the processing of streaming data, also referred to as pipelined data. The technologies described herein have particular applicability in distributed computing systems that are required to process streams of data and provide at-most-once and/or exactly-once service levels. In a preferred embodiment, a system comprises many nodes configured in a network topology, such as a hierarchical tree structure. Data is generated at leaf nodes. Intermediate nodes process the streaming data in a pipelined fashion, sending towards the root aggregated or otherwise combined data from the source data streams towards. To reduce overhead and provide locally handled failure recovery, system nodes transfer data using a protocol that controls which node owns the data for purposes of failure recovery as it moves through the network.
Refresh management for DRAM
A memory controller interfaces with a dynamic random access memory (DRAM). The memory controller selectively places memory commands in a memory interface queue, and transmits the commands from the memory interface queue to a memory channel connected to at least one dynamic random access memory (DRAM). The transmitted commands are stored in a replay queue. A number of activate commands to a memory region of the DRAM is counted. Based on this count, a refresh control circuit signals that an urgent refresh command should be sent to the memory region. In response to detecting a designated type of error, a recovery sequence initiates to re-transmit memory commands from the replay queue. Designated error conditions can cause the recovery sequence to restart. If an urgent refresh command is pending when such a restart occurs, the recovery sequence is interrupted to allow the urgent refresh command to be sent.
Dynamic, distributed, and scalable single endpoint solution for a service in cloud platform
A first forwarding VM may execute in a first availability zone and have a first IP address. Similarly, a second forwarding VM may execute in a second availability zone and have a second IP address. The first and second IP addresses may be recorded with a cloud DNS web service of a cloud provider such that both receive requests from applications directed to a particular DNS name acting as a single endpoint. A service cluster may include a master VM node and a standby VM node. An IPtable in each forwarding VM may forward a request having a port value to a cluster port value associated with the master VM node. Upon a failure of the master VM node, the current standby VM node may be promoted to execute in master mode and the IPtables may be updated to now forward requests having the port value to a cluster port value associated with the newly promoted master VM node (which was previously the standby VM node).
DATA COLLECTION MANAGEMENT DEVICE AND DATA COLLECTION SYSTEM
The data collection management device (10) is connected via a network to a plurality of communication devices (20) performing cyclic communication and includes: a network configuration storage (17) to store network configuration information indicating the communication devices participating in the cyclic communication; a data receiving unit (11) to receive communication data multicast from each communication device (20); a received data storage (12) to store the received communication data as collected data; a received data determination unit (13) to determine whether there is missing data in the collected data and identify unreceived communication data, based on information specifying communication cycles included in the collected data, on information specifying sender communication devices included in the collected data, and on network configuration information; and a retransmission requesting unit (15) to transmit a retransmission request of the unreceived communication data to one of the plurality of communication devices (20).
Individual data unit and methods and systems for enhancing the security of user data
An individual data unit for enhancing the security of a user data record is provided that includes a processor and a memory configured to store data. The individual data unit is associated with a network and the memory is in communication with the processor. The memory has instructions stored thereon which, when read and executed by the processor cause the individual data unit to perform basic operations only. The basic operations include communicating securely with computing devices, computer systems, and a central user data server. Moreover, the basic operations include receiving a user data record, storing the user data record, retrieving the user data record, and transmitting the user data record. The individual data unit can be located in a geographic location associated with the user which can be different than the geographic locations of the computer systems and the central user data server.
Quantum key distribution in a multi-cloud environment
In an approach to improve the field of multi-cloud environments by detecting data corruption between storage systems. Embodiments perform information tunneling on data transferring between a source storage system and a target storage system. Further, embodiments determine a checksum data of a data payload does not match an Internet Protocol (IP) packet extracted checksum and a blockchain based checksum and compare the checksum data at the target storage system with the IP packet extracted checksum and the blockchain based checksum to identify one or more checksum mismatches. Additionally, embodiments identify a corruption in a data payload based on the comparison between the checksum data at the target storage system and the IP packet extracted checksum and the blockchain based checksum, validate the corruption in the data payload, and update respective entities of identified corruption in the data payload.
REDUNDANT CONTROL IN A DISTRIBUTED AUTOMATION SYSTEM
A method for redundant control in a distributed automation system, preferably a real-time automation system, for operating a client device of the distributed automation system is discussed. The method includes using the client device to monitor for the occurrence of a fault in communication between the client device and a first computing infrastructure that is part of the distributed automation system and operates the client device. The method may also include using the client device, once the fault occurs, to instruct a second computing infrastructure of the distributed automation system to operate the client device.