G06F11/1415

SYSTEMS AND METHODS FOR APPLICATION FAILOVER MANAGEMENT USING A DISTRIBUTED DIRECTOR AND PROBE SYSTEM

Systems and methods for application management are disclosed. The system may include a probe system and one or more director systems, each comprising at least one memory and one or more processors configured to execute instructions. The instructions may include monitoring an availability of an application; updating a status associated with the availability of the application in a first data store; polling the first data store in intervals to retrieve the status associated with the availability of the application; upon retrieving at least a consecutive predetermined number of statuses associated with the application being unavailable, determining the application is unavailable; determining whether at least one other director system of the one or more director systems has determined the application is unavailable; and upon determining the at least one other director system of the one or more director systems has determined the application is unavailable, triggering a failover process.

SECURITY INSPECTIONS

In some examples, a method for performing an out-of-band security inspection of a device comprises generating a snapshot of the state of the device, storing data representing the snapshot to a non-volatile storage of the device, and storing a hash of the snapshot in a device BIOS, transitioning the power state of the device, triggering boot of a trusted diskless operating system image, providing the data representing the snapshot and the hash of the snapshot to the trusted diskless operating system image, and executing a script selected on the basis of a trigger event and the hash of the snapshot to analyse at least a portion of the non-volatile storage of the device.

SYSTEMS AND METHODS FOR POST-DETECT COMBINING OF A PLURALITY OF DOWNLINK SIGNALS REPRESENTATIVE OF A COMMUNICATION SIGNAL

Embodiments of systems and methods for combining downlink signals representative of a communication signal are provided herein. An example method comprises receiving samples of the downlink signals from multiple antenna feeds; generating first symbols for a first signal and second symbols for a second signal based on performing timing recovery operations on the first signal and the second signal, respectively; generating offset information based on performing a correlator operation on the first and second symbols; and combining the first and second signals based on performing a weighted combiner operation. At least one of the first timing recovery operation, the second timing recovery operation, the correlator operation, and the combing are performed in a plurality of processing blocks in one or more processors, wherein the first and second processing block operate in parallel.

TECHNIQUES TO PROVIDE SELF-HEALING DATA PIPELINES IN A CLOUD COMPUTING ENVIRONMENT
20220382620 · 2022-12-01 · ·

Embodiments may generally be directed to systems and techniques to detect failure events in data pipelines, determine one or more remedial actions to perform, and perform the one or more remedial actions.

Coexisting differing erasure codes

A method for proactively rebuilding user data in a plurality of storage nodes of a storage cluster is provided. The method includes distributing user data and metadata throughout the plurality of storage nodes such that the plurality of storage nodes can read the user data, using erasure coding, despite loss of two of the storage nodes. The method includes determining that one of the storage nodes is unreachable and determining to rebuild the user data for the one of the storage nodes that is unreachable. The method includes reading the user data across a remainder of the plurality of storage nodes, using the erasure coding and writing the user data across the remainder of the plurality of storage nodes, using the erasure coding. A plurality of storage nodes within a single chassis that can proactively rebuild the user data stored within the storage nodes is also provided.

System and method for secure access to a distributed virtual firmware network drive

An information handling system includes a virtual network access module configured to access a virtual network drive that has a first partition in a local storage resource and a second partition in a remote storage resource. In response to detection of an exception, a processor may trigger an exception handler that directs a service processor to initialize a network stack. The processor initializes a mailbox to transmit a mailbox request to retrieve network configuration settings to be used in the initialization of the network stack. The service processor transmits a request to the processor to initialize the mailbox, and initializes the network stack based on the network configuration settings. Subsequent to the initialization of the network stack, a universal network device interface request may be sent to mount and secure communication with the virtual network drive.

Storage System and Method for Data Recovery After Detection of an Uncorrectable Error

A storage system caches, in volatile memory, data read from non-volatile memory. After detecting an uncorrectable error in the data cached in the volatile memory, the storage system replaces the cached data with data re-read from the non-volatile memory and updated to reflect any changes made to the data after it was stored in the non-volatile memory. The storage system can also analyze a pattern in data adjacent to the uncorrectable error and predict corrected data based on the pattern.

METHOD, ELECTRONIC DEVICE AND COMPUTER PROGRAM PRODUCT FOR STORAGE MANAGEMENT
20220342753 · 2022-10-27 ·

Techniques involve: writing, when a first disk for dirty page storage has a failure, a first target page description generated by a first node and associated with a first set of target dirty pages of the first node and a second set of target dirty pages of a second node to a first page description storage layer for the first node in a second disk; writing a second target page description generated by the second node and associated with the first set of target dirty pages and the second set of target dirty pages to a second page description storage layer for the second node in the second disk; and restoring, when the failure has been eliminated, the first set of target dirty pages and the second set of target dirty pages in the first disk based on at least one of the descriptions.

Method and system for managing fault recovery in system-on-chips
11609821 · 2023-03-21 · ·

A fault recovery system including a fault controller is disclosed. The fault controller is coupled between a processor and an interconnect, and configured to receive a time-out signal that is indicative of a failure of the processor to execute a transaction after a fault is detected in the processor. The failure in the execution of the transaction results in queuing of the interconnect. Based on the time-out signal, the fault controller is further configured to generate and transmit a control signal to the processor to disconnect the processor from the interconnect. Further, the fault controller is configured to execute the transaction, and in turn, dequeue the interconnect. When the transaction is successfully executed, the fault controller is further configured to generate a status signal to reset the processor, thereby managing a fault recovery of the processor.

PARITY DATA MODIFICATION FOR PARTIAL STRIPE DATA UPDATE
20230082636 · 2023-03-16 ·

A system and method for recovery data generation for partial memory block modifications. An example system including a memory device and a processing device, operatively coupled with the memory device, to perform operations including: receiving a command to modify a portion of a memory block that is stored by the memory device, wherein the command comprises user data and a location in the memory block; reading user data at the location and existing recovery data for the memory block; generating recovery data for the memory block based on the existing recovery data, the user data at the location, and the user data of the command; and writing the user data of the command and the generated recovery data to the memory device, wherein the user data overwrites the portion of the memory block at the location.