Patent classifications
G06F11/1612
Securing crash dump files
In a computer storage system, crash dump files are secured without power fencing in a cluster of a plurality of nodes connected to a storage system. Upon an occurrence of a panic of a crashing node and prior to receiving a panic message of the crashing node by a surviving node loading, in the cluster, a capturing node to become active, prior to a totem token being declared lost by the surviving node, for capturing the crash dump files of the crashing node.
FAULT-TOLERANT STORAGE SYSTEM
A fault-tolerant storage system includes a first strand comprising a first initiating device connected to a first expansion device of a first set of serially connected expansion devices, a second strand comprising a second initiating device connected to a first expansion device of a second set of serially connected expansion devices, and a plurality of zoning switches that each connect an expansion device in the first set of serially connected expansion devices to a corresponding expansion device in the second set of serially connected expansion devices. A computer program product and method executed by a computer corresponding to the above system are also disclosed herein.
Snapshot-based data corruption detection
Embodiments described herein detect data corruption in a distributed data set system. For example, a system comprises node(s) for processing queries with respect to a distributed data set comprising a plurality of storage segments. A write transaction resulting from a query with respect to a particular storage segment is logged in a log record that describes a modification to the storage segment. A log service provides the log record to a data server managing a portion of the distributed data set in which the storage segment is included, which performs the write transaction with respect to the storage segment. For redundancy purposes, the data server has replica(s) that manage respective replicas of the portion of the distributed data set managed thereby. For backup purposes, snapshots of the replica(s) are periodically generated. To determine a data corruption, a snapshot of one replica is cross-validated with a snapshot of another replica.
Data Rebuild Based On Solid State Memory Characteristics
A plurality of storage nodes within a single chassis is provided. The plurality of storage nodes is configured to communicate together as a storage cluster. The plurality of storage nodes has a non-volatile solid-state storage for user data storage. The plurality of storage nodes is configured to distribute the user data and metadata associated with the user data throughout the plurality of storage nodes, with erasure coding of the user data. The plurality of storage nodes is configured to recover from failure of two of the plurality of storage nodes by applying the erasure coding to the user data from a remainder of the plurality of storage nodes. The plurality of storage nodes is configured to detect an error and engage in an error recovery via one of a processor of one of the plurality of storage nodes, a processor of the non-volatile solid state storage, or the flash memory.
TWO-WAY REAL TIME COMMUNICATION SYSTEM THAT ALLOWS ASYMMETRIC PARTICIPATION IN CONVERSATIONS ACROSS MULTIPLE ELECTRONIC PLATFORMS
Methods, systems, and apparatuses, including computer programs encoded on computer-readable media, for receiving a first communication as part of a conversation, from an unauthenticated user of a web browser. A conversation identifier is determined based on the first communication. A first responder, a communication protocol, and a communication address of the first responder is determined. The first communication is sent to the first responder and a first reply is received. The conversation identifier is determined based on the first reply and the first reply is mapped to the web browser. The first reply is sent to the web browser.
Data processing apparatus and method
The disclosure provides a data processing device and method. The data processing device may include: a task configuration information storage unit and a task queue configuration unit. The task configuration information storage unit is configured to store configuration information of tasks. The task queue configuration unit is configured to configure a task queue according to the configuration information stored in the task configuration information storage unit. According to the disclosure, a task queue may be configured according to the configuration information.
Recovering Error Corrected Data
A plurality of storage nodes within a single chassis is provided. The plurality of storage nodes is configured to communicate together as a storage cluster. The plurality of storage nodes has a non-volatile solid-state storage for user data storage. The plurality of storage nodes is configured to distribute the user data and metadata associated with the user data throughout the plurality of storage nodes, with erasure coding of the user data. The plurality of storage nodes is configured to recover from failure of two of the plurality of storage nodes by applying the erasure coding to the user data from a remainder of the plurality of storage nodes. The plurality of storage nodes is configured to detect an error and engage in an error recovery via one of a processor of one of the plurality of storage nodes, a processor of the non-volatile solid state storage, or the flash memory.
Data storage system with metadata check-pointing
A data storage system includes multiple head nodes and data storage sleds. Volume data is replicated between a primary and one or more secondary head nodes for a volume partition and is further flushed to a set of mass storage devices of the data storage sleds. Volume metadata is maintained in a primary and one or more secondary head nodes for a volume partition and is updated in response to volume data being flushed to the data storage sleds. Also, the primary and secondary head nodes store check-points of volume metadata to the data storage sleds, wherein in response to a failure of a primary or secondary head node for a volume partition, a replacement secondary head node for the volume partition recreates a secondary replica for the volume partition based, at least in part, on a stored volume metadata checkpoint.
Erasure coding magnetic tapes for minimum latency and adaptive parity protection feedback
A magnetic tape device or system can store erasure encoded data that generates a multi-dimensional erasure code corresponding to an erasure encoded object comprising a code-word (CW). The multi-dimensional erasure code enables using a single magnetic tape in response to a random object/file request, and correct for an error within the single magnetic tape without using other tapes. Encoding logic can further utilize other magnetic tapes to generate additional parity tapes that recover data from an error of the single magnetic tape in response to the error satisfying a threshold severity for a reconstruction of the erasure coded object or chunk(s) of the CW. The encoding logic can be controlled, at least in part, by one or more iterative coding processes between multiple erasure code dimensions that are orthogonal to one another.
Host system, process, object, self-determination apparatus, and host device
A method including executing a portion of a service which is part of at least one service provided by a system including a distributed computing platform; determining object capability parameters required to perform the executing; storing information about at least one target host device; generating an announcement message reporting presence of a service type and the object capability parameters; receiving information from other announcement messages; evaluating current host device capability parameters with respect to the object capability parameters; determining when the current host device capability parameters meet a criterion; initiating a migration request message from the object for migration of the object, the object including software code and processing instructions and service function instructions, the migration to a target object host device, when the module capability parameters meet a criterion; and managing the migration of the object to the target host device.