H04L49/901

STANDING QUERIES IN MEMORY
20170286483 · 2017-10-05 ·

A network device includes a database that includes a number of tables, an observer, and a query service (QS) manager. The observer monitors an element of a first table of the number of tables and provide a value of the element to a query instance of a number of query instances. The QS manager generate the number of query instances. The query instance of the number of query instances executes a compiled query in response to receiving the value.

Application Centric Distributed Storage System and Method
20170251058 · 2017-08-31 ·

A software defined storage network comprising one or more storage nodes, each storage node including a computer processor and one or more data storage devices; the one or more storage devices including a computer readable medium storing data partitioned into one or more volumes; wherein the one or more volumes are visible to at least a subset of the storage nodes and to non-storage nodes on the network; and a computer system in communication with the network having a computer processor executing instructions stored on a computer readable medium to define a plurality of actors providing a storage service; wherein each actor defines a virtual representation of at least one of the volumes and acts as a controller for each of the at least one data storage devices; wherein each of the plurality of actors places data for each volume on the storage devices according to at least one policy.

TECHNOLOGIES FOR MANAGING REPLICA CACHING IN A DISTRIBUTED STORAGE SYSTEM
20170251073 · 2017-08-31 ·

Technologies for managing replica caching in a distributed storage system include a storage manager device. The storage manager device is configured to receive a data write request to store replicas of data. Additionally, the storage manager device is configured to designate one of the replicas as a primary replica, select a first storage node to store the primary replica of the data in a cache storage and at least a second storage node to store a non-primary replica of the data in a non-cache storage. The storage manager device is further configured to include a hint in a first replication request to the first storage node that the data is to be stored in the cache storage of the first storage node as the primary replica. Further, the storage manager device is configured to transmit replication requests to the respective storage nodes. Other embodiments are described and claimed.

SENDING PACKETS USING OPTIMIZED PIO WRITE SEQUENCES WITHOUT SFENCES AND OUT OF ORDER CREDIT RETURNS
20170249079 · 2017-08-31 ·

Methods and apparatus for sending packets using optimized PIO write sequences without sfences and out-of-order credit returns. Sequences of Programmed Input/Output (PIO) write instructions to write packet data to a PIO send memory are received by a processor in an original order and executed out of order, resulting in the packet data being written to send blocks in the PIO send memory out of order, while the packets themselves are stored in sequential order once all of the packet data is written. The packets are egressed out of order by egressing packet data contained in the send blocks to an egress block using a non-sequential packet order that is different than the sequential packet order. In conjunction with egressing the packets, corresponding credits are returned in the non-sequential packet order. A block list comprising a linked list and a free list are used to facilitate out-of-order packet egress and corresponding out-of-order credit returns.

Switching of a transmit queue associated with a virtual machine
09742707 · 2017-08-22 · ·

Methods and systems for managing multiple transmit queues of a networking device of a host machine in a virtual machine system. The networking device includes multiple transmit queues that are used by multiple guests of the virtual machine system for the transmission of packets in a data communication. A hypervisor of the virtual machine system manages the switching from one or more transmit queues (i.e., old transmit queues) to one or more other queues (i.e., new transmit queues) by managing a flow of packets in the virtual machine system to maintain a proper sequence of packets and avoid a need to re-order the transmitted packets at a destination.

PACKET PROCESSING WITH REDUCED LATENCY

Generally, this disclosure provides devices, methods, and computer readable media for packet processing with reduced latency. The device may include a data queue to store data descriptors associated with data packets, the data packets to be transferred between a network and a driver circuit. The device may also include an interrupt generation circuit to generate an interrupt to the driver circuit. The interrupt may be generated in response to a combination of an expiration of a delay timer and a non-empty condition of the data queue. The device may further include an interrupt delay register to enable the driver circuit to reset the delay timer, the reset postponing the interrupt generation.

Instruction Writing Method and Apparatus, and Network Device

An instruction writing method, apparatus, and network device are provided to reduce a requirement for a storage space of a microcode processor. The method includes obtaining, by a first device, first indication information, where the first indication information indicates the first device to enable a first service function, and writing, by the first device, a first microcode instruction set corresponding to the first service function into an unused storage space of a target microcode processor in a network processor, where a size of the unused storage space is greater than or equal to a size of the first microcode instruction set.

DELAYED PROCESSING FOR ELECTRONIC DATA MESSAGES IN A DISTRIBUTED COMPUTER SYSTEM

A distributed computer system is provided. The distributed computer system includes at least one sequencer computing node and at least one matcher computing node. Electronic data messages are sequenced by the sequencer and sent to at least matcher computing node. The matcher computing node receives the electronic data messages and a reference value from an external computing source. New electronic data messages are put into a pending list before they can be acted upon by the matcher. A timer is started based on a comparison of the reference value (or a calculation based thereon) to at least one attribute or value of a new electronic data message. When the timer expires, the electronic data message is moved from the pending list to another list—where it is eligible to be matched against other, contra-side electronic data messages.

DELAYED PROCESSING FOR ELECTRONIC DATA MESSAGES IN A DISTRIBUTED COMPUTER SYSTEM

A distributed computer system is provided. The distributed computer system includes at least one sequencer computing node and at least one matcher computing node. Electronic data messages are sequenced by the sequencer and sent to at least matcher computing node. The matcher computing node receives the electronic data messages and a reference value from an external computing source. New electronic data messages are put into a pending list before they can be acted upon by the matcher. A timer is started based on a comparison of the reference value (or a calculation based thereon) to at least one attribute or value of a new electronic data message. When the timer expires, the electronic data message is moved from the pending list to another list—where it is eligible to be matched against other, contra-side electronic data messages.

MULTI-STREAM SCHEDULING FOR TIME SENSITIVE NETWORKING
20220311710 · 2022-09-29 ·

A network interface device for implementing multi-stream scheduling for time sensitive networking includes direct memory access (DMA) circuitry, comprising: descriptor parsing circuitry to read a packet descriptor from a descriptor cache, wherein the packet descriptor includes at least one scheduling control parameter including: a launch time offset, a gate cycle offset, or a reduction ratio; wherein the packet descriptor is associated with a packet stream having a traffic class; and scheduling circuitry to schedule packets from the packet stream for transmission using the at least one scheduling control parameter.