Patent classifications
H04L49/9005
Blockchain-based dynamic storage provisioner
Methods and systems for dynamically provisioning storage on a blockchain are provided. In one embodiment, a method is provided that includes receiving a request that includes a data unit for storage. The data unit may be buffered in a data stack that stores one or more data units. The data unit may be stored in the data stack until (i) a predetermined amount of time has passed and/or (ii) a size of the data stack exceeds a predetermined threshold. The data units stored in the data stack may then be encrypted and included in one or more storage transactions. The storage transactions may also include encryption keys used to encrypt the data units. The storage transactions may then be transmitted to nodes for storage on a blockchain.
Blockchain-based dynamic storage provisioner
Methods and systems for dynamically provisioning storage on a blockchain are provided. In one embodiment, a method is provided that includes receiving a request that includes a data unit for storage. The data unit may be buffered in a data stack that stores one or more data units. The data unit may be stored in the data stack until (i) a predetermined amount of time has passed and/or (ii) a size of the data stack exceeds a predetermined threshold. The data units stored in the data stack may then be encrypted and included in one or more storage transactions. The storage transactions may also include encryption keys used to encrypt the data units. The storage transactions may then be transmitted to nodes for storage on a blockchain.
2 LAYER ALPHA BASED BUFFER MANAGEMENT WITH DYNAMIC RED
Methods and apparatus for two-layer Alpha-based buffer management with dynamic RED. A two-layer hierarchical sharing scheme using alpha parameters is provided. A buffer is dynamically shared across upper-level entities, such as hosts, using one set of alpha parameters, then a dynamically-adjusted buffer portion allocated for an upper level entity is shared among its lower level entities (e.g., sub queues) using a separate set of low-level alpha parameters. The memory spaces for the upper- and lower-level entities may be dynamically redistributed. Determinations to drop and/or mark and ECN field of received packets are performed using Dynamic RED, which employs dynamic thresholds and associated dynamic probabilities.
2 LAYER ALPHA BASED BUFFER MANAGEMENT WITH DYNAMIC RED
Methods and apparatus for two-layer Alpha-based buffer management with dynamic RED. A two-layer hierarchical sharing scheme using alpha parameters is provided. A buffer is dynamically shared across upper-level entities, such as hosts, using one set of alpha parameters, then a dynamically-adjusted buffer portion allocated for an upper level entity is shared among its lower level entities (e.g., sub queues) using a separate set of low-level alpha parameters. The memory spaces for the upper- and lower-level entities may be dynamically redistributed. Determinations to drop and/or mark and ECN field of received packets are performed using Dynamic RED, which employs dynamic thresholds and associated dynamic probabilities.
Packet processing system, method and device utilizing a port client chain
A packet processing system having each of a plurality of hierarchical clients and a packet memory arbiter serially communicatively coupled together via a plurality of primary interfaces thereby forming a unidirectional client chain. This chain is then able to be utilized by all of the hierarchical clients to write the packet data to or read the packet data from the packet memory.
Packet processing system, method and device utilizing a port client chain
A packet processing system having each of a plurality of hierarchical clients and a packet memory arbiter serially communicatively coupled together via a plurality of primary interfaces thereby forming a unidirectional client chain. This chain is then able to be utilized by all of the hierarchical clients to write the packet data to or read the packet data from the packet memory.
INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND COMPUTER-READABLE RECORDING MEDIUM STORING INFORMATION PROCESSING PROGRAM
An information processing apparatus including: a memory; and a processor coupled to the memory, the processor being configured to perform processing including: executing a buffer management processing that, under flow control over communication executed by an arithmetic processing device, sequentially obtains a plurality of packets transmitted and destined for the arithmetic processing device, stores the packets in a buffer, generates one aggregated packet by aggregating the packets, and transmits the aggregated packet to the arithmetic processing device; executing an ACK management processing that decides transmission timing for ACKs to a transmission source of the packets based on a flow rate for the aggregated packet; and executing a window management processing that decides a receive window size representing a data amount to be transmitted by one flow to the arithmetic processing device based on the flow rate for the aggregated packet.
INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND COMPUTER-READABLE RECORDING MEDIUM STORING INFORMATION PROCESSING PROGRAM
An information processing apparatus including: a memory; and a processor coupled to the memory, the processor being configured to perform processing including: executing a buffer management processing that, under flow control over communication executed by an arithmetic processing device, sequentially obtains a plurality of packets transmitted and destined for the arithmetic processing device, stores the packets in a buffer, generates one aggregated packet by aggregating the packets, and transmits the aggregated packet to the arithmetic processing device; executing an ACK management processing that decides transmission timing for ACKs to a transmission source of the packets based on a flow rate for the aggregated packet; and executing a window management processing that decides a receive window size representing a data amount to be transmitted by one flow to the arithmetic processing device based on the flow rate for the aggregated packet.
TECHNOLOGIES FOR DYNAMIC ACCELERATOR SELECTION
Technologies for dynamic accelerator selection include a compute sled. The compute sled includes a network interface controller to communicate with a remote accelerator of an accelerator sled over a network, where the network interface controller includes a local accelerator and a compute engine. The compute engine is to obtain network telemetry data indicative of a level of bandwidth saturation of the network. The compute engine is also to determine whether to accelerate a function managed by the compute sled. The compute engine is further to determine, in response to a determination to accelerate the function, whether to offload the function to the remote accelerator of the accelerator sled based on the telemetry data. Also the compute engine is to assign, in response a determination not to offload the function to the remote accelerator, the function to the local accelerator of the network interface controller.
METHOD AND SYSTEM FOR FACILITATING LOSSY DROPPING AND ECN MARKING
Methods and systems are provided for performing lossy dropping and ECN marking in a flow-based network. The system can maintain state information of individual packet flows, which can be set up or released dynamically based on injected data. Each flow can be provided with a flow-specific input queue upon arriving at a switch. Packets of a respective flow are acknowledged after reaching the egress point of the network, and the acknowledgement packets are sent back to the ingress point of the flow along the same data path. As a result, each switch can obtain state information of each flow and perform per-flow packet dropping and ECN marking.