Patent classifications
H04L12/743
Power reduction methods for variable sized tables
Disclosed are devices, systems, and methods for the use of memory including a data table configured to store a plurality of elements, wherein the plurality of elements are arranged into a plurality of buckets and each of the plurality of buckets comprising a plurality of entries. A first power domain can be associated with an entry of each bucket or with a first bucket. A second power domain can be associated with a second entry of each bucket or a second bucket. Processing logic can be configured to search for a particular value stored in an element of the plurality of elements by selecting buckets of the plurality of buckets and selecting at least one entry of each of the buckets. A programmable register can be used to select a powered state of the second power domain based on a configuration of the programmable register.
TECHNOLOGIES FOR PROVIDING STREAMLINED PROVISIONING OF ACCELERATED FUNCTIONS IN A DISAGGREGATED ARCHITECTURE
Technologies for providing streamlined provisioning of accelerated functions in a disaggregated architecture include a compute sled. The compute sled includes a network interface controller and circuitry to determine whether to accelerate a function of a workload executed by the compute sled, and send, to a memory sled and in response to a determination to accelerate the function, a data set on which the function is to operate. The circuitry is also to receive, from the memory sled, a service identifier indicative of a memory location independent handle for data associated with the function, send, to a compute device, a request to schedule acceleration of the function on the data set, receive a notification of completion of the acceleration of the function, and obtain, in response to receipt of the notification and using the service identifier, a resultant data set from the memory sled. The resultant data set was produced by an accelerator device during acceleration of the function on the data set. Other embodiments are also described and claimed.
INTELLIGENT AND DYNAMIC OVERLAY TUNNEL FORMATION VIA AUTOMATIC DISCOVERY OF CITRIVITY/SDWAN PEER IN THE DATAPATH IN A PURE PLUG AND PLAY ENVIRONMENT WITH ZERO NETWORKING CONFIGURATION
Described embodiments provide systems and methods of forming overlay tunnels for delivery of data between networked devices. A first intermediary device may transmit, responsive to a connection request from a client, a request having a source IP address corresponding to a first virtual IP address of the first device and a first payload including first security hash information to be processed by a second intermediary device. The first device may receive, from the second intermediary device, a response. The response may have a source IP address corresponding to the IP address of the server and a second payload including a virtual IP address of the second device, responsive to second security hash information corresponding to the first security hash information. The first device may establish an overlay tunnel using the first virtual IP address and the second virtual IP address for communicating data between the client and the server.
Network Load Distribution Device and Method
A network load balancing apparatus has a data buffer provided to each communication path of transfer destinations of a received packet and being associated with a virtual function, determines a destination virtual function based on a field value of the received packet, determines a communication path of a transfer destination of a packet to be subject to priority control based on a first hash value calculated using the field value, determines a communication path of a transfer destination of a packet to be subject to load balancing control, to match a preset load balancing situation of the data buffer, based on a second hash value based on the first hash value, and transmits the packet to a data buffer corresponding to the destination virtual function and the communication path of the transfer destination.
DISTRIBUTED ROUTING AND LOAD BALANCING IN A DYNAMIC SERVICE CHAIN
Disclosed is distributed routing and load balancing in a dynamic service chain, receiving a packet at a first service instance, including a NSH imposed on the by a service classifier. The NSH includes a stream affinity code consistent for packets in a stream. The method also includes processing the packet at the first instance where the instance performs a first service in a service chain that includes second and third services. The first service instance accesses a flow table using the stream affinity code to select a second service instance performing the second service from among service instances performing the second service, and the first instance routes the packet to the selected second service instance upon egress from the first service instance. The method can include hashing the stream affinity code to access the flow table and access an available instance using the hash as a key to a CHT.
Storing keys with variable sizes in a multi-bank database
A network device determines, based on a size of a lookup value, that the lookup value is to be stored across a set of two or more memory banks including a first memory bank and a second memory bank of a database. A first hash function is for determining locations for storing lookup values entirely in the first memory bank, whereas a second hash function is for determining locations for storing lookup values entirely in the second memory bank. A hash operation is performed on the lookup value using the first hash function to determine a memory location for storing the lookup value. A first segment of the lookup value is stored in the first memory bank at the memory location determined using the first hash function, and a second segment of the lookup value is stored in the second memory bank at the memory location determined using the first hash function.
Methods and systems for selecting a tunnel of an aggregated connection
The present invention discloses methods and systems for transmitting a received packet at a first network node through an aggregated connection. The first network node determines session information of the received packet and determines whether a new tunnel needs to be selected for transmitting the received packet. When a new tunnel needs to be selected, a hash result is determined. The hash result is substantially based on the session information and the number of available tunnels. A first tunnel is determined for transmitting the received packet according to the hash result. The session information and tunnel ID of the first tunnel is then stored in a first database. The received packet is transmitted through the first tunnel. When a new tunnel need not be selected, a lookup is performed to determine a tunnel ID substantially based on the session information. The received packet is transmitted through the determined tunnel.
SYSTEM AND METHODS FOR TAMPER PROOF INTERACTION RECORDING AND TIMESTAMPING
A system and method for securely recording voice communications, comprising a network-connected computer server and an authentication system which verifies the validity of voice communications.
System and method for dynamic name configuration in content-centric networks
One embodiment of the present invention provides a system for automatic configuration of a computing device in a content-centric network (CCN). During operation, the system sends, by the computing device on at least one of the computing device's faces, an interest in configuration information. The interest has a predetermined name prefix. The system then receives a content object in response to the interest. The content object includes at least a default name prefix, to which the computing device can send other interests. The system further configures the computing device based on the received content object.
Media streaming system supporting several streaming engines and using content addressing
A method of streaming media content over a network from a media cache node is described. The method includes receiving a request for a media content item from a client device, the request comprising an address identifying a media content item to be streamed. In response to the request, a streaming engine process is allocated to the media content item for fulfilling the request. Based on the address identifying the media content item, a location comprising a media cache node able to provide the media content item is determined and the media content item is streamed to the client device using the streaming engine process allocated to the media content item. Further methods of streaming a media content item and providing access to media content are also described.