Patent classifications
H04L12/863
ELECTRONIC DEVICE AND METHOD FOR PROCESSING DATA IN ELECTRONIC DEVICE
Disclosed are an electronic device and a method of processing data by an electronic device. The method may include: detecting an operation mode of an application processor; storing data received through a communication processor in the communication processor when the application processor operates in a first mode; and transmitting at least some of the stored data to the application processor when the application processor operates in a second mode.
Switching method
A method for providing identifiers for virtual devices in a network. The method comprises receiving a discovery data packet directed to a physical network node associated with a physical endpoint device. A response to the discovery data packet directed to a physical network node is provided, the response comprising an identifier of a virtual device. At least one further discovery data packet directed at least to said virtual device is received. A response to a first one of the further discovery data packets is provided, the response comprising an identifier of a virtual endpoint device. At least some functionality of the virtual endpoint device is provided by the physical endpoint device.
Large receive offload for virtual machines
A network interface controller (NIC) that includes a set of receive NIC queues capable of performing large receive offload (LRO) operations by aggregating incoming receive packets is provided. Each NIC queue turns on or off its LRO operation based a set of LRO enabling rules or parameters, whereby only packets that meet the set of rules or parameters will be aggregated in the NIC queue. Each NIC queue is controlled by its own set of LRO enabling rules such that the LRO operations of the different NIC queues can be individually controlled.
Techniques for enabling packet prioritization without starvation in communications networks
A method is provided in one example embodiment and includes determining whether a packet received at a network node in a communications network is a high priority packet; determining whether a low priority queue of the network node has been deemed to be starving; if the packet is a high priority packet and the low priority queue has not been deemed to be starving, adding the packet to a high priority queue, wherein the high priority queue has strict priority over the low priority queue; and if the packet is a high priority packet and the low priority queue has been deemed to be starving, adding the packet to the low priority queue.
Method and system for storing packets for a bonded communication links
Method and system for storing packets received from a bonded communication links according to latency of the communication link that has the largest latency among all communication links of the bonded communication links. Embodiments of present inventions can be applied to bonded communication links, including wireless connection, Ethernet connection, Internet Protocol connection, asynchronous transfer mode, virtual private network, WiFi, high-speed downlink packet access, GPRS, LTE, and X.25. The present invention presents methods comprising the steps of estimating storage size of a queue, wherein the queue is for storage the one or more packets received from the bonded communication links. The storage size is based on one or more factors, including largest latency, bandwidth of each of the plurality of communication links, and allowed time duration of packet storage
Network Overlay Systems and Methods Using Offload Processors
A network overlay system can include a data transport module having a network interface and a translation module configured to generate offload processing addresses for the network packet data; a system bus; at least one host processor connected to the system bus; and at least one offload processor module coupled to the system bus and configured to receive network packet data associated. Offload processor modules include processing circuits associated with at least one of the offload processing addresses that are configured to encapsulate the network packet data for transport on a logical network or decapsulate the network packet data received from the logical network. The offload processing circuits encapsulate or decapsulate network packet data independent of any host processor.
DATA CACHING METHOD AND DEVICE, AND STORAGE MEDIUM
Disclosed is a data caching method, comprising: according to an input port number of a cell, storing the cell in a corresponding first-in first-out queue; determining that a cell to be dequeued can be dequeued in the current Kt.sup.th cycle, scheduling for the cell to be dequeued to be dequeued, acquiring the actual value of the number of splicing units occupied by the cell to be dequeued, and storing the cell to be dequeued in a register the same number of bits wide as a bus in a cell splicing manner, wherein determining that the cell to be dequeued can be dequeued is conducted in accordance with the fact that a first back pressure count value of the (K−1).sup.th cycle is less than or equal to a first preset threshold value, and the first back pressure count value of the (K−1).sup.th cycle is obtained in accordance with an estimated value of the number of the splicing units occupied when the previous cell to be dequeued is dequeued, the number of splicing units capable of being transmitted by the bus in each cycle, and a first back pressure count value of the (K−2).sup.th cycle. Also disclosed at the same time are a data caching device and a storage medium.
End-to end lossless Ethernet in Ethernet fabric
One embodiment of the present invention provides a computing system. The computing system includes a packet processor, a buffer management module, a data monitoring module, and a flow control module. The packet processor identifies a class of service indicating priority-based flow control associated with a remote computing system from a notification message. The buffer management module creates a buffer dedicated for frames belonging to the class of service from the remote computing system in response to identifying the class of service. The data monitoring module detects a potential overflow of the buffer. The flow control module operates in conjunction with the packet processor to generate a pause frame in response to detecting a potential overflow.
Flow-based adaptive private network with multiple WAN-paths
Systems and techniques are described which improve performance, reliability, and predictability of networks without having costly hardware upgrades or replacement of existing network equipment. An adaptive communication controller provides WAN performance and utilization measurements to another network node over multiple parallel communication paths across disparate asymmetric networks which vary in behavior frequently over time. An egress processor module receives communication path quality reports and tagged path packet data and generates accurate arrival times, send times, sequence numbers and unutilized byte counts for the tagged packets. A control module generates path quality reports describing performance of the multiple parallel communication paths based on the received information and generates heartbeat packets for transmission on the multiple parallel communication paths if no other tagged data has been received in a predetermined period of time to ensure performance is continually monitored. An ingress processor module transmits the generated path quality reports and heartbeat packets.
Transmission apparatus and transmission method
There is provided a transmission apparatus including: a buffer provided with a queue for each of user flows, the buffer configured to store data in the queue for the user flow to which the data belongs; a storage unit configured to store a link list indicating an output order of the user flows; an issuing unit configured to issue a predetermined output permissible amount to the queue for the user flow by referring to the link list; and an output control unit configured to output data from the queue to which the output permissible amount is issued.