Patent classifications
H04L47/626
Method and device for transmitting a data stream with selectable ratio of error correction packets to data packets
A method of transmitting data determines a measure of consecutive packet loss in a network; a ratio of a number of data packets and a number of error correction packets is selected in dependence on the measure. A stream of data packets is generated, and a stream of error correction packets is generated in dependence on the stream of data packets such that the proportion of error correction packets generated to the data packets generated is commensurate with the selected ratio.
Method and system for providing network egress fairness between applications
Methods and systems are provided to facilitate network egress fairness between applications. At an egress port of a network, an arbitrator can provide fairness-based traffic shaping to data associated with applications. The desired fairness-based traffic shaping can be provided based on bandwidth, traffic classes, or other parameters. Consequently, the egress link's bandwidth can be allocated with fairness among the applications.
Optimized adaptive routing to reduce number of hops
A switch is provided, which can receive a data communication at an edge of a network. The network may be made up of a plurality of switches. The switch may generate a flow channel based upon an identified source and destination for the data communication. The data communication can be routed across the plurality of switches based on minimizing a number of hops between a subset of the plurality of switches and in accordance with the flow channel.
QoS MANAGEMENT FOR MULTI-USER AND SINGLE USER EDCA TRANSMISSION MODE IN WIRELESS NETWORKS
A communication method in a communication network comprising a plurality of nodes, at least one node comprising a plurality of traffic queues for serving data traffic at different priorities, each traffic queue being associated with a respective queue backoff value computed from respective queue contention parameters having first and second values in, respectively, a first and a second contention modes, obtaining quality of service requirements of data stored in a traffic queue of the node; checking whether the quality of service requirements can be fulfilled when accessing the communication channel using the second contention mode; if the requirements cannot be fulfilled as the result of the checking, disabling access to resource units provided by the other node within one or more transmission opportunities granted to the other node on the communication channel; and transmitting data stored in the traffic queue using the first contention mode.
System and method for facilitating efficient packet injection into an output buffer in a network interface controller (NIC)
A network interface controller (NIC) capable of efficient packet injection into an output buffer is provided. The NIC can be equipped with an output buffer, a plurality of injectors, a prioritization logic block, and a selection logic block. The plurality of injectors can share the output buffer. The prioritization logic block can determine a priority associated with a respective injector based on a high watermark and a low watermark associated with the injector. The selection logic block can then determine, from the plurality of injectors, a subset of injectors associated with a buffer class and determine whether the subset of injectors includes a high-priority injector. Upon identifying a high-priority injector in the subset of injectors, the selection logic block can select the high-priority injector for injecting a packet in the output buffer.
METHOD AND SYSTEM FOR PROVIDING NETWORK EGRESS FAIRNESS BETWEEN APPLICATIONS
Methods and systems are provided to facilitate network egress fairness between applications. At an egress port of a network, an arbitrator can provide fairness-based traffic shaping to data associated with applications. The desired fairness-based traffic shaping can be provided based on bandwidth, traffic classes, or other parameters. Consequently, the egress link’s bandwidth can be allocated with fairness among the applications.
Switch device for facilitating switching in data-driven intelligent network
- Abdulla M. Bataineh ,
- Jonathan P. Beecroft ,
- Thomas L. Court ,
- Anthony M. Ford ,
- Edwin L. Froese ,
- David Charles Hewson ,
- Joseph G. Kopnick ,
- Andrew S. Kopser ,
- Duncan Roweth ,
- Gregory Faanes ,
- Michael Higgins ,
- Timothy J. Johnson ,
- Trevor Jones ,
- James Reinhard ,
- Edward J. Turner ,
- Steven L. Scott ,
- Robert L. Alverson
A switch architecture for a data-driven intelligent networking system is provided. The system can accommodate dynamic traffic with fast, effective congestion control. The system can maintain state information of individual packet flows, which can be set up or released dynamically based on injected data. Each flow can be provided with a flow-specific input queue upon arriving at a switch. Packets of a respective flow are acknowledged after reaching the egress point of the network, and the acknowledgement packets are sent back to the ingress point of the flow along the same data path. As a result, each switch can obtain state information of each flow and perform flow control on a per-flow basis.
PACKET TRAFFIC MANAGEMENT
Examples described herein relate to a system within a package. In some examples, the system includes a communication fabric and circuitry to adjust a packet throughput rate associated with the communication fabric based at least in part on incoming receive rate across multiple input ports and fabric usage. In some examples, the communication fabric is to communicatively couple devices in the package including one or more of: an accelerator, a processor, a memory, or a network interface device.
System and method for facilitating efficient host memory access from a network interface controller (NIC)
A network interface controller (NIC) capable of efficient memory access is provided. The NIC can be equipped with an operation logic block, a signaling logic block, and a tracking logic block. The operation logic block can maintain an operation group associated with packets requesting an operation on a memory segment of a host device of the NIC. The signaling logic block can determine whether a packet associated with the operation group has arrived at or departed from the NIC. Furthermore, the tracking logic block can determine that a request for releasing the memory segment has been issued. The tracking logic block can then determine whether at least one packet associated with the operation group is under processing in the NIC. If no packet associated with the operation group is under processing in the NIC, tracking logic block can notify the host device that the memory segment can be released.
Systems and methods for transport based network slicing orchestration and management
A network device obtains service requirements associated with a customer identifier, obtains a first profile describing an infrastructure design of multiple transport domains associated with at least one network slice of a network, and obtains a second profile describing performance characteristics of the multiple transport domains of the at least one network slice. The network device receives training data associated with performance measurements of the multiple transport domains of the at least one network slice, and updates a machine learning model based on the training data. The network device selects at least one of the multiple transport domains for orchestration using the updated machine learning model, the service requirements, the first profile, and the second profile.