H04L49/9084

POWER THROTTLE FOR NETWORK SWITCHES

The disclosed systems and methods provide methods and systems for providing power throttling adapted for high performance network switches. A method includes determining, for each of a plurality of measurement periods within a thermal average period, an energy usage estimate for a packet processing block configured to process ingress packets at a power gated clock rate. The method includes determining, for each of the plurality of measurement periods, a target clock rate for the packet processing block based on the determined energy usage estimates to meet a target energy value that is averaged for the thermal average period. The method includes adjusting, for each of the plurality of measurement periods, the power gated clock rate towards the target clock rate, wherein the adjusting causes the packet processing block to process the ingress packets at the adjusted power gated clock rate.

DYNAMICALLY CONTROLLING A LOCAL BUFFER OF A MODEM OF A WIRELESS DEVICE

Apparatuses, methods, and systems for dynamically controlling a local buffer of a modem of a wireless device are disclosed. One method includes receiving and queuing transmission packets in the local buffer of the modem of the wireless device for wireless transmission to a receiving device, purging each transmission packet from the local buffer after receiving an acknowledgement of reception of the transmission packet from the receiving device, and requesting acknowledgement from the receiving device when a queue of the transmission packets within the local buffer exceeds a threshold level, wherein the receiving device aggregates acknowledgment responses to a plurality of unpurged transmission packets in the local buffer and transmits an aggregated acknowledgment to the modem.

Apparatus, system, and method for detecting and remedying memory buffer leaks in network devices

A disclosed method may include (1) identifying a memory buffer that is allocated to a packet on a computing device, (2) identifying one or more characteristics of the memory buffer allocated to the packet on the computing device, (3) determining, based at least in part on the characteristics of the memory buffer, that the memory buffer allocated to the packet has leaked, and then in response to determining that the memory buffer has leaked, (4) performing at least one action to remedy the leak of the memory buffer. Various other apparatuses, systems, and methods are also disclosed.

BUFFER MANAGEMENT METHOD AND APPARATUS
20210392092 · 2021-12-16 ·

A memory management method includes: determining that available storage space of a first memory in a network device is less than a first threshold, where the first threshold is greater than 0 and the first memory stores a first packet queue; and deleting at least one packet at the tail of the first packet queue from the first memory based on the available storage space of the first memory being less than the first threshold. When the available storage space of the first memory is less than the first threshold, a packet queue, namely, the first packet queue, is selected and a packet at the tail of the packet queue is deleted from the first memory.

CONTENT MANAGEMENT SYSTEM FRAMEWORK FOR CLOUD DEPLOYMENT AND OPERATION AS MICROSERVICES
20210377363 · 2021-12-02 ·

The disclosure provides a new content server framework in which functionalities of a content server are implemented as lightweight microservices. At startup of the content server framework, a content server container and a set of microservices are launched. The content server container only has a content server application programming interface (API) which has a controller that can instantiate controller applications, each having a master module and worker(s). When a request is received, the content server API routes it to an appropriate microservice which stores the request in a repository. The master module retrieves the request from the repository and places it in a queue. The worker picks up the request from the queue and processes it. The controller keeps track of details of each controller application container that it instantiated (e.g., load and status) and automatically scale up or down the number of instances.

Power throttle for network switches

The disclosed systems and methods provide methods and systems for providing power throttling adapted for high performance network switches. A method includes determining, for each of a plurality of measurement periods within a thermal average period, an energy usage estimate for a packet processing block configured to process ingress packets at a power gated clock rate. The method includes determining, for each of the plurality of measurement periods, a target clock rate for the packet processing block based on the determined energy usage estimates to meet a target energy value that is averaged for the thermal average period. The method includes adjusting, for each of the plurality of measurement periods, the power gated clock rate towards the target clock rate, wherein the adjusting causes the packet processing block to process the ingress packets at the adjusted power gated clock rate.

SYSTEM AND METHOD FOR MODEM STABILIZATION WHEN WAITING FOR AP-DRIVEN LINK RECOVERY
20220166729 · 2022-05-26 ·

Various embodiments of methods and systems for a modem-directed application processor boot flow in a portable computing device (“PCD”) are disclosed. An exemplary method includes an application processor that transitions into an idle state, such as a WFI state, for durations of time during a boot sequence that coincide with processing by a DMA engine and/or crypto engine. That is, the application processor may “sleep” while the DMA engine and/or crypto engine process workloads in response to instructions they received from the application processor.

HIERARCHICAL STATISICALLY MULTIPLEXED COUNTERS AND A METHOD THEREOF

Embodiments of the present invention relate to an architecture that uses hierarchical statistically multiplexed counters to extend counter life by orders of magnitude. Each level includes statistically multiplexed counters. The statistically multiplexed counters includes P base counters and S subcounters, wherein the S subcounters are dynamically concatenated with the P base counters. When a row overflow in a level occurs, counters in a next level above are used to extend counter life. The hierarchical statistically multiplexed counters can be used with an overflow FIFO to further extend counter life.

Systems and methods for managing a need-to-know domain name system
11336639 · 2022-05-17 · ·

The disclosed computer-implemented method for managing a need-to-know domain name system may include (i) intercepting, by an agent of the computing device, network traffic received on the computing device, (ii) generating, by the agent, a one-time password based on a unique identifier of the agent of the computing device, (iii) wrapping, by the agent, the network traffic with the one-time password, and (iv) pushing, by the agent, the wrapped network traffic to a cloud server using a local domain name system (DNS) of the agent of the computing device, wherein the local DNS comprises a private domain name unpublished in a global DNS. Various other methods, systems, and computer-readable media are also disclosed.

Fine Grain Traffic Shaping Offload For A Network Interface Card
20220124039 · 2022-04-21 ·

A network interface card with traffic shaping capabilities and methods of network traffic shaping with a network interface card are provided. The network interface card and method can shape traffic originating from one or more applications executing on a host network device. The applications can execute in a virtual machine or containerized computing environment. The network interface card and method can perform or include several traffic shaping mechanisms including, for example and without limitation, a delayed completion mechanism, a time-indexed data structure, a packet builder, and a memory manager.