Patent classifications
H04L47/722
Vehicle network and method of communication
A method of communication in a vehicle network is provided. An example method includes transmitting a network allocation map in a TDMA cycle, indicating reservation of time slots in the TDMA cycle. The method further includes transmitting a synchronization signal in the TDMA cycle, to synchronize the timing of nodes in the vehicle network. Each of the reserved time slots is identified by at least a network ID of a transmitting node in the vehicle network, and a slot type comprising one of a low latency traffic slot, and a bulk traffic slot. Further, the low latency traffic slots are repeated in the TDMA cycle at least as frequently as a guaranteed QoS latency parameter. Further, the bulk traffic slots are at least as long as a guaranteed QoS throughput parameter.
SERVICE-TO-SERVICE SCHEDULING IN CONTAINER ORCHESTRATORS
Methods, systems, and computer program products for service-to-service scheduling in container orchestrators are provided herein. A computer-implemented method includes reserving, by a network orchestrator, network resources requested between a plurality of services, wherein each of the services is implemented as one or more replicas running on a set of nodes of a cluster, managed by the network orchestrator, that use the network resources to serve incoming requests to the plurality services; monitoring utilization of the network resources; and scheduling, by the network orchestrator based on the monitoring, one or more new replicas of the plurality of services and the incoming requests to the plurality of services in a collaborative manner to increase at least one network performance characteristic.
ADAPTIVE PROGRESSIVE DOWNLOADING OF A CONTENT BROADCAST IN REAL TIME ON A MOBILE RADIOCOMMUNICATION NETWORK, ASSOCIATED COMPUTER PROGRAM AND MULTIMEDIA-STREAM PLAYER TERMINAL
The adaptive progressive downloading of a content broadcast in real time on a radiocommunication network, implemented by a multimedia stream player terminal connected to the radiocommunication network is disclosed. A method implements, before playing the content, a configuration, according to at least one characteristic of the radiocommunication network, of a depth of buffer dedicated to the temporary storage of the content before playing, in the multimedia stream player terminal.
Packet Buffer Spill-Over in Network Devices
A packet processor of a network device receives packets ingressing from a plurality of network links via a plurality of network ports of the network device. The packet processor buffers the packets in an internal packet memory in a plurality of queues, including a first queue. In response to the packet processor detecting congestion in the internal packet memory, the packet processor selectively forwards a group of multiple packets in the first queue from the internal packet memory to a first port, among one or more ports coupled to one or more external memories, to transfer the group of multiple packets to a first external memory that is coupled to the first port so that the first queue is stored across the internal packet memory and the first external packet memory.
Packet Buffer Spill-Over in Network Devices
A packet processor of a network device receives packets ingressing from a plurality of network links via a plurality of network ports of the network device. The packet processor buffers the packets in an internal packet memory in a plurality of queues, including a first queue. In response to the packet processor detecting congestion in the internal packet memory, the packet processor selectively forwards a group of multiple packets in the first queue from the internal packet memory to a first port, among one or more ports coupled to one or more external memories, to transfer the group of multiple packets to a first external memory that is coupled to the first port so that the first queue is stored across the internal packet memory and the first external packet memory.
Reservation-Based Resource Configuration for Dynamic Workspaces
Reservations of resources within a dynamic workspace are made and facilitated including by pushing configurations specific to a worker for whom a reservation is made to the resource for a limited time period for the reservation. When the worker arrives for the reservation and checks in, configurations associated with the worker are obtained and pushed to the resource to configure the resource for use by the worker for a time period of the reservation. In some cases, an identity of the worker may be verified to complete the check in process for the reservation. In response to an end time of the reservation being determined, such as based on a scheduled end time for the reservation or based on a determination that the resource has gone unused for some amount of time, the configurations are removed from the resource to prepare the resource for use by a next worker.
Processing of ethernet packets at a programmable integrated circuit
Methods, systems, and computer programs are presented for processing Ethernet packets at a Field Programmable Gate Array (FPGA). One programmable integrated circuit includes: an internal network on chip (iNOC) comprising rows and columns; clusters, coupled to the iNOC, comprising a network access point (NAP) and programmable logic; and an Ethernet controller coupled to the iNOC. When the controller operates in packet mode, each complete inbound Ethernet packet is sent from the controller to one of the NAPs via the iNOC, where two or more NAPs are configurable to receive the complete inbound Ethernet packets from the controller. The controller is configurable to operate in quad segment interface (QSI) mode where each complete inbound Ethernet packet is broken into segments, which are sent from the controller to different NAPs via the iNOC, where two or more NAPs are configurable to receive the complete inbound Ethernet packets from the controller.
Processing of ethernet packets at a programmable integrated circuit
Methods, systems, and computer programs are presented for processing Ethernet packets at a Field Programmable Gate Array (FPGA). One programmable integrated circuit includes: an internal network on chip (iNOC) comprising rows and columns; clusters, coupled to the iNOC, comprising a network access point (NAP) and programmable logic; and an Ethernet controller coupled to the iNOC. When the controller operates in packet mode, each complete inbound Ethernet packet is sent from the controller to one of the NAPs via the iNOC, where two or more NAPs are configurable to receive the complete inbound Ethernet packets from the controller. The controller is configurable to operate in quad segment interface (QSI) mode where each complete inbound Ethernet packet is broken into segments, which are sent from the controller to different NAPs via the iNOC, where two or more NAPs are configurable to receive the complete inbound Ethernet packets from the controller.
Packet buffer spill-over in network devices
Packets to be transmitted from a network device are buffered in queues in a first packet memory. In response to detecting congestion in a queue in the first packet memory, groups of multiple packets are transferred from the first packet memory to a second packet memory, the second packet memory configured to buffer a portion of traffic bandwidth supported by the network device. Prior to transmission of the packets among the one or more groups of multiple packets from the network device, packets among the one or more groups of multiple packets are transferred from the second packet memory back to the first packet memory. The packets transferred from the second packet memory back to the first packet memory are retrieved from the first packet memory and are forwarded to one or more network ports for transmission of the packets from the network device.
Packet buffer spill-over in network devices
Packets to be transmitted from a network device are buffered in queues in a first packet memory. In response to detecting congestion in a queue in the first packet memory, groups of multiple packets are transferred from the first packet memory to a second packet memory, the second packet memory configured to buffer a portion of traffic bandwidth supported by the network device. Prior to transmission of the packets among the one or more groups of multiple packets from the network device, packets among the one or more groups of multiple packets are transferred from the second packet memory back to the first packet memory. The packets transferred from the second packet memory back to the first packet memory are retrieved from the first packet memory and are forwarded to one or more network ports for transmission of the packets from the network device.