Patent classifications
H04L12/835
Network node for controlling transport of data in a wireless communication network
A first network node, for supporting transport of data in a wireless communication network. One or more first data transmission connections, for first data flows, are established between the first network node and one or more service providing entities. One or more second data transmission connections, for second data flows associated with the first data flows, are established between the first network node and one or more wireless terminals via a target node. The first network node obtains a buffer measure that is indicative of a buffering degree at the target node for a first group of the second data flows. The first network node controls, based on the obtained buffer measure, a transmission pace between the first network node and the target node for a second group of the second data flows.
APPARATUS AND METHOD FOR ROUTING DATA IN A SWITCH
Apparatuses, methods and storage medium associated with routing data in a switch are provided. In embodiments, the switch may include route lookup circuitry determine a first set of output ports that are available to send a data packet to a destination node. The lookup circuitry may further select, based on respective congestion levels associated with the first set of output ports, a plurality of output ports for a second set of output ports from the first set of output ports. An input queue of the switch may buffer the data packet and route information associated with the second set of output ports. The switch may further include route selection circuitry to select a destination output port from the second set of output ports, based on updated congestion levels associated with the output ports of the second set of output ports. Other embodiments may be described and/or claimed.
Method and system for flow control between a base station controller and a base transceiver station
According to a disclosed embodiment, a flow indication counter is incremented each time a data packet is transmitted from a buffer. When the number of data packets transmitted equals or exceeds a threshold number, a flow indication message comprising the buffer window size is generated and transmitted to the base station controller. Further, flow indication messages can be generated and transmitted every threshold time interval, independently of the number of data packets transmitted to ensure that flow indication messages are sent at least every preset time interval. Moreover, a system for flow control can be constructed comprising a flow indication counter module configured to provide an updated number of data packets transmitted. The system further comprises a window size monitoring module which determines the buffer window size and a message generating module which generates a flow indication message comprising the buffer window size.
Buffer control for multi-transport architectures
A system and method for automating connection management in a manner that may be transparent to any actively communicating applications operating in a Network on Terminal Architecture (NoTA). An application level entity may access another node by making a request to a high level communication structure via an interface. The high level structure may interact with a lower level structure configured to manage communication by establishing communication with another device via one or more transports. In at least one embodiment, provisions may be made to guard against data being lost when a transport fails, including storing data that is passed from a transport-independent buffer to a transport-specific buffer in case the transport fails. When a failure occurs, the stored data may readily be forwarded for sending using another transport.
Methods for an automatic scaling of data consumers and apparatuses using the same
The invention introduces a method for an automatic scaling of data consumers, performed by a processing unit of a data-consumer server, which contains at least the following steps. State-update requests are received from data consumers, where each state-update request includes a busy state or an idle state and the busy state indicates that the data consumer reads data from a queue and processes the read data. A first quantity of data consumers, a second quantity of busy states and a third quantity of idle states are calculated. A fourth quantity of data consumers to be produced is estimated according to the first, second and third quantities. Data consumers of the fourth quantity are produced.
Offload of data transfer statistics from a mobile access gateway
In general, techniques are described for offloading data transfer statistics from a mobile access gateway. The mobile access gateway comprises a forwarding unit. The forwarding unit comprises a packet forwarding engine (PFE). When the PFE receives a packet, the PFE updates a data transfer statistic based on a quantity of data in the packet. The data transfer statistic is initially stored in a memory of the PFE. The PFE is configured to push the data transfer statistic from the memory of the PFE to a memory of the forwarding unit.
NETWORK NODE AND METHOD FOR HANDLING A PROCESS OF CONTROLLING A DATA TRANSFER RELATED TO VIDEO DATA OF A VIDEO STREAMING SERVICE
A method in a network node relating to a process of controlling a data transfer related to video data of a video streaming service from a server to a wireless device is provided. The network node and wireless device operates in a wireless communications network. The network node determines a scheduling weight value for the wireless device to be used in the data transfer based on a target rate scheduling weight value and a proportional rate fair weight value. The network node then determines a size of data segment to be used in the data transfer based on at least part of the scheduling weight value. The network node further determines a pending data volume for the transferring of the video data to a play back buffer of the wireless device based on at least part of the scheduling weight value.
PROCESSING HIGH VOLUME NETWORK DATA
Disclosed are a system comprising a computer-readable storage medium storing at least one program, and a computer-implemented method for event messaging over a network. A subscription interface receives data indicative of a subscription request for sessionized data. An allocation module allocates a sessionizer bank linked to the subscription request. A messaging interface module provisions identifiers linked to the respective processing engines of the sessionizer bank. The messaging interface module registers the allocated sessionizer bank as available to process event messages matching the subscription request by providing the provisioned identifiers. The messaging interface module receives event messages from a producer device linked by a collection server to a selected one of the processing engines of the sessionizer bank. The selected one of the processing engine processes the received event messages in accordance with session rule data linked to the subscription request to generate sessionized data.
Data transfer device and data transfer system using adaptive compression algorithm
A data transfer device compresses and transfers data according to a priority given to a CPU-constraint process imposing a constraint to a compression processing speed over a NW bandwidth-constraint process imposing a constraint to a transfer processing speed. It is necessary to select a compression algorithm, applied to the CPU-constraint process or the NW bandwidth-constraint process, based on a NW bandwidth, compressibility, and compression processing speed maximizing an effective throughput. When the amount of compressed data held in a temporary hold part is smaller than the predetermined value, the compressed data of the NW bandwidth-constraint process is stored in a temporary hold part. When the amount of compressed data held by the temporary hold part is larger than the predetermined value, the compressed data of the CPU-constraint process is stored in the temporary hold part. Thus, it is possible to improve an effective throughput by effectively using NW bandwidths.
Message processing using dynamic load balancing queues in a messaging system
A system, method, and computer-readable medium are disclosed for dynamically managing message queues to balance processing loads in a message-oriented middleware environment. A first source message associated with a first target is received, followed by generating a first dynamic load balancing message queue when a first message queue associated with the first target is determined to not be optimal. The first dynamic load balancing message queue is then associated with the first target, followed by enqueueing the first source message to the first dynamic load balancing message queue for processing by the first target.