H04L67/1004

ROUTING APPLICATION CONTROL AND DATA-PLANE TRAFFIC IN SUPPORT OF CLOUD-NATIVE APPLICATIONS

Techniques for using computer networking protocol extensions to route control-plane traffic and data-plane traffic associated with a common application are described herein. For instance, a traffic flow associated with an application may be established such that control-plane traffic is sent to a control-plane node associated with the application and data-plane traffic is sent to a data-plane node associated with the application. When a client device sends an authentication request to connect to the application, the control-plane node may send an indication of a hostname to be used by the client device to send data-plane traffic to the data-node. As such, when a packet including the hostname corresponding with the data-plane node is received, the packet may be forwarded to the data-plane node.

SERVICE ROAMING BETWEEN EDGE COMPUTING PLATFORMS

Systems and techniques for service roaming between edge computing platforms are described herein. A service executing on a first edge computing platform may be identified to be migrated to a second edge computing platform. A first service component may be determined that is being executed by the first edge computing platform. Transmission of the service to the second edge platform may be initiated to execute a second service component for execution of the service.

SERVICE ROAMING BETWEEN EDGE COMPUTING PLATFORMS

Systems and techniques for service roaming between edge computing platforms are described herein. A service executing on a first edge computing platform may be identified to be migrated to a second edge computing platform. A first service component may be determined that is being executed by the first edge computing platform. Transmission of the service to the second edge platform may be initiated to execute a second service component for execution of the service.

AUTO-DOCUMENTATION FOR APPLICATION PROGRAM INTERFACES BASED ON NETWORK REQUESTS AND RESPONSES
20230224359 · 2023-07-13 ·

Disclosed embodiments are directed at systems, methods, and architecture for providing auto-documentation to APIs. The auto documentation plugin is architecturally placed between an API and a client thereof and parses API requests and responses in order to generate auto-documentation. In some embodiments, the auto-documentation plugin is used to update preexisting documentation after updates. In some embodiments, the auto-documentation plugin accesses an on-line documentation repository. In some embodiments, the auto-documentation plugin makes use of a machine learning model to determine how and which portions of an existing documentation file to update.

AUTO-DOCUMENTATION FOR APPLICATION PROGRAM INTERFACES BASED ON NETWORK REQUESTS AND RESPONSES
20230224359 · 2023-07-13 ·

Disclosed embodiments are directed at systems, methods, and architecture for providing auto-documentation to APIs. The auto documentation plugin is architecturally placed between an API and a client thereof and parses API requests and responses in order to generate auto-documentation. In some embodiments, the auto-documentation plugin is used to update preexisting documentation after updates. In some embodiments, the auto-documentation plugin accesses an on-line documentation repository. In some embodiments, the auto-documentation plugin makes use of a machine learning model to determine how and which portions of an existing documentation file to update.

Multi-access edge computing low latency information services

A multi-access edge computing (MEC) platform may receive an indication that a user device has downloaded a MEC application client associated with a MEC application and may send, to the user device, instructions to install a device client. The device client may transmit device information associated with the user device to the MEC platform. The MEC platform may receive the device information associated with the user device and determine, based on the received device information, performance information associated with the MEC application.

CONTENT DELIVERY SYSTEMS AND METHODS

Aspects of the present disclosure involve a CDN utilizing a Direct Server Return (DSR) request command to create a tunnel from one content server to another content server of the CDN. Through the DSR tunnel, a request for content may be served from a content server that is more advantageous for the CDN for any number of reasons, such as from a content server that is logically closer to the content server. In determining when a DSR tunnel is created to provide the content, the CDN may access a database storing various information concerning the CDN, such as the location of egress gateways of the CDN in relation to one or more content servers, connecting network location information, the capabilities of one or more content servers, and/or load information of available content servers.

Multi-threaded processing of search responses
11695830 · 2023-07-04 · ·

Multi-threaded processing of search responses returned by search peers is disclosed. An example method may include transmitting, by a computer system, a search request to a plurality of search peers of a data aggregation and analysis system; receiving, by a first processing thread, a plurality of data packets from the plurality of search peers; parsing, by a second processing thread, one or more data packets of the plurality of data packets to produce a first partial response to the search request; parsing, by a third processing thread, the one or more data packets to produce a second partial response to the search request; and generating, based on the first partial response and the second partial response, an aggregated response to the search request.

Multi-threaded processing of search responses
11695830 · 2023-07-04 · ·

Multi-threaded processing of search responses returned by search peers is disclosed. An example method may include transmitting, by a computer system, a search request to a plurality of search peers of a data aggregation and analysis system; receiving, by a first processing thread, a plurality of data packets from the plurality of search peers; parsing, by a second processing thread, one or more data packets of the plurality of data packets to produce a first partial response to the search request; parsing, by a third processing thread, the one or more data packets to produce a second partial response to the search request; and generating, based on the first partial response and the second partial response, an aggregated response to the search request.

Technologies for assigning workloads to balance multiple resource allocation objectives

Technologies for allocating resources of managed nodes to workloads to balance multiple resource allocation objectives include an orchestrator server to receive resource allocation objective data indicative of multiple resource allocation objectives to be satisfied. The orchestrator server is additionally to determine an initial assignment of a set of workloads among the managed nodes and receive telemetry data from the managed nodes. The orchestrator server is further to determine, as a function of the telemetry data and the resource allocation objective data, an adjustment to the assignment of the workloads to increase an achievement of at least one of the resource allocation objectives without decreasing an achievement of another of the resource allocation objectives, and apply the adjustments to the assignments of the workloads among the managed nodes as the workloads are performed. Other embodiments are also described and claimed.