H04L47/765

INCREASING AN EFFICIENCY OF A FILE TRANSFER BY USING MULTIPLE COMMUNICATION RESOURCES

A device may receive an indication to establish a communication session between a first device and a second device for a file transfer. The device may receive, from the second device, metadata. The metadata may be used to identify files of a set of files. The device may identify the files for transfer from the second device to the first device based on the metadata. The device may cause multiple communication sessions to be established between the first device and the second device. The multiple connections may be used to transfer the files from the second device to the first device. The device may receive, from the second device, the files. The files may be received via the multiple communication sessions.

Techniques to support multiple interconnect protocols for a common set of interconnect connectors

Embodiments may be generally direct to apparatuses, systems, method, and techniques to determine a configuration for a plurality of connectors, the configuration to associate a first interconnect protocol with a first subset of the plurality of connectors and a second interconnect protocol with a second subset of the plurality of connectors, the first interconnect protocol and the second interconnect protocol are different interconnect protocols and each comprising one of a serial link protocol, a coherent link protocol, and an accelerator link protocol, cause processing of data for communication via the first subset of the plurality of connectors in accordance with the first interconnect protocol, and cause processing of data for communication via the second subset of the plurality of connector in accordance with the second interconnect protocol.

Method and apparatus for dynamically deploying software agents

Methods and apparatus for dynamically deploying software agents are disclosed. For example, a user of an electronic record management system may deploy a plurality of different background processes (e.g., OCR, dedup, etc.) that may each take several hours to complete. Subsequently, the user may decide to change the number of servers dedicated to one or more background process. In addition, the user may decide to deploy additional background processes. Neither of these user actions interrupts any of the background processes.

SCALABLE EDGE COMPUTING

There is disclosed in one example an application-specific integrated circuit (ASIC), including: an artificial intelligence (Al) circuit; and circuitry to: identify a flow, the flow including traffic diverted from a core cloud service of a network to be serviced by an edge node closer to an edge of the network than to the core of the network; receive telemetry related to the flow, the telemetry including fine-grained and flow-level network monitoring data for the flow; operate the Al circuit to predict, from the telemetry, a future service-level demand for the edge node; and cause a service parameter of the edge node to be tuned according to the prediction.

Network system, topology management method, and program

A network system includes a storage unit that stores topology information in a graph database in chronological order, the topology information including end point information and connection information of logical resources of the virtual network, end point information and connection information of physical resources of the physical network, and correspondence information between the logical resources and the physical resources; a setting information acquirer that acquires setting information of the virtual network from the orchestrator using a change notification from the orchestrator as a trigger; and a topology information management unit that updates the topology information stored in the graph database of the storage unit on the basis of the acquired setting information.

Network system, topology management method, and program

A network system includes a storage unit that stores topology information in a graph database in chronological order, the topology information including end point information and connection information of logical resources of the virtual network, end point information and connection information of physical resources of the physical network, and correspondence information between the logical resources and the physical resources; a setting information acquirer that acquires setting information of the virtual network from the orchestrator using a change notification from the orchestrator as a trigger; and a topology information management unit that updates the topology information stored in the graph database of the storage unit on the basis of the acquired setting information.

Scalable edge computing

There is disclosed in one example an application-specific integrated circuit (ASIC), including: an artificial intelligence (Al) circuit; and circuitry to: identify a flow, the flow including traffic diverted from a core cloud service of a network to be serviced by an edge node closer to an edge of the network than to the core of the network; receive telemetry related to the flow, the telemetry including fine-grained and flow-level network monitoring data for the flow; operate the Al circuit to predict, from the telemetry, a future service-level demand for the edge node; and cause a service parameter of the edge node to be tuned according to the prediction.

Scalable edge computing

There is disclosed in one example an application-specific integrated circuit (ASIC), including: an artificial intelligence (Al) circuit; and circuitry to: identify a flow, the flow including traffic diverted from a core cloud service of a network to be serviced by an edge node closer to an edge of the network than to the core of the network; receive telemetry related to the flow, the telemetry including fine-grained and flow-level network monitoring data for the flow; operate the Al circuit to predict, from the telemetry, a future service-level demand for the edge node; and cause a service parameter of the edge node to be tuned according to the prediction.

Method and Apparatus for Controlling Data Transmission

A method and apparatus for controlling data transmission includes: obtaining an amount of data, the data requested by a requesting server, an instructed deadline, and candidate source servers storing the requested data; obtaining, based on the amount of data, a minimum bandwidth required for transmitting the requested data before the deadline as a required bandwidth corresponding to the requested data; and selecting one of the candidate source servers as a source server for transmitting the requested data, based on the required bandwidth, bandwidth resources of the candidate source servers and the requesting server, and remaining bandwidth of links between the candidate source servers and the requesting server. This implementation implements the selection of the most appropriate source server from the holistic perspective, ensures that data transmission can be completed before the deadline, and effectively utilize the link bandwidth.

CLOUD TRANSLATION MECHANISM
20210399994 · 2021-12-23 ·

A system to facilitate communication between a cloud system and an edge server is described. The system includes one or more processors to execute a translation service to receive a first request operation from a cloud service, wherein the first request operation comprises a first communication protocol supported by the cloud service, translate the first request operation to a second communication protocol supported by a first edge server, wherein the second communication protocol is different from the first communication protocol and transmit the translated first request operation to the first edge server.