Patent classifications
H04L41/0803
Method and apparatus for low latency transmissions
A resource assignment can be received. A first set of time-frequency resources in a subframe can be determined from the resource assignment. A second set of time-frequency resources in the subframe can be determined. The second set of time-frequency resources can be used for a second latency data transmission. The second set of time-frequency resources can overlap with at least a portion of the first set of time-frequency resources. A first latency data transmission in the subframe can be decoded based on the determined first and second set of time-frequency resources. The first latency transmission can have a longer latency than the second latency transmission.
Improving performance of multi-processor computer systems
Embodiments of the invention may improve the performance of multi-processor systems in processing information received via a network. For example, some embodiments may enable configuration of a system such that information received via a network may be distributed among multiple processors for efficient processing. A user (e.g., system administrator) may select from among multiple configuration options, each configuration option being associated with a particular mode of processing information received via a network. By selecting a configuration option, the user may specify how information received via the network is processed to capitalize on the system's characteristics, such as by aligning processors on the system with certain NICs. As such, the processor(s) aligned with a NIC may perform networking-related tasks associated with information received by that NIC. If initial alignment causes one or more processors to become over-burdened, processing tasks may be dynamically re-distributed to other processors so as to achieve a more even distribution of the overall processing burden across the system.
Improving performance of multi-processor computer systems
Embodiments of the invention may improve the performance of multi-processor systems in processing information received via a network. For example, some embodiments may enable configuration of a system such that information received via a network may be distributed among multiple processors for efficient processing. A user (e.g., system administrator) may select from among multiple configuration options, each configuration option being associated with a particular mode of processing information received via a network. By selecting a configuration option, the user may specify how information received via the network is processed to capitalize on the system's characteristics, such as by aligning processors on the system with certain NICs. As such, the processor(s) aligned with a NIC may perform networking-related tasks associated with information received by that NIC. If initial alignment causes one or more processors to become over-burdened, processing tasks may be dynamically re-distributed to other processors so as to achieve a more even distribution of the overall processing burden across the system.
User-defined custom storage classes for hybrid-cloud and multicloud data management
A method includes: receiving, by a computing device, input defining a custom storage class in a first dispersed storage network; receiving, by the computing device, input defining a data management rule in the first dispersed storage network; determining, by the computing device, a data object in the first dispersed storage network satisfies the data management rule; and moving, by the computing device and in response to the determining, the data object to a second dispersed storage network according to the custom storage class.
Semiconductor device and information processing method
A semiconductor device capable of performing filter processing while suppressing an increase in processing time is provided. The semiconductor device includes a microcontroller. The microcontroller comprises a CPU, a memory and a CAN-controller. The memory stores software. The CPU executes the software stored in the memory. The CAN controller is configured to add label information to the message information. The CAN routing software stored in the memory implements a filtering function for performing a filter processing for determining whether or not to route the message information by using the label information.
Communication between independent containers
Techniques related to communication between independent containers are provided. In an embodiment, a first programmatic container includes one or more first namespaces in which an application program is executing. A second programmatic container includes one or more second namespaces in which a monitoring agent is executing. The one or more first namespaces are independent of the one or more second namespaces. A monitoring agent process hosts the monitoring agent. The monitoring agent is programmed to receive an identifier of the application program. The monitoring agent is further programmed to switch the monitoring agent process from the one or more second namespaces to the one or more first namespaces. After the switch, the monitoring agent process continues to execute in the second programmatic container, but communication is enabled between the application program and the monitoring agent via the monitoring agent process.
Communication between independent containers
Techniques related to communication between independent containers are provided. In an embodiment, a first programmatic container includes one or more first namespaces in which an application program is executing. A second programmatic container includes one or more second namespaces in which a monitoring agent is executing. The one or more first namespaces are independent of the one or more second namespaces. A monitoring agent process hosts the monitoring agent. The monitoring agent is programmed to receive an identifier of the application program. The monitoring agent is further programmed to switch the monitoring agent process from the one or more second namespaces to the one or more first namespaces. After the switch, the monitoring agent process continues to execute in the second programmatic container, but communication is enabled between the application program and the monitoring agent via the monitoring agent process.
LOCATION-INDEPENDENT PROGRAMMING DATA PLANE FOR PACKET PROCESSING
Techniques are disclosed for efficient communications over a network path between an accelerator of a smart network interface card (smartNIC) and a remote programming data plane of a remote device. In one example, the accelerator receives an instruction to register a pairing between the accelerator and the remote programming data plane, and then stores registration data indicating the pairing. The accelerator then receives from the remote programming data plane a second instruction associated with processing one or more flows. The accelerator then stores instruction data corresponding to the second instruction based on confirming the registered pairing with the remote programming data plane. Subsequently, the accelerator receives a data packet and processes the data packet in accordance with the stored instruction data. In some embodiments, the accelerator may transmit packets to the pair remote programming data plane, for example, requesting further instructions associated with processing a packet.
Endpoint computing device multi-network slice remediation/productivity system
An endpoint computing device multi-network slice remediation/productivity system includes a core network system coupled to a RAN system and configured to allocate network slices and make them available for use in wireless communications via the RAN system. While operating in a pre-boot environment, an endpoint computing device determines that it is unable to transition to operating in a runtime environment and, in response, establishes a remediation network connection with a first network slice, and establishes a productivity network connection with a second network slice. While operating in a pre-boot environment and performing the remediation operations, the endpoint computing device then performs remediation operations via remediation wireless communications over the remediation network connection with the first network slice, and provides access to productivity application(s) that are configured to allow a user to perform productivity operations via productivity wireless communications over the productivity network connection with the second network slice.
Process isolation via application-level blue-green topology
A method for routing requests to a plurality of server clusters is disclosed. The method comprises establishing a first server cluster responding to requests concerning a first software via a first version of that software and to requests concerning a second software via a first version of that software, and a second server cluster responding to requests concerning the first software via a second version of that software and to requests concerning the second software via a second version of that software. A gateway router initially routes requests concerning the first and second software by default to the first server cluster. Upon receiving a request to change default routing of requests, a configuration of the gateway router is updated. Subsequent requests concerning the first software are routed to the second server cluster while subsequent request concerning the second software remain routed to the first server cluster.