G06F2209/543

Providing a global queue through replication
11500701 · 2022-11-15 · ·

A global queue service of a provider network allows clients to automatically replicate a queue to another queue of one or more other regions (creating a global queue). A local instance of the service receives a send request from a producer. In response, the local instance stores a message in a local queue and replicates the send request to a remote queue of a remote instance of the global queue service (e.g., at another region). The local instance receives a receive request from a consumer and sends the message to the consumer (without replicating the request to the remote queue). The local instance then receives a delete request from a consumer. In response, the local instance deletes the message from the local queue and replicates the delete request to the remote queue. The local instance also receives replicated requests from the remote instance and in response, applies them locally.

Configuring an electronic device using artificial intelligence

The devices, systems, and methods described herein enable automatically configuring an electronic device using artificial intelligence (AI). The devices, systems, and methods enable accessing telemetry data representing device usage data, inputting the accessed telemetry data into machine learning models that are matched to device metadata, and determining notifications to publish to components of the electronic device. The notifications represent events predicted to occur on the electronic device. The notifications are published to the components of the electronic device such that the electronic device is configured according to the published notifications. The determined notifications enable the identification of optimal settings for the electronic device based on the usage pattern of the device and enable components of the electronic device to preemptively take action on events which are predicted to occur in the future.

SYSTEMS AND METHODS FOR ESTABLISHING AND MANAGING FAST DATA CHANNELS AMONG MODERN WORKSPACES

Systems and methods for establishing and managing fast data channels among modern workspaces are described. In an embodiment, an Information Handling System (IHS) may include a processor and a memory coupled to the processor, the memory having program instructions stored thereon that, upon execution, cause the IHS to: detect access by a first workspace and a second workspace to an IHS resource, and establish a fast data channel between the first and second workspaces, at least in part, based upon context information.

Data processing system and method

A data processing system (10) connected to a plurality of first applications (80) and a plurality of second applications (90) includes a setting information generation function section (20), wherein the setting information generation function section includes a first interface generator (22) configured to generate a first interface for the first applications, a second interface generator (24) configured to generate a second interface for the second applications, and a storage (30) to store common data structure generation source information which is common information based on which the first and second interfaces are generated, and when the common data structure generation source information is updated, the first interface generator and the second interface generator automatically generate the first interface and the second interface, respectively, based on the updated common data structure generation source information.

PROCESS RUNNING METHOD AND APPARATUS

A process running method and apparatus are provided. The method includes: receiving a process startup request; executing an executable file of a target process in a kernel based on the process startup request, to obtain first running function data that supports running of the target process in the kernel; migrating the first running function data to a functional safety partition to obtain second running function data, where the functional safety partition and the kernel are located at a same privilege layer; and running the target process in the functional safety partition based on the second running function data. According to the present disclosure, both the functional safety partition and the kernel are located at a kernel layer. In this way, during process running, a service at the kernel layer can be called without switching from a user layer to the kernel layer.

Processing multi-frame tasks in a multi-threaded digital design system
11645107 · 2023-05-09 · ·

Embodiments are disclosed for processing multi-frame tasks in a multi-threaded digital design system. In particular, in one or more embodiments, the disclosed systems and methods comprise receiving, by a first thread, a first set of inputs on a graphical user interface, determining that the first set of inputs represent a first multi-frame request, wherein a multi-frame request includes a series of inputs that produce a continuous response, generating a first multi-frame edit task, the first multi-frame edit task including a communications channel between the first thread and a second thread for caching the first set of inputs, enqueuing the first multi-frame edit task in a task queue of the second thread, processing, by the second thread, the first multi-frame edit task.

Multicast message filtering in virtual environments

Various systems, processes, and products may be used to filter multicast messages in virtual environments. In particular implementations, a system, process, and product for filtering multicast messages in virtual environments may include the ability to determine whether a multicast message has been received, examine a local filtering store for a match for the destination address of the multicast message, and send the message to a number of virtual machines if a match for the address is found. The system, process, and product may also include the ability to analyze a filtering store of at least one of the virtual machines for a match for the address if a match is not found in the local filtering store and send the message to a number of the virtual machines if a match for the address is found in a filtering store of one of the virtual machines.

Framework to provide time bound execution of co-processor commands

When a main processor issues a command to co-processor, a timeout value is included in the command. As the co-processor attempts to execute the command, it is determined whether the attempt is taking time beyond what is permitted by the timeout value. If the timeout is exceeded then responsive action is taken, such as the generation of a command timeout type failure message. The receipt of the command with the timeout value, and the consequent determination of a timeout condition for the command, may be determined by: the co-processor that receives the command, or a watchdog timer that is separate from the co-processor. Also, detection of co-processor hang and/or hung co-processor conditions during the time that a co-processor is executing a command for the main processor.

Application dehydration and rehydration during application-to-application calls

Methods, systems, and computer program products are provided that enable a first application (i.e., a caller application) to call a second application (i.e., a callee application) to perform a function in a manner such that the first application may be dehydrated during the call. Dehydrating includes terminating execution of an instance of the first application, and freeing memory space in a memory of a mobile device that stored the instance of the first application. In such case, the second application may be active while the first application is no longer present in memory. The second application is enabled to respond to the call, causing the first application to be rehydrated. The first application continues execution at a location where the first application was dehydrated, and receives the response to the call.

INTER-NODE MESSAGING CONTROLLER

A processor package comprises a first core, a local cache in the first core, and an inter-node messaging controller (INMC) in the first core. The INMC is to receive an inter-node message from a sender thread executing on the first core, wherein the message is directed to a receiver thread executing on a second core. In response, the INMC is to store a payload from the inter-node message in a local message queue in the local cache of the first core. After storing the payload, the INMC is to use a remote atomic operation to reserve a location at a tail of a shared message queue in a local cache of the second core. After reserving the location, the INMC is to use an inter-node-put operation to write the payload directly to the local cache of the second core. Other embodiments are described and claimed.