Patent classifications
G06F9/546
WRITE ORDER PRESERVING DATA STREAM CONSUMPTION
A system, a method, and a computer program product for scalable processes for write-order preserving data stream consumption. A data partition in a plurality of data partitions of data stream is selected based on a request received from a client processing node. The plurality of data partitions are distributed among a plurality of broker nodes in a distributed messaging system. A broker node in the plurality of broker nodes hosting the selected data partition is identified. The identified broker node provides the selected data partition to the client processing node for performing at least one function.
Message Retargeting In Machine-to-Machine Service Layer Communications
Two methods are described to enable the communication between two service layer entities that do not have registration relations. In the first method, service layer entities employ a retargeting table that stores the service layer entity to retarget the message in order to deliver the message to the destination. In a second method, an infrastructure node in a service provider employs a retargeting chain table that stores a chain of entities that can retarget the message to the destination.
Issue tracking system having temporary notification suppression corresponding to group activity
An issue tracking system configured to track issues, tickets, or tasks is described herein. The issue tracking system may be used to create, modify, and delete issue records during a live meeting or event. The issue tracking system may be adapted to determine a likelihood of user attendance or user proximity and, in accordance with the likelihood exceeding a threshold or meeting a criteria, suppress transmission of notifications to a subset of user devices. Normal or unsuppressed notification transmission is maintained to another subset of user devices.
METHODS AND APPARATUS TO FACILITATE READ-MODIFY-WRITE SUPPORT IN A COHERENT VICTIM CACHE WITH PARALLEL DATA PATHS
Methods, apparatus, systems and articles of manufacture are disclosed facilitate read-modify-write support in a coherent victim cache with parallel data paths. An example apparatus includes a random-access memory configured to be coupled to a central processing unit via a first interface and a second interface, the random-access memory configured to obtain a read request indicating a first address to read via a snoop interface, an address encoder coupled to the random-access memory, the address encoder to, when the random-access memory indicates a hit of the read request, generate a second address corresponding to a victim cache based on the first address, and a multiplexer coupled to the victim cache to transmit a response including data obtained from the second address of the victim cache.
AGGRESSIVE WRITE FLUSH SCHEME FOR A VICTIM CACHE
A caching system including a first sub-cache and a second sub-cache in parallel with the first sub-cache, wherein the second sub-cache includes: line type bits configured to store an indication that a corresponding cache line of the second sub-cache is configured to store write-miss data, and an eviction controller configured to evict a cache line of the second sub-cache storing write-miss data based on an indication that the cache line has been fully written.
Distributed Computing based on Memory as a Service
Systems, methods and apparatuses of distributed computing based on memory as a service are described. For example, a set of networked computing devices can each be configured to execute an application that accesses memory using a virtual memory address region. Each respective device can map the virtual memory address region to the local memory for a first period of time during which the application is being executed in the respective device, map the virtual memory address region to a local memory of a remote device in the group for a second period of time after starting the application in the respective device and before terminating the application in the respective device, and request the remote device to process data in the virtual memory address region during at least the second period of time.
CARD SHARING METHOD, ELECTRONIC DEVICE, AND COMMUNICATIONS SYSTEM
This application provides a card sharing method, an electronic device, and a communications system. The method includes: displaying an interface of a first application in response to a first operation of a user, where a first card is displayed on the interface of the first application, the first card includes a first card module and a second card module, and interface resources of the first card are provided by a second application; sending a second message to a third application in response to a second operation of the user, where the second operation is associated with the first card module, and the second message includes an identifier of the first card module; and displaying an interface of the third application, where the first card module is displayed on the interface of the third application, and interface resources of the first card module are provided by a fourth application.
Message and user profile indications in speech-based systems
A speech-based system utilizes a speech interface device located in the home of a user. The system may interact with different users based on different user profiles. The system may include messaging services that generate and/or provide messages to the user through the speech interface device. The speech interface device may have indicators that are capable of being illuminated in different colors. To notify a user regarding the currently active user profile, each user profile is associated with a different color and the color of the active profile is displayed on the speech interface device when the user is interacting with the system. To notify the user regarding awaiting messages, different types of messages are associated with different colors and the colors of the message types of waiting messages are displayed on the speech interface whenever the user is not interacting with the system.
Method of dynamically configuring FPGA and network security device
Provided are a method of dynamically configuring a FPGA and a network security device. The network security device includes a CPU and at least one FPGA coupled with the CPU. The CPU generates a configuration entry for a target FPGA in response to a user instruction. The configuration entry includes a classification number and a configuration content for the target FPGA. The CPU sends the configuration entry to each FPGA coupled with the CPU, Each FPGA obtains its own classification number, compares its own classification number with the classification number in the configuration entry, and stores the configuration content when the own classification number the same with the classification number in the configuration entry.
Control registers to store thread identifiers for threaded loop execution in a self-scheduling reconfigurable computing fabric
Representative apparatus, method, and system embodiments are disclosed for configurable computing. A representative system includes an interconnection network; a processor; and a plurality of configurable circuit clusters. Each configurable circuit cluster includes a plurality of configurable circuits arranged in an array; a synchronous network coupled to each configurable circuit of the array; and an asynchronous packet network coupled to each configurable circuit of the array. A representative configurable circuit includes a configurable computation circuit and a configuration memory having a first, instruction memory storing a plurality of data path configuration instructions to configure a data path of the configurable computation circuit; and a second, instruction and instruction index memory storing a plurality of spoke instructions and data path configuration instruction indices for selection of a master synchronous input, a current data path configuration instruction, and a next data path configuration instruction for a next configurable computation circuit.