Patent classifications
G06F15/17331
CROSS BUS MEMORY MAPPING
A computerized system for efficient interaction between a host, the host having a first operating system, and a second operating system, the system comprising a subsystem on the second operating system which extracts data, directly from a buffer which is local to the host, wherein the system is operative for mapping memory from one bus associated with the first operating system to a different bus, associated with the second operating system and from which different bus the memory is accessed, thereby to emulate a connection between the first and second operating systems by cross-bus memory mapping.
CENTRALIZED ACCESS CONTROL FOR CLOUD RELATIONAL DATABASE MANAGEMENT SYSTEM RESOURCES
Methods for centralized access control for cloud relational database management system resources are performed by systems and devices. The methods utilize a central policy storage, managed externally to database servers, which stores external policies for access to internal database resources at up to fine granularity. Database servers in the processing system each receive external access policies that correspond to users of the system by push or pull operations from the central policy storage, and store the external access policies in a cache of the database servers for databases. For resource access, access conditions are determined via policy engines of database servers based on an external access policy in the cache that corresponds to a user, responsive to a resource access request from a device of the user specifying the internal resource. Data associated with the resource is provided to the user based on the access condition being met.
RoCE over wireless
A wireless device can achieve higher predictability for its transmissions by inserting a placeholder frame in a transmission queue before time sensitive data has been received. In addition, a contention countdown associated with the placeholder frame can start before the time sensitive data is ready for transmission. Once the data is available, the device can insert the data into the payload of the placeholder frame, thereby reducing the wait time before the data can be transmitted wirelessly. Additionally, the device can improve reliability by transmitting data using multiple subcarrier RUs in a channel. The data blocks and the duplicative data can be transmitted in parallel using the subcarrier RUs. If a subset of the subcarrier RUs are blocked because of narrowband interference, the receiving device can nonetheless recover the data blocks and reconstruct the packet from the data transported on the RUs that did not have interference.
Virtualized remote direct memory access
Particular embodiments described herein provide for an electronic device that can be configured to receive a remote direct memory access (RDMA) message from a first virtual machine located on a first network element, determine that the RDMA message is destined for a second virtual machine that is located on the first network element, and use a local direct memory access engine to process the RDMA message, where the local direct memory access engine is located on the first network element. In an example, the electronic device can be further configured to determine that the RDMA message is destined for a third virtual machine on a second network element, wherein the second network element is different than the first network element and use an other device acceleration driver to process the RDMA message instead of the local direct memory access engine.
Distributed raid rebuild
A technique is disclosed for generating rebuild data of a RAID configuration having one or more failed drives. The RAID configuration includes multiple sets of drives coupled to respective computing nodes, and the computing nodes are coupled together via a network. A lead node directs rebuild activities, communicating with the other node or nodes and directing such node(s) to compute partial rebuild results. The partial rebuild results are based on data of the drives of the RAID configuration coupled to the other node(s). The lead node receives the partial rebuild results over the network and computes complete rebuild data based at least in part on the partial rebuild results.
Randomized data distribution in highly parallel database management system
A massively parallel database management system includes an index store and a payload store including a set of storage systems of different temperatures. Both the index store and the storage system each include a list of clusters. Each cluster includes a set of nodes with storage devices forming a group of segments. Nodes and clusters are connected over high speed links. Each cluster receives data and splits the data into data rows based on a predetermined size. The data rows are randomly and evenly distributed between all nodes of the cluster.
TECHNOLOGIES FOR DYNAMIC ACCELERATOR SELECTION
Technologies for dynamic accelerator selection include a compute sled. The compute sled includes a network interface controller to communicate with a remote accelerator of an accelerator sled over a network, where the network interface controller includes a local accelerator and a compute engine. The compute engine is to obtain network telemetry data indicative of a level of bandwidth saturation of the network. The compute engine is also to determine whether to accelerate a function managed by the compute sled. The compute engine is further to determine, in response to a determination to accelerate the function, whether to offload the function to the remote accelerator of the accelerator sled based on the telemetry data. Also the compute engine is to assign, in response a determination not to offload the function to the remote accelerator, the function to the local accelerator of the network interface controller.
Communication method and apparatus
A communication method includes monitoring, by a shared agent, shared memory, wherein the shared memory is used by a first application, wherein the first application runs on a virtual device, wherein the virtual device is located on a host, wherein the shared memory belongs to a part of memory of the host and does not belong to memory specified by the host for the virtual device, and wherein the shared agent is disposed on the host independent of the virtual device, determining, by the shared agent, whether data of the first application is written to the shared memory, reading, by the shared agent, the data from the shared memory and sending the data to a second application in response to the data of the first application is written to the shared memory, wherein the second application is a data sharing party specified by the first application.
METHOD AND SYSTEM FOR FACILITATING LOSSY DROPPING AND ECN MARKING
Methods and systems are provided for performing lossy dropping and ECN marking in a flow-based network. The system can maintain state information of individual packet flows, which can be set up or released dynamically based on injected data. Each flow can be provided with a flow-specific input queue upon arriving at a switch. Packets of a respective flow are acknowledged after reaching the egress point of the network, and the acknowledgement packets are sent back to the ingress point of the flow along the same data path. As a result, each switch can obtain state information of each flow and perform per-flow packet dropping and ECN marking.
RDMA-enabled key-value store
According to one or more embodiments, lookup, insertion, and deletion operations are allowed to continue during actions required for collision remediation. When relocation operations are used to resolve a collision, information encoded in header portions of the hash table entries that store the key-value pairs indicates when the associated key-value pairs are undergoing relocation. This information facilitates continued access to the RKVS during the relocation process by allowing other processes that access the RKVS to handle relocations without failure. Furthermore, when hash table expansion is needed in order to resolve a collision, a second, larger, hash table is allocated, and lookup operations continue on both the old hash table and the new hash table. One or more embodiments further prevent insertion, lookup, and deletion failures in the RKVS using flags, encoded in header information in hash table entries, that reflect the state of the respective key-value pairs in the store.