G06F9/45

System and method for pre-operating system memory map management to minimize operating system failures

A method includes booting an information handling system, providing by an EFI of the information handling system a memory segment for a first EFI type memory access, reserving a first portion of the segment from access by an operating system of the information handling system, determining a size of the first portion, determining a size of a second portion of the segment based upon the size of the first portion, allocating a third portion of the segment for the first EFI type memory access, the third portion including the first portion and the second portion, and passing a memory map to the operating system, the memory map including the third portion, wherein the third portion is reserved from access by the operating system.

Compiling graph-based program specifications

A graph-based program specification includes: a plurality of components, each corresponding to a processing task and including one or more ports, including scalar data ports for sending or receiving a single data element and collection data ports for sending or receiving a collection of multiple data elements; and one or more links, each connecting an output port of an upstream component to an input port of a downstream component. Prepared code is generated representing subsets of the plurality of components, including: identifying one or more subset boundaries, including identifying one or more links connecting a collection data port of a component to a scalar data port of a component; forming the subsets based on the identified subset boundaries; and generating prepared code for each formed subset that when used for execution by a runtime system causes processing tasks corresponding to the components in each formed subset to be performed.

Dynamic provision of debuggable program code
09841960 · 2017-12-12 · ·

According to one aspect of the present disclosure, a method comprises receiving a command to load first compiled program code for execution by a processor. The first compiled program code is decompiled to generate source code. The source code is compiled to generate second compiled program code, the second compiled program code comprising information associated with the source code. The second compiled program code is provided to a debugger.

Pattern matching techniques in analog and mixed signal circuits
09830414 · 2017-11-28 · ·

A system and method for adding hierarchy to a netlist. A netlist is received and converted into a connected graph. Location parameters for the nodes of the connected graph are mapped onto the connected graph. Landmark structures are identified in the connected graph, wherein identifying includes recording a location associated with each landmark structure. Patterns are searched for in the connected graph, wherein searching proceeds outward from an anchor defined by the location of each of the identified landmark structures.

Generating object code from intermediate code that includes hierarchical sub-routine information
09830134 · 2017-11-28 · ·

Examples are described for a device to receive intermediate code that was generated from compiling source code of an application. The intermediate code includes information generated from the compiling that identifies a hierarchical structure of lower level sub-routines in higher level sub-routines, and the lower level sub-routines are defined in the source code of the application to execute more frequently than the higher level sub-routines that identify the lower level sub-routines. The device is configured to compile the intermediate code to generate object code based on the information that identifies lower level sub-routines in higher level sub-routines, and store the object code.

Remote procedure calls in a replicated server system
09830147 · 2017-11-28 · ·

A system for command handling for replicated repositories includes an input interface and a processor. The input interface is to receive a command. The processor is to determine a set of servers to receive the command, determine whether responses match for the command for each server of the set of server, and in the event the responses do not match, determine whether the responses are required to match for the command and in the event the responses are not required to match for the command, determine a unified response.

Automatic management of low latency computational capacity

A system for providing automatic management of low latency computational capacity is provided. The system may be configured to maintain a plurality of virtual machine instances. The system may be further configured to identify a trend in incoming code execution requests to execute program code on a virtual compute system, determine, based on the identified trend, that the plurality of virtual machine instances should be adjusted, and adjust the plurality of virtual machine instances based on the identified trend.

Methods and apparatus for automatic communication optimizations in a compiler based on a polyhedral representation

Methods, apparatus and computer software product for source code optimization are provided. In an exemplary embodiment, a first custom computing apparatus is used to optimize the execution of source code on a second computing apparatus. In this embodiment, the first custom computing apparatus contains a memory, a storage medium and at least one processor with at least one multi-stage execution unit. The second computing apparatus contains at least one local memory unit that allows for data reuse opportunities. The first custom computing apparatus optimizes the code for reduced communication execution on the second computing apparatus. This Abstract is provided for the sole purpose of complying with the Abstract requirement rules. This Abstract is submitted with the explicit understanding that it will not be used to interpret or to limit the scope or the meaning of the claims.

Asynchronous parallel processing of log data
09830368 · 2017-11-28 · ·

Techniques to process machine generated log data are disclosed. In various embodiments, a parser definition associated with a set of log data is determined. The parser definition is compiled to create an instance of a parser to parse the set of log data. The parser has a hierarchical structure comprising a plurality of hierarchically related nodes, each of at least a subset of said nodes having associated therewith one or more actors each configured to parse data associated with that node. At least a portion of the set of log data is sent to the parser instance prior to compilation of said parser instance being completed. A first node of the parser instance is configured to receive and parse log data associated with the first node even if compilation of the parser definition has not been completed with respect to a second node of said parser instance.

Proxy simulation service for testing stateless systems

An enhanced proxy simulation service can be used to test a complex stateless system, such as a fulfillment network, that depends on upstream and downstream processes in order to operate. The enhanced proxy simulation service can be configured to send an event, such as an order request, so as to appear to be sent from a process upstream of the system; and can be configured to receive a request from the system, such as a request to charge payment information, so as to appear as a downstream process of the system. The proxy simulation service can store events received from the system and apply a set of rules to the stored events to construct a simulated response to the last event received from the system. The simulated response can then be sent to the system being tested.