Patent classifications
G06F8/447
GENERATING RULES FOR MIGRATING DEPENDENCIES OF A SOFTWARE APPLICATION
Rules can be generated for migrating dependencies of a software application. For example, a computing device can receive a source version of a dependency of a software application and a target version of the dependency of the software application. The computing device can compare the source version to the target version to determine a difference between the source version and the target version. The computing device can receive a template for a rule indicating a location in the source version to be modified for the software application to support the target version. The template can include a fillable section. The computing device can populate the fillable section of the template with a value based on the difference between the source version and the target version.
NPU IMPLEMENTED FOR ARTIFICIAL NEURAL NETWORKS TO PROCESS FUSION OF HETEROGENEOUS DATA RECEIVED FROM HETEROGENEOUS SENSORS
A neural processing unit (NPU) includes a controller including a scheduler, the controller configured to receive from a compiler a machine code of an artificial neural network (ANN) including a fusion ANN, the machine code including data locality information of the fusion ANN, and receive heterogeneous sensor data from a plurality of sensors corresponding to the fusion ANN; at least one processing element configured to perform fusion operations of the fusion ANN including a convolution operation and at least one special function operation; a special function unit (SFU) configured to perform a special function operation of the fusion ANN; and an on-chip memory configured to store operation data of the fusion ANN, wherein the schedular is configured to control the at least one processing element and the on-chip memory such that all operations of the fusion ANN are processed in a predetermined sequence according to the data locality information.
Off-load servers software optimal placement method and program
A software deployment method includes: analyzing a source code of an application; designating off-loadable processes of the application; performing a code conversion of the application according to a deployment destination environment; measuring the performance of the converted application on a verification device; making a setting for resource amounts according to the deployment destination environment; selecting a deployment place by calculating a deployment destination on the basis of a performance and a cost when the converted application is deployed while ensuring the resource amounts; performing, after deployment to an actual environment, a performance measurement test process to measure an actual performance of application; and performing, after performing the performance measurement test process, one or more of performing the code conversion, making the setting for resource amounts, selecting the deployment place, measuring the performance of the application on the verification device, and performing the performance measurement test process.
SELF-DEBUGGING
In overview, methods, computer programs products and devices for securing software are provided. In accordance with the disclosure, a method may comprise attaching a debugger process to a software process. During execution of the software process, operations relevant to the functionality of the code process are carried out within the debugger process. As a result, the debugger process cannot be replaced or subverted without impinging on the functionality of the software process. The software process can therefore be protected from inspection by modified or malicious debugging techniques.
METHOD, APPARATUS, AND COMPUTER-READABLE MEDIUM FOR INTELLIGENT EXECUTION OF A SOLUTION ON A COMPUTER NETWORK
A deployment platform, computer-readable medium, and computer-implemented method for intelligent execution of a solution on a computer network, including receiving an instruction to execute a solution in a local runtime environment on the deployment platform, the solution including solution code written in a solution language, determining, by a helper program on the deployment platform, whether the solution is executable on the deployment platform based on the solution language and either launching, by the helper program, the solution on the deployment platform when the solution is executable on the deployment platform or launching, by the helper program, the solution on a remote platform on the computer network that is configured to execute the solution when the solution is not executable on the deployment platform, the helper program being configured to communicate with the launched solution to enable the launched solution to interface with the local runtime environment on the deployment platform.
METHOD AND SYSTEM TO ENABLE PRINT FUNCTIONALITY IN HIGH-LEVEL SYNTHESIS (HLS) DESIGN PLATFORMS
This disclosure generally relates to high-level synthesis (HLS) platforms, and, more particularly, enable print functionality in high-level synthesis (HLS) platforms. The recent availability FPGA-HLS is a great success due to availability of compilers for FPGAs as opposed to hardware description languages (HDLs) that requires special skills. However, the compilers within the HLS design platform includes limited support for all the standard libraries, wherein features like print functionality is not supported. The invention discloses techniques to enable print functionality in HLS design platforms based on source-to-source transformations and stream combining scheme. In addition to enabling print functionality, the invention also discloses a formatter technique to receive-format FPGA data into human interpretable data.
KERNEL GENERATION FOR NEURAL NETWORKS
Apparatuses, systems, and techniques to automatically generate a reduced number of compute kernels for performing operations of one or more neural networks. In at least one embodiment, one or more operations of one or more neural network graph nodes of the one or more neural network are automatically adjusted to generate an optimized one or more operations that are compiled to generate the reduced number of compute kernels.
Systems and methods for providing a hybrid mobile application
Disclosed are systems and methods for providing a hybrid mobile application. The method may include: receiving specification of at least one feature of a mobile application. The mobile application may be specific to a first programming platform. The method may further include generating the at least one feature based on the specification. The at least one feature may be written in a programming language specific to a second programming platform, and the second programming platform may be different from the first programming platform. The method may also include compiling a programming wrapper based on the at least one feature of the mobile application, combining the at least one feature and the programming wrapper into a programming component, packaging the programming component into the mobile application, and distributing the mobile application in a repository of the first programming platform.
Dynamic code management
Systems and methods for performing dynamic code management, such as dynamic management of JavaScript tags in webpages or code segments in native applications, are disclosed. A user device loading a web or native application can access a factor, such as a user device-specific attribute or a piece of content of the webpage or native application being loaded. That factor can be applied to a rule that is evaluated (e.g., by the user device or a code server) to select one or more desired segments of code (e.g., JavaScript tags or native application code) to be executed by the user device from a pool of available code (e.g., pre-embedded code or dynamically injected code).
Checking source code validity at time of code update
A computer-implemented method includes receiving, by a processor, an updated version of a computer program that includes several source code changes. A compiler preprocesses the source code changes for a target computing platform. The preprocessing includes identifying a compile time condition associated with one or more computer instructions enclosed by a macro, determining a current value of the compile time condition at the time of compiling, and determining corresponding object code based on the current value. Further, a macro information record for the macro is generated that includes the compile time condition, the current value, and an identification of the computer instructions. Further, a linker preprocesses the source code changes. This preprocessing includes determining that the object code is not included in an executable file, and updating the macro information record to indicate that the macro is not included in the executable file.