Patent classifications
G06F8/41
SOFTWARE APPLICATION DEPLOYMENT
Methods, computer program products, and systems can include obtaining a first computing environment specific application deployment software code instance associated to a first computing environment, the first computing environment specific application deployment software code instance for deployment of a certain application on the first computing environment; parsing the first computing environment specific application deployment software code instance, wherein the parsing includes determining attributes of the first computing environment specific application deployment software code instance and generating, using the determined attributes of the first computing environment specific application deployment software code instance, a computing environment agnostic semantic tree data structure that expresses a workflow pattern for deployment of the certain application; and composing, with use of the computing environment agnostic semantic tree data structure, a second computing environment specific application deployment software code instance associated to a second computing environment for deployment of the certain application on the second computing environment.
Two-Phase Application Development Device
A mobile application development device having a platform processor, a native application converter engine, and a mobile platform framework engine configured to facilitate the development and deployment of mobile applications configured to be run on different mobile operating systems from code that is developed independently and agnostic of the mobile operating system on which it will ultimately run.
Cognitve Automation-Based Engine to Propagate Data Across Systems
Aspects of the disclosure relate to cognitive automation-based engine processing to propagate data across multiple systems via a private network to overcome technical system, resource consumption, and architecture limitations. Data to be propagated can be manually input or extracted from a digital file. The data can be parsed by analyzing for correct syntax, normalized into first through sixth normal forms, segmented into packets for efficient data transmission, validated to ensure that the data satisfies defined formats and input criteria, and distributed into a plurality of data stores coupled to the private network, thereby propagating data without repetitive manual entry. The data may also be enriched by, for example, correcting for any errors or linking with other potentially related data. Based on data enrichment, recommendations of additional target(s) for propagation of data can be identified. Reports may also be generated. The cognitive automation may be performed in real-time to expedite processing.
MULTI-LINGUAL CODE GENERATION WITH ZERO-SHOT INFERENCE
A neural transformer model with attention is trained to predict candidates to complete a line of source code with a zero-inference capability. The model is trained on an unsupervised training dataset that includes features from source code written in multiple programming languages. The features include a file-level context and a local context, where the file-level context includes a global context, a class context, a function context, and/or a method context for each class, function and/or method of the source code programs used in the training dataset. The local context includes method bodies, function bodies, and/or stand-alone code of main method routines. From these features, the model is able to learn to predict an ordered sequence of code elements that complete a line of source code in a programming language seen and not seen during training.
PROGRAM GENERATION APPARATUS, PROGRAM GENERATION METHOD AND PROGRAM
A program generation apparatus includes a generation unit that inputs a specification of a program to be generated described in natural language into a model trained on a relationship between a specification of a program described in natural language and the program to generate a first program, and a change unit that changes the first program to generate a second program satisfying a set of one or more input values and output values, and thus the possibility of a desired program being automatically generated can be increased.
OFFLOAD SERVER, OFFLOAD CONTROL METHOD, AND OFFLOAD PROGRAM
An offload server includes: an application code analysis section configured to analyze source code of an application; a PLD processing designation section configured to identify loop statements of the application, generate a plurality of offload processing patterns designating, for each of the identified loop statements, parallel processing and/or pipeline processing on the PLD according to OpenCL, and perform compilation; an arithmetic intensity calculation section configured to calculate arithmetic intensities of the loop statements of the application; and a PLD processing pattern generation section configured to, on the basis of the arithmetic intensities calculated by the arithmetic intensity calculation section, narrow down the loop statements to, as candidates to be offloaded, those having an arithmetic intensity higher than a predetermined threshold.
Extensible platform for orchestration of data using probes
In a computer system, an orchestration platform includes extensible components that interact with external systems and technology. The platform extension deploys a surrogate component or probe that acts as a bridge between the core platform and the extension technology.
METHOD AND SYSTEM FOR SOFTWARE ENHANCEMENT AND MANAGEMENT
A software enhancement and management system (E&M System) can include two ways to decompose existing software such that new functionality can be added: functional decomposition and time-affecting linear pathway (TALP) decomposition. Functional decomposition captures the inputs and outputs of the existing software's functions and attaches the new algorithmic constructs presented as other functions that receive the outputs of the existing software's functions. TALP decomposition allows for the generation of time-prediction polynomials that approximate time-complexity functions, speedup, and automatic dynamic loop-unrolling-based parallelization for each TALP.
Reduced memory consumption of compiler-transformed asynchronous methods
An asynchronous method is implemented in a manner that reduces the amount of runtime overhead needed to execute the asynchronous method. The data elements needed to suspend an asynchronous method to await completion of an asynchronous operation, to resume the asynchronous method at a resumption point, and to provide a completion status of the caller of the asynchronous method are consolidated into one or two reusable objects. An asynchronous method may be associated with a distinct object pool of reusable objects. The size of a pool and the total size of all pools can be configured statically or dynamically based on runtime conditions.
Register sharing mechanism to equally allocate disabled thread registers to active threads
An apparatus is disclosed. The apparatus includes one or more processors comprising register sharing circuitry to receive meta-information indicating a number of threads that are to be disabled and provide an indication that an associated thread is disabled, a plurality of General Purpose Register Files (GRFs), wherein one or more of the plurality of GRFs is associated with one of the plurality of threads and a plurality of multiplexers coupled to the one or more GRFs to receive the indication from the register sharing circuitry and disable thread access to an associated GRF based on an indication that a thread is to be disabled.