Patent classifications
G06F11/3698
Program lifecycle testing
In one embodiment, a system for program lifecycle testing includes receiving a request to test a program update at an interface. Using a processor, the system may then execute a validation test associated with the program update, wherein the validation test is conducted in a testing environment comprising a plurality of testing environment systems. The system may then use the processor to capture a current state of the testing environment at a start of the validation test, and confirm that the plurality of testing environment systems are operating according to the validation test. The system may then use the interface to receive testing results from the validation test and compare the testing results to previous test results from a prior program update. The system may then store the validation test results, the current state of the testing environment, and a name of the program update, in a performance database.
Systems and methods for unit testing of functions on remote kernels
The disclosed computer-implemented method may include (1) providing a framework that includes (A) a user-space component that runs at a client site and (B) a kernel-space component that runs at a remote site, (2) identifying attributes of objects that reside at the remote site and whose addresses are unknown at the client site, (3) generating a script to test a function of a kernel running on the remote site based at least in part on the attributes, and (4) performing a remote unit testing of the function of the kernel by executing the script such that the user-space component (A) generates a message that identifies the attributes and (B) sends the message to the kernel-space component to facilitate (I) obtaining references to the objects by way of the attributes and (II) invoking the function by way of the references. Various other methods, systems, and computer-readable media are also disclosed.
Unified interface for development and testing of deployment resource architecture
Embodiments of the disclosure pertain to providing, via a graphical user interface, a set of user-selectable resources available for defining a resource architecture for an application in development; receiving, via the graphical user interface, a selection of a first resource for the application; receiving, via the graphical user interface, a selection of a second resource for the application; functionally connecting the first resource with the second resource, wherein the first resource, the second resource, and the logical relationship between the first resource and the second resource define the resource architecture for the application; verifying the resource architecture, wherein verifying the resource architecture comprises communicating over a network with the first resource and with the second resource and verifying that the first resource can communicate over a network with the second resource; and providing a graphical indication of a result of the verification of the resource architecture.
Test and automated remediation of depolyed computing products
A method of evaluating a computer-implemented product that is deployed on one or more endpoints. The method includes identifying a first program and a second program of a product deployed on a first endpoint of multiple endpoints. The method includes implementing a diagnostic process at the first endpoint. The diagnostic process includes a first subroutine directed to the first program and a second subroutine directed to a second program. The subroutines each execute installation and functional parameter tests of the programs. Responsive to the first subroutine indicating that the first program is operational, the method includes outputting data that the first subroutine passed. Responsive to the second subroutine returning an unexpected result, the method includes outputting data indicating details of the unexpected result and implementing a remediation that modifies the second program or a condition at the first endpoint to mitigate the unexpected result.
Presentation module for webinterface production and deployment system
The technology disclosed relates to a webinterface production and deployment system. In particular, it relates to a presentation module that applies a selected candidate individual to a presentation database to determine frontend element values corresponding to dimension values identified by the selected candidate individual, and which presents toward a user a funnel having the determined frontend element values.
SIMULATED DEPLOYMENT OF DATABASE ARTIFACTS
Systems and techniques for performing a simulated deployment to a database include generating a dependency graph of multiple artifacts contained in a container for deployment to a database and receiving an indication to perform a simulated deployment of the artifacts. In response to receiving the indication, the container is marked for simulated deployment. The artifacts are deployed to the database using the marked container and the generated dependency graph without committing the artifacts to the database. The deployed artifacts are rolled back such that no state changes are made to the database due to the simulated deployment.
GENERATING TEST CASES
Systems and techniques are described for generating test cases. In one example, a manual test of code is performed. The manual test uses proxy code to detect at least one input to the code and a data transfer between the code and other code. The data transfer includes an input to the other code and an output from the other code. A file is generated that includes the at least one input to the code, the input to the other code, the output from the other code, and data that specifies that the manual test of the code resulted in an expected output. A mock test automates the manual test of the code based at least in part on the file.
Testing software enhancements in database applications
A test system registers a series of tests that reference different software enhancements added to an application within a database system. The test system enables individual software enhancements referenced by the tests and then runs log lines through the application to produce test structured query language (SQL). The database system compares the test SQL with baseline SQL produced by the application without enabling the software enhancements. The database system executes the test SQL and captures performance metrics when the test SQL is different from the baseline SQL. Comparing SQL results avoids processing and capturing performance metrics for log lines not affected by the software enhancements. Incrementally running the log lines with one software enhancement enabled at a time also allows the test system to isolate the performance impact of individual software enhancements on the database system.
Dynamic trace level control
A method for adjusting a filtering mechanism within a trace logging system. The method may include receiving a plurality of messages from a software program, whereby each of the plurality of messages includes a message logging level. The method may also include storing the plurality of received messages in a buffer. The method may further include determining an error has occurred within the software program. The method may also include identifying each of the plurality of stored messages that aid in debugging the determined error. The method may further include updating an alert status configuration based on the message logging level associated with each of the plurality of identified messages.
Dynamic trace level control
A method for adjusting a filtering mechanism within a trace logging system. The method may include receiving a plurality of messages from a software program, whereby each of the plurality of messages includes a message logging level. The method may also include storing the plurality of received messages in a buffer. The method may further include determining an error has occurred within the software program. The method may also include identifying each of the plurality of stored messages that aid in debugging the determined error. The method may further include updating an alert status configuration based on the message logging level associated with each of the plurality of identified messages.