Patent classifications
G06F11/3688
Smart test case generator
Embodiments provide systems, methods, and computer-readable storage media for automated and objective testing of applications or processes. Graphical representations of the application may be analyzed to derive attribute data and identify flows (e.g., possible processing paths that may be accessed during utilization of the application by a user). Test cases may be automatically generated based on the attribute data and the identified flows. Additionally, testing scripts for testing the portions of the application corresponding to each identified flow may be generated using machine learning logic. Once generated, the testing scripts may be executed against the application to test different portions of the application functionality (or processes). Execution of the testing scripts may be monitored to generate feedback used to train the machine learning logic. Reports may be generated based on the monitoring and provided to users to enable the users to resolve any errors encountered during the testing.
Coverage of web application analysis
A method for detecting a defect may include extracting, from application code and using a framework support specification corresponding to a framework, a framework interaction between the application code and the framework. The framework interaction specifies an object used by the application code and managed by the framework. The method may further include performing, using the framework interaction, a dynamic analysis of the application code to obtain a heap snapshot, performing, using the heap snapshot and the framework interaction, a static analysis of the application code, and detecting, by the static analysis, the defect.
Big data application lifecycle management
Aspects of the present disclosure involve systems, methods, devices, and the like for creating an application lifecycle management platform for big data applications. In one embodiment the lifecycle management platform can include a multiple-layer container file that integrates multiple big-data tools/platforms. The system may create a generic template application, create a build environment for the generic template application, create a test environment for the generic template application, and run the built generic template application in the test environment prior to the user writing any new code in the generic template application. In one embodiment, the test environment includes a container management system or virtual machine that launches the big data application (which may be the generic template application before a developer edits the file) on a separate big-data server cluster.
Method and system for remote testing of devices
A method and system for remote testing of a plurality of devices is disclosed. The method may include receiving a request from a client system to perform testing on a set of remote devices. The local system and the client system are connected via a first network connection and the plurality of remote devices are connected to the local system via a second network connection. The method may further include receiving an input from the client system with reference to a test-suite to perform a testing action on the set of remote devices, generating a test command corresponding to the input, and transmitting the test command to each of the set of remote devices. The method may further include receiving feedback from each of the set of remote devices and transmitting the feedback to the client system.
Method and system for synchronous development and testing of live, multi-tenant microservices based SaaS systems
The present disclosure provides techniques for configuring and provisioning a tenant for testing microservices in a multi-tenant instance. Code is committed for a modified microservice, and a configuration is received for a production tenant of the multi-tenant instance. The configuration is updated to include a reference to the updated microservice, and then provided to a provisioner that provisions a test tenant based on the configuration. The microservices for the test tenant are compared with versions in a code version management system and updated, then a reference to the test tenant is provided to a developer to test the modified microservice. The test tenant may be deprovisioned after a predetermined amount of time, by a command of the developer, or other automated method.
NETWORK SERVICE MANAGEMENT SYSTEM AND NETWORK SERVICE MANAGEMENT METHOD
A CI/CD assist device accepts configuration data and test data collectively from a terminal (a vendor terminal) of a provider that provides a network service to a customer. The configuration data specifies a functional unit required to provide the network service, and the test data specifies test content for the network service or the functional unit. Each of an NOS, an NOS, and an NOS automatically builds the functional unit specified by the configuration data that is accepted by the CI/CD assist device. Each of a test device, a test device, and a test device automatically conducts a test on the network service or the functional unit based on the test data. The network service and the functional unit are built in each environment.
DEBUGGING MANAGEMENT PLATFORM AND OPERATING METHOD FOR THE SAME
A debugging management platform and an operating method for the same are provided. In the operating method, the debugging management platform operates a debugging agent service for establishing a debugging channel between a software development platform and a test platform. When receiving debugging packets are issued by the software development platform or the test platform, the debugging agent service analyzes the debugging packets and checks if the debugging packets meet an information security standard. The debugging packets are forwarded to the test platform or the software development platform if the debugging packets meet the information security standard. If the debugging packets do not meet the information security standard, the debugging packets are not forwarded, so as to ensure information security of the debugging packets that are forwarded between different environments.
SIMULATION METHOD FOR AUTONOMOUS VEHICLE AND METHOD FOR CONTROLLING AUTONOMOUS VEHICLE
The present document relates to a simulation method for an autonomous vehicle, a method for controlling the autonomous vehicle, a device, an electronic apparatus, a computer-readable storage medium, and a computer program product. The method for the simulation of the autonomous vehicle comprises acquiring current state information of the autonomous vehicle; performing the simulation based on the current state information to acquire the prediction information of the autonomous vehicle; and sending the prediction information to the autonomous vehicle.
SYSTEMS AND METHODS FOR AUTOMATED TEST DATA MICROSERVICES
Systems and methods for automated test data microservices are provided. Test versions of software (such as an Application Programming Interface (API)) may be configured to automatically generate test data and to call a microservice to manage the test data. The microservice may automatically add and remove the test data from an operational data store to facilitate the testing process and to automatically perform setup and teardown stages of the testing process.
SYSTEMS AND METHODS FOR TESTING COMPONENTS OR SCENARIOS WITH EXECUTION HISTORY
Systems and methods for testing components or scenarios with execution history are disclosed. A method may include: receiving, at a testing interface and from an application or program executed by a user electronic device, an identification of a test and one or more data layers of a plurality of data layers in pod to test, the plurality of data layers including a data collection layer, a data ingestion layer, a data messaging layer, a data enrichment layer, and a data connect layer; receiving, by the testing interface, a selection of testing parameters or values for the identified test; retrieving, by the testing interface, the identified test; executing, by the testing interface, the identified test on the identified one or more data layers using the selected testing parameters or values; retrieving, by testing interface, results of the execution of the test; and outputting, by the testing interface, the results.