Patent classifications
G06F11/3664
REDUCING TIME TO TEST CYCLE FIRST FAIL
A system that automatically reduces test cycle time to save resources and developer time. The present system selects a subset of tests from a full test plan that should be selected for a particular test cycle, rather than running the entire test plan. The subset of tests is intelligently selected using metrics such as tests associated with changed code and new and modified tests.
SYSTEM AND METHOD FOR DETECTING INSIDER THREATS IN SOURCE CODE
A code repository stores source code. An insider threat detection system stores instructions for detecting code defects and criteria indicating predetermined types of code defects that, when present, are associated with intentional obfuscation of one or more functions of the source code. The insider threat detection system receives an entry of source code and detects, using the model, a set of code defects in the entry of source code. A defect type is determined for each code defect, thereby determining a set of defect types included in the entry of source code. If it is determined that each of the predetermined types of code defects indicated by the criteria is included in the determined set of defect types, the entry of source code is determined to include an insider threat.
Signing and verifying mutable structured documents
A structured document is verified for changes that are made during and after deployment of an application. The structured document includes first fields that are designated as mutable, and second fields that are designated as immutable. An attempted change is detected to the structured document during or after deployment of the application. Upon detecting the attempted change, a digital signature is generated of the second fields of the structured document. A determination is made whether the generated digital signature of the second fields matches a reference digital signature of the second fields. Upon determining that the generated digital signature matches the reference digital signature, the change to the structured document is permitted. Upon determining that the generated digital signature does not match the reference digital signature, the change is blocked to the structured document.
Processor with debug pipeline
A processor includes an execution pipeline that includes a plurality of execution stages, execution pipeline control logic, and a debug system. The execution pipeline control logic is configured to control flow of an instruction through the execution stages. The debug system includes a debug pipeline and debug pipeline control logic. The debug pipeline includes a plurality of debug stages. Each debug pipeline stage corresponds to an execution pipeline stage, and the total number of debug stages corresponds to the total number of execution stages. The debug pipeline control logic is coupled to the execution pipeline control logic. The debug pipeline control logic is configured to control flow through the debug stages of debug information associated with the instruction, and to advance the debug information into a next of the debug stages in correspondence with the execution pipeline control logic advancing the instruction into a corresponding stage of the execution pipeline.
Techniques for large-scale functional testing in cloud-computing environments
Techniques are disclosed for generating an execution plan for performing functional tests in a cloud-computing environment. Infrastructure resources and capabilities (e.g., system requirements) may be defined within an infrastructure object (e.g., a resource of a declarative infrastructure provisioner) that stores a code segment that implements the resource or capability. Metadata may be maintained that indicates what particular capabilities are applicable to each infrastructure resource. Using the metadata, the system can generate an execution plan by combining code segments for each resource with code segments defining each capability in accordance with the metadata. The execution plan may include programmatic instructions that, when executed, generate a set of test results. The system can execute instructions that cause the set of test results to be presented at a user device.
Agentless distributed monitoring of microservices through a virtual switch
Disclosed are systems, computer-readable media and methods for monitoring performance data across microservices. One example method includes establishing a service policy configured on a centralized switch controller, applying the service profile to a virtual interface associated with a microservice, mapping a microservice name for the microservice to an IP address and a port number, tracking a protocol flow for the microservice, wherein the protocol flow is associated with a virtual switch, to yield data, aggregating the data to yield aggregated data and presenting the aggregated data on a user interface.
Crowd-sourced automatic generation of user interface tests for enterprise-specific mobile applications
A computer-implemented method includes downloading respective instances of an enterprise mobile application to a plurality of mobile devices. The instances of the enterprise mobile applications, while executing on respective mobile devices, capture, for each session, a session log that includes indications of ordered user actions occurring during the session, and optionally time intervals between user actions and/or user attributes. Captured session logs stored at and are mined by one or more servers to discover a particular pattern or sequence of user actions that occurred across multiple, different user sessions. If the number and/or rate of occurrences of the particular pattern is greater than a threshold, a new test case corresponding to the pattern is automatically generated and added to a suite of test cases for the UI functionality of the enterprise mobile application. The updated test suite may be automatically executed on a test version of the enterprise mobile application.
DEFECT REPORTING IN APPLICATION TESTING
The present subject matter relates to defect reporting in application testing. In an implementation, a category of application testing is determined based on a testing instance of an application. The category of application testing is indicative of an aspect of the application, being tested. A list of previously reported defects associated with the determined category of application testing is displayed in a display layer over the testing instance of the application. A first user-input indicative of one of acceptance and rejection of a previously reported defect, from the list, with respect to the testing instance of the application is received. The first user-input is aggregated with previous user-inputs indicative of one of acceptance and rejection of the previously reported defect. It is determined whether the previously reported defect is irrelevant to the testing instance of the application based on the aggregation.
Cloud Assisted Behavioral Automated Testing
A computer readable storage medium, system and method for improving automated testing systems to include a first and second behavioral data. The first behavioral data is collected periodically and the second behavioral data is collected in real time. The receipt of the first behavioral data and a second behavioral data are followed by the receipt of a system configuration template. A test case is updated based on the first and second behavioral data, and an automated test environment is reconfigured based on the first behavioral data, second behavioral data, and the system configuration template. The test executes in the automated test environment producing a test result.
DEBUGGING FAILURE OF A SERVICE VALIDATION TEST
The present disclosure generally discloses a testing capability related to service testing in a communication network. The testing capability may be configured to support debugging of failures identified during service validation testing of a service in a communication network. The testing capability may be configured to support debugging of failures (e.g., transmission failures or the like) associated with a failed service validation test (e.g., a transmission verification test or the like). The testing capability may be configured to support debugging of failures identified during service validation testing of an Ethernet service. The testing capability may be configured to support debugging of failures (e.g., frame loss or the like) identified during service validation testing of an Ethernet service where the service validation testing of the Ethernet service is performed based on International Telecommunication Union (ITU)-Standardization (ITU-T) Y.1564, which is an Ethernet-based service activation test methodology.