Patent classifications
G06F11/302
Centralized error telemetry using segment routing header tunneling
A network device receives a data packet including a source address and a destination address. The network device drops the data packet before it reaches the destination address and generates an error message indicating that the data packet has been dropped. The network device encapsulates the error message with a segment routing header comprising a list of segments. The first segment of the list of segments in the segment routing header identifies a remote server, and at least one additional segment is an instruction for handling the error message. The network device sends the encapsulated error message to the remote server based on the first segment of the segment routing header.
SYSTEMS AND METHODS FOR DETECTION OF DEGRADATION OF A VIRTUAL DESKTOP ENVIRONMENT
Described embodiments provide systems and methods for detection of the degradation of a virtual desktop environment. A computing device may receive data from a plurality of client devices. The computing device may identify a subset of client devices from the plurality of client devices with at least one characteristic in common based on the received data. The computing device may determine a ratio of the identified subset of client devices, the ratio being a comparison of client devices of the subset with a value above a first threshold to a total number of client devices of the subset, and the value being indicative of a characteristic of performance for that client device. The computing device may identify a cause of an anomaly in the performance of the application based on the ratio exceeding a second threshold.
PREDICTION OF BUFFER POOL SIZE FOR TRANSACTION PROCESSING WORKLOADS
Techniques are described herein for prediction of an buffer pool size (BPS). Before performing BPS prediction, gathered data are used to determine whether a target workload is in a steady state. Historical utilization data gathered while the workload is in a steady state are used to predict object-specific BPS components for database objects, accessed by the target workload, that are identified for BPS analysis based on shares of the total disk I/O requests, for the workload, that are attributed to the respective objects. Preference of analysis is given to objects that are associated with larger shares of disk I/O activity. An object-specific BPS component is determined based on a coverage function that returns a percentage of the database object size (on disk) that should be available in the buffer pool for that database object. The percentage is determined using either a heuristic-based or a machine learning-based approach.
ENVIRONMENTAL IMPACT POWER CONSUMPTION RATING FOR APPLICATIONS
Systems and methods for generating a power consumption rating include receiving instrumentation data corresponding to a plurality of applications. The received instrumentation data is processed to calculate a relative power consumption value for each application of the plurality of applications. The relative power consumption value for each application is compared and a power consumption rating for each application based on the comparison is generated, thereby providing a visual indicator of power consumption for the applications that can be easily evaluated.
System and method for creating buffered firewall logs for reporting
A system for firewall data log processing, comprising a firewall logging system operating on a first processor and configured to cause the first processor to receive firewall log data and to process the firewall log data on a periodic basis to reduce the size of the firewall log data and a firewall reporting system operating on a second processor and configured to process the reduced size firewall log data to generate a report on a user interface that includes one or more analytics from the reduced size firewall data.
COMPARING THE PERFORMANCE OF MULTIPLE APPLICATION VERSIONS
Comparing the performance of multiple versions or branches/paths of an application (e.g., a web service or application) may be conducted within a suitable computing environment. Such an environment may be virtual in nature, cloud-based, or server-based, and is hosted with tools for simultaneously (or nearly simultaneously) executing multiple containers or other code collections with the same or similar operating conditions (e.g., network congestion, resource contention, memory management schemes). By arranging the performance test of different application versions in different sequences executed in parallel in separate containers, fair comparisons of the tested applications will be obtained. Testing sequences may be executed multiple times, and metrics are collected during each execution. Afterward, the results for each metric for each code version are aggregated and displayed to indicate their relative performance quantitatively and/or qualitatively.
Allocation of resources for a plurality of hosts
It is presented a method for enabling allocation of resources for a plurality of hosts. The method is performed by a server (1) and comprises identifying (S100) a service running on one or more of the plurality of hosts, determining (S140) a stretch factor for a recurring load pattern for the service running on the one or more of the plurality of hosts, and storing (S150) the identified service together with the determined stretch factor. It is also presented a server, a computer program and a computer program product.
PREVENTION AND REMEDIATION OF MALWARE BASED ON SELECTIVE PRESENTATION OF FILES TO PROCESSES
Malware prevention and remediation is provided by monitoring actions performed processes and maintaining indications of which processes are trusted; selectively presenting canary files to these processes, which includes presenting the canary files to processes not indicated as being trusted and hiding the canary files from processes indicated as being trusted, and where the monitoring includes monitoring for access of canary files with change privileges; scoring each of the processes based on the actions performed, including any access of canary files with change privileges, which scoring produces a malice score for each process; and automatically terminating any process for which its malice score indicates at least a threshold level of malice in the execution of the process.
Determining a future operation failure in a cloud system
Examples described relate to determining a future operation failure in a cloud system. In an example, a historical utilization of resources for performing an operation in a cloud system may be determined. A current utilization of resources in the cloud system may be determined. Based on the historical utilization of resources for performing the operation in the cloud system and the current utilization of resources in the cloud system, a determination may be made whether a future performance of the operation in the cloud system is likely to be a failure. In response to a determination that the future performance of the operation in the cloud system is likely to be a failure, an alert may be generated.
Test equipment interface add-on having a production support equipment module and a selectively removable test support equipment module
Devices, systems, and methods for providing an engine control system configured with a two-part test equipment monitor where at least one part is selectively removable are disclosed. An engine control system for an aircraft includes an electronic control unit (ECU). The ECU is configured to implement a production support equipment module and a selectively removable test support equipment module. The production support equipment module enables restricted data monitoring of the engine control system. The test support equipment module enables a comprehensive interface with the engine control system when installed with the ECU.