Abstract:
The techniques described herein provide software testing that may concurrently process a user request using a live version of software and a shadow request, which is based on the user request, using a shadow version of software (e.g., trial or test version, etc.). The live version of software, unlike the shadow version, is user-facing and transmits data back to the users while the shadow request does not output to the users. An allocation module may vary allocation of the shadow requests to enable a ramp up of allocations (or possibly ramp down) of the shadow version of software. The allocation module may use allocation rules to dynamically initiate the shadow request based on various factors such as load balancing, user attributes, and/or other rules or logic. Thus, not all user requests may be issued as shadow requests.
Abstract:
A mechanism is described for facilitating dynamic data management for computing devices according to one embodiment. A method of embodiments, as described herein, includes tracking one or more factors relating to a plurality of data sets, evaluating the plurality of data sets based on the one or more factors. The evaluating may include speculating at least one of relevancy and accessibility of each of the plurality of data sets. The method may further include generating data scores, the data scores being associated with the plurality of data sets based on the evaluation of the plurality of data sets, performing a first comparison of the data scores of the plurality of data sets with a criteria score, and classifying each data set based on the first comparison. The classifying may include setting caching order for each data set of the plurality of data sets.
Abstract:
Embodiments of the present invention provide systems, methods, and computer program products for configuring auto-scaling parameters of a computing environment, as well as alerting a user when auto-scaling operations are not attainable given current operating configurations.
Abstract:
A system for maintaining a two-site configuration for continuous availability over long distances may include a first computing site configured to execute a first instance associated with a priority workload, the first instance being designated as an active instance; a second computing site configured to execute a second instance of the priority workload, the second instance being designated as a standby instance; a software replication module configured to replicate a unit of work data associated with the priority workload from a first data object associated with the active instance to a second data object associated with the standby instance, and a hardware replication module configured to replicate an image from a first storage volume to a copy on a second storage volume, wherein the first storage volume is associated with the first computing site, and the second storage volume is associated with a third computing site.
Abstract:
Systems and methods are provided for analyzing operating metrics of monitored metric sources. Aspects of the present disclosure may present for display information associated with the monitored metric source and the analysis of its operating metrics. Analysis comprises determination of reference values and tolerance levels which represent allowable deviations from the reference values. Input data includes a measurement of an operating parameter and a time stamp. Input data may be saved to a data store for using in future analysis of other input data. When input data is determined to be outside the tolerance level, notifications may be issued to alert administrators or systems of the anomaly.
Abstract:
Method and system for asynchronously dispersing Disaster Recovery (DR) enabling data between a plurality of storage sites. The method comprises: receiving, at a primary storage site, a written block and a write frequency counter associated with the written block. In case the write frequency counter is below a threshold: receiving information dispersal parameters including number indicative of a size difference between said written block and DR enabling data based on said written block; number of slices to slice said DR enabling data into and data indicative of DR storage sites of said plurality of storage sites for storing said slices. Further calculating DR enabling data based on written block, wherein DR enabling data is larger than said written block by size difference; slicing DR enabling data in accordance with number of slices; and dispersing slices in accordance with data indicative of DR storage sites.
Abstract:
Embodiments of the invention provide for controlling message flow across a parallel computer system having multiple compute nodes by selectively grouping compute nodes of such a system into node pools and assigning message flow control policies to nodes in the node pools. The message flow control policies specify logging and/or tracing activities to be performed by instances of applications running on nodes assigned to the node pools. As the application is executed, logging and/or tracing messages are generated on the compute nodes according to message flow control policies assigned to the nodes. Optionally, the message flow is analyzed, the message flow control policies are adjusted, and duplicate messages are eliminated.
Abstract:
A network-based production service is configured to process client requests for the production service via a network, capture production request data defining the requests and store the production request data in a data store. A test system comprising one or more controllers creates test jobs according to a test plan for testing the production service. The test plan creates a test profile for using specified production request data to simulate a load on the production service. Each job created by the test plan specifies a portion of production request data. A job queue receives and queues test jobs from one or more controllers configured to add test jobs to the job queue according to the test plan. Workers access jobs from the job queue and the production request data from the data store as specified in each job and replay the production request data to the production service.
Abstract:
A non-transitory computer program product with computer readable program code for system detection and flow control. The throughput and response time of the system are acquired, where the throughput reflects the number of requests from outside the system that has been processed by the system. The response time reflects time required by the system to process requests from outside the system. Acquiring throughput and response time includes directly sampling the throughput and the response time of the system; and conducting noise filtering on the sampled throughput and response time. A system performance characteristic parameter based on the filtered throughput and response time is obtained and is used to describe congestion condition of the system to determine whether there is a bottleneck in the system. Detecting a bottleneck and flow control is provided in a simple and effective way to ensure that requests are processed in time.
Abstract:
Described herein are systems and methods for providing software administration tools, for use in administering server configurations, such as in a traffic director or other type of server environment. In accordance with an embodiment, the system comprises a traffic director having one or more traffic director instances, which is configured to receive and communicate requests, from clients, to origin servers having one or more pools of servers. An administration server can be used to manage the traffic director, including a REpresentational State Transfer (REST) infrastructure and management service which maps REST calls to mbeans or other management components registered on the administration server, for use in managing the traffic director.