摘要:
Implementations of the present disclosure provide systems and methods directed to receiving, at a computing device, trace data provided in a trace log file corresponding to a server, parameterizing, using the computing device, a first Markovian Arrival Process (MAP) model based on the trace data, parameterizing, using the computing device, a second MAP model based on the trace data, defining a queuing model that models a performance of the server and that includes the first MAP model and the second MAP model, and storing the queuing model in computer-readable memory.
摘要:
Implementations of the present disclosure provide computer-implemented methods including defining a workload comprising a plurality of service requests, each service request corresponding to a class of a plurality of classes, applying the workload to a computer system that receives and processes service requests, measuring a response time of the computer system for each request of the workload, estimating a mean service demand for each class based on the response times and a base queuing model that represents the computer system, and generating the queuing model based on the mean service demands and characteristics of the workload.
摘要:
Methods for performance evaluation of admission control policies (ACPs) include storing a simulation model including an admission queue and a finite capacity region (FCR), the admission queue admitting queries to the FCR based on an active ACP, the FCR modeling a resource constraint, generating traces, each trace being generated based on processing a single query using the database system, the single query being associated with a query type of a plurality of query types, for each query type, determining a query fork size and a service time from an associated trace to provide a plurality of tuples, each tuple being associated with a query type and including the fork size and the service time associated with the query type, parameterizing the simulation model based on the plurality of tuples, and for each of the ACPs, generating a performance parameter by applying a workload to the simulation model.
摘要:
Methods for performance evaluation of admission control policies (ACPs) include storing a simulation model including an admission queue and a finite capacity region (FCR), the admission queue admitting queries to the FCR based on an active ACP, the FCR modeling a resource constraint, generating traces, each trace being generated based on processing a single query using the database system, the single query being associated with a query type of a plurality of query types, for each query type, determining a query fork size and a service time from an associated trace to provide a plurality of tuples, each tuple being associated with a query type and including the fork size and the service time associated with the query type, parameterizing the simulation model based on the plurality of tuples, and for each of the ACPs, generating a performance parameter by applying a workload to the simulation model.
摘要:
Implementations of the present disclosure provide computer-implemented methods for predicting a performance of a consolidated virtualized computing environment. Methods include processing benchmark workloads using a plurality of virtual machines to generate a plurality of traces, each trace including data corresponding to requests issued by a respective virtual machine operating in an isolated environment, storing the plurality of traces in one or more trace repositories, each trace repository provided as a computer-readable storage medium, selecting a trace from the plurality of traces stored in the one or more trace repositories, parameterizing a queuing model based on the trace, the queuing model representing request queuing in the consolidated virtualized computing environment, and processing the queuing model using one or more processors to generate one or more response time estimates for the consolidated virtualized computing environment.
摘要:
Implementations of the present disclosure provide computer-implemented methods including defining a workload comprising a plurality of service requests, each service request corresponding to a class of a plurality of classes, applying the workload to a computer system that receives and processes service requests, measuring a response time of the computer system for each request of the workload, estimating a mean service demand for each class based on the response times and a base queuing model that represents the computer system, and generating the queuing model based on the mean service demands and characteristics of the workload.
摘要:
Implementations of the present disclosure provide computer-implemented methods for predicting a performance of a consolidated virtualized computing environment. Methods include processing benchmark workloads using a plurality of virtual machines to generate a plurality of traces, each trace including data corresponding to requests issued by a respective virtual machine operating in an isolated environment, storing the plurality of traces in one or more trace repositories, each trace repository provided as a computer-readable storage medium, selecting a trace from the plurality of traces stored in the one or more trace repositories, parameterizing a queuing model based on the trace, the queuing model representing request queuing in the consolidated virtualized computing environment, and processing the queuing model using one or more processors to generate one or more response time estimates for the consolidated virtualized computing environment.
摘要:
Implementations of the present disclosure provide systems and methods directed to receiving, at a computing device, trace data provided in a trace log file corresponding to a server, parameterizing, using the computing device, a first Markovian Arrival Process (MAP) model based on the trace data, parameterizing, using the computing device, a second MAP model based on the trace data, defining a queuing model that models a performance of the server and that includes the first MAP model and the second MAP model, and storing the queuing model in computer-readable memory.