摘要:
Embodiments of the present invention address deficiencies of the art in respect to deploying heterogeneous workloads in separate resource pools and provide a method, system and computer program product for on-demand application scheduling in a heterogeneous environment. In one embodiment of the invention, a method for balancing nodal allocations in a resource pool common to both transactional workloads and long running workloads can include parsing a service policy for both transactional workloads and also long running workloads. An allocation of nodes for a common resource pool for the transactional and long running workloads can be determined to balance performance requirements for the transactional workloads and long running workloads specified by the service policy. Subsequently, the determined allocation can be applied to the common resource pool.
摘要:
The present invention provides a method, system, and computer program product for supporting a large number of intermittently used application clusters without consuming all the resources needed to run server processes for all the application clusters all the time. A method in accordance with the present invention comprises: holding an application request for a dormant application cluster in a holding zone; changing a status of the dormant application cluster to active; changing a status of an unused active application cluster to dormant; stopping server processes for the newly dormant application cluster; starting server processes for the newly active application cluster; and directing the application request held in the holding zone to the server processes for the newly active application cluster.
摘要:
Incoming work units (e.g., requests) in a computing workload are analyzed and classified according to predicted execution. Preferred embodiments track which instrumented wait points are encountered by the executing work units, and this information is analyzed to dynamically and autonomically create one or more recognizers to programmatically recognize similar, subsequently-received work units. When a work unit is recognized, its execution behavior is then predicted. Execution resources are then allocated to the work units in view of these predictions. The recognizers may be autonomically evaluated or tuned, thereby adjusting to changing workload characteristics. The disclosed techniques may be used advantageously in application servers, message-processing software, and so forth.
摘要:
In alternate embodiments, the invention is a message-passing process for routing communications between a transmitting parallel process and a receiving parallel process executing in an application server environment, or a machine or computer-readable memory having the message-passing process programmed therein, the message-passing process comprising: linking a context key to an addressable computing resource in the application server environment; linking the receiving parallel process to the context key; receiving a communication from the transmitting parallel process, wherein the communication transmits the context key; and routing the communication to the addressable computing resource linked to the context key.
摘要:
The invention described is a new and useful process for minimizing the overall rebate a provider disburses to customers when a service level agreement (SLA) breach occurs in a utility computing environment. Specifically, the process compares performance data and resource usage with the SLAs of the customers, and reallocates shared resources to those customers who represent a lesser penalty to the provider in the event of an SLA breach. The process determines which resources, used by customers representing the lesser penalty, are operating below peak capacity. The process then reallocates these under-utilized resources to those customers requiring additional resources to meet SLA thresholds. If all resources are operating at peak capacity, the process reallocates the resources to those customers whose SLAs represent a greater penalty in the event of an SLA breach as compared to those customers whose SLAs provide for a lesser penalty, thereby minimizing the total rebate due upon an SLA breach.