摘要:
A method and apparatus for controlling the number of servers in an information handling system in which incoming work requests belonging to a first service class are placed in a queue for processing by one or more servers. The system also has units of work assigned to a second service class that acts as a donor of system resources. In accordance with the invention, a performance measure is defined for the first service class as well as for the second service class. Before adding servers to the first service class, there is determined not only the positive effect on the performance measure for the first service class, but also the negative effect on the performance measure for the second service class. Servers are added to the first service class only if the positive effect on the performance measure for the first service class outweighs the negative effect on the performance measure for the second service class.
摘要:
A method and apparatus for controlling the number of servers in a multisystem cluster. Incoming work requests are organized into service classes, each of which has a queue serviced by servers across the cluster. Each service class has defined for it a local performance index for each particular system of the cluster and a multisystem performance index for the cluster as a whole. Each system selects one service class as a donor class for donating system resources and another service class as a receiver class for receiving system resources, based upon how well the service classes are meeting their goals. Each system then determines the resource bottleneck causing the receiver class to miss its goals. If the resource bottleneck is the number of servers, each system determines whether and how many servers should be added to the receiver class, based upon whether the positive effect of adding such servers on the performance index for the receiver class outweighs the negative effect of adding such servers on the performance measure for the donor class. If a system determines that servers should be added to the receiver class, it then determines the system in the cluster to which the servers should be added, based upon the effect on other work on that system. To make this latter determination, each system first determines whether another system has enough idle capacity and, if so, lets that system add servers. If no system has sufficient idle capacity, each system then determines whether the local donor class will miss its goals if servers are started locally. It not, the servers are started on the local system. Otherwise, each system determines where the donor class will be hurt the least and acts accordingly. To ensure the availability of a server capable of processing each of the work requests in the queue, each system determines whether there is a work request in the queue with an affinity only to a subset of the cluster that does not have servers for the queue and, if so, starts a server for the queue on a system in the subset to which the work request has an affinity.
摘要:
A technique is disclosed for managing a workload distributed across multiple data processing systems to enhance shared resource access to meet a common performance standard. The technique includes on at least one system, measuring performance of the work units on the system to create local performance data, and on at least some of the systems sending the local performance data to at least one other system of the multiple data processing systems. The method further includes on at least one of the systems, receiving the performance data from the sending systems to create remote performance data, and adjusting at least one control parameter for accessing shared resources in response to the local and remote performance data to modify the performance of the work units distributed across the data processing systems to achieve the common performance standard. A dynamic resource clustering process is also employed to enhance the shared resource management.
摘要:
A workload manager creates goal control data, defining two or more classes of system work units, in response to specification of goals of two or more goal types for the classes, and specification of importance values for each of the goal types. A system resource manager causes the goals to be met by periodically sampling work unit status; calculating a performance index for each class; selecting a receiver class to receive improved service based on the relative performance indexes and goal importance; a system bottleneck impacting achievement of goal by the receiver class is identified; and one or more system control data elements are identified and adjusted to cause the goal to be met for the receiver class.
摘要:
A technique is disclosed for managing a workload distributed across multiple data processing systems to enhance shared resource access to meet a common performance standard. The technique includes on at least one system, measuring performance of the work units on the system to create local performance data, and on at least some of the systems sending the local performance data to at least one other system of the multiple data processing systems. The method further includes on at least one of the systems, receiving the performance data from the sending systems to create remote performance data, and adjusting at least one control parameter for accessing shared resources in response to the local and remote performance data to modify the performance of the work units distributed across the data processing systems to achieve the common performance standard. A dynamic resource clustering process is also employed to enhance the shared resource management.
摘要:
Central processing unit (CPU) resources are managed within a computing environment. When the allocation of CPU resources to a partition of the computing environment is to be adjusted, the allocation is adjusted dynamically. This dynamic adjustment is across at least two partitions of the computing environment. The adjusting includes modifying processor weights associated with the partitions.
摘要:
Disclosed is a method and computer program device for dynamically managing the assignment of alias addresses to base addresses referencing an input/output (I/O) device, such as a direct access storage device (DASD). Two distinct methods are disclosed. In one method, alias addresses are assigned based on the performance of the I/O devices. In this method, alias addresses are assigned to highly utilized devices, as indicated by device performance data, in order to maximize the efficient utilization of I/O device resources. In a second method, workload management principles are utilized to assign alias addresses. In this method, a correlation is made between each I/O device and the service classes utilizing each device. As in the first method, performance data is generated for each I/O device. Alias addresses are assigned to I/O devices experiencing queue delays as indicated by their performance data, if the device is associated with a service class that has failed to meet one or more processing goals. These methods may operate on a single host, or in a multi-host environment. The methods may be operated individually, or concurrently. Methods are disclosed to manage contention between concurrently operating assignment methods, and between multiple hosts concurrently operating one or more assignment methods.
摘要:
A method and apparatus for managing CPU resources of a logically partitioned computing environment without shared memory access. A logical partition needing additional resources sends a message requesting such resources to a central domain manager, which sends messages to other partitions in the same group requesting that they assess their ability to donate resources to the requesting partition. Upon receiving such assessment request, each logical partition assesses its ability to donate resources to the requesting partition and responds accordingly to the domain manager. If at least one partition responds that it can donate resources to the requesting partition, the domain manager sends a message to a selected donor partition requesting that it reconfigure itself to donate resources to the requesting partition. Upon receiving a notification from the donor partition that it has successfully reconfigured itself, the domain manager notifies the requesting partition, which reconfigures itself to accept the donated resources.
摘要:
A method and apparatus for managing CPU resources of a logically partitioned computing environment without shared memory access. A logical partition needing additional resources sends a message requesting such resources to a central domain manager, which sends messages to other partitions in the same group requesting that they assess their ability to donate resources to the requesting partition. Upon receiving such assessment request, each logical partition assesses its ability to donate resources to the requesting partition and responds accordingly to the domain manager. If at least one partition responds that it can donate resources to the requesting partition, the domain manager sends a message to a selected donor partition requesting that it reconfigure itself to donate resources to the requesting partition. Upon receiving a notification from the donor partition that it has successfully reconfigured itself, the domain manager notifies the requesting partition, which reconfigures itself to accept the donated resources.
摘要:
A method and apparatus for enforcing capacity limitations such as those imposed by software license agreements in an information handling system in which a physical machine is divided into a plurality of logical partitions, each of which is allocated a defined portion of processor resources by a logical partition manager. A software license manager specifies a maximum allowed consumption of processor resources by a program executing in one of the logical partitions. A workload manager also executing in the partition measures the actual consumption of processor resources by the logical partition over a specified averaging interval and compares it with the maximum allowed consumption. If the actual consumption exceeds the maximum allowed consumption, the workload manager calculates a capping pattern and interacts with the logical partition manager to cap the actual consumption of processor resources by the partition in accordance with the calculated capping pattern. To provide additional capping flexibility, partitions are assigned phantom weights that the logical partition manager adds to the total partition weight to determine whether the partition has exceeded its allowed share of processor resources for capping purposes. The logical partition thus becomes a “container” for the licensed program with an enforced processing capacity less than that of the entire machine.