摘要:
A software method and computer system using said method are disclosed for efficiently reallocating unused system resources in a system in which unused system resources are initially allocated among groups that may become active or inactive. The method reallocates unused resources among active groups by scaling up these groups' initial entitlement shares, subject to maximum caps that may apply. For each group, the method calculates a scaling ratio equal to the cap divided by the entitlement, and then sorts the groups according their scaling ratios. The method processes each group in increasing order of their scaling ratios and reaches a maximal share of system resources to be allotted to each group, where the maximal share for an active group is between, or equal to, the group's initial entitlement value and its cap. Inactive groups have maximal shares equal to zero, reflecting that they receive no system resources. As the method processes the groups, it attempts to simply scale each active group by its scaling ratio. If scaling all unprocessed groups by the current group's scaling ratio would exceed the unallocated resources, then the remaining groups are scaled by the remaining resources divided by the sum of the entitlements of the remaining groups. Because the groups are sorted in order of increasing scaling ratios, the remaining active groups may be scaled as described without exceeding their caps.
摘要:
A method is disclosed for allocating computer system resources, such as memory, among applications. Input parameters are received for applications. The parameters include a minimum entitlement value, a maximum entitlement value, and a weight. The minimum entitlement is the minimum amount of resources that the application should receive. The maximum entitlement is the maximum amount of resources that the application should receive. The weight specifies the priority of the application relative to other applications, for purposes of distributing system resources. Computer system resources are then allocated among applications based on the parameters.
摘要:
A method comprises defining a scheduler parameter for a maximum allocation of capacity of a shared resource to a consumer for a scheduling interval. Utilization of an allocated capacity of the shared resource by the consumer during a given scheduling interval is measured, and when the allocated capacity of the shared resource is completely utilized by the consumer during the given scheduling interval, the scheduler increases the allocated capacity of the shared resource to the defined maximum allocation for the consumer for a next scheduling interval. Thus, rather than gradually increasing the allocation of capacity over many intervals, the scheduler immediately increases the allocation to a predefined maximum amount in response to an allocated amount of capacity being completely utilized during a scheduling interval.
摘要:
Apparatus, methods, and other embodiments associated with providing a correlation between a power distribution unit(s) and a device(s) are described. One example method includes storing first time series data that identifies, on a per power distribution unit (PDU) basis, current drawn from a set of PDUs. The example method may include storing second time series data that identifies, on a per device basis, power used by a set of related devices. With the two time series data available, the method may then provide a PDU-device correlation signal that identifies a correlation between current drawn from a PDU and power used by a device.
摘要:
In one embodiment, a system comprises a plurality of partitions that provide isolated operating environments for execution of software processes, wherein the plurality of partitions are arranged in a tiered manner with different partition types at different tiers, a data structure for representing the plurality of partitions, wherein the data structure comprises multiple nodes corresponding to the plurality of partitions and links between the nodes representing how the plurality of partitions are arranged in the tiers, and an arbiter software module for allocating resources between the plurality of partitions, wherein the arbiter software module receives requests to allocate resources to the plurality of partitions and traverses the data structure to determine which requests to satisfy.
摘要:
A system comprises a workload manager evaluator operable to receive a representative workload that is representative of competing workloads that share access to at least one shared computing resource. The workload manager evaluator is operable to evaluate performance of a scheduler that schedules access of the competing workloads to the shared computing resource according to defined control parameter values, wherein the workload manager evaluator evaluates performance of the scheduler under the representative workload for a plurality of different values of the control parameters. In certain embodiments, the workload manager evaluator determines an optimal value for the control parameters of the scheduler for scheduling access to the at least one shared computing resource for the representative workload to satisfy defined performance desires of the system.
摘要:
The inventive work load manager (WLM) dynamically responds to changes in workload characteristics. The WLM bases response on performance goals set by the administrator, and manipulates the resources to achieve these goals. The WLM receives performance information from performance monitors. The WLM examines the information from the performance monitors and compares the information with the goals using a Proportional Integral and Derivative controller. Based on the comparison, the WLM may increase, decrease, or leave unchanged, the resources devoted to an application. If the performance of an application is performance is lagging, e.g., if transactions are taking longer than the goal, then the WLM increases the entitlement. If an application is over-achieving, then the WLM will decrease its entitlement and allocate it to another application. The WLM uses an arbiter which allocates the resources between the different applications, based on the priorities of the applications. Consequently, the WLM creates a feedback loop between application performance and CPU entitlements.