摘要:
Automated or autonomic techniques for managing deployment of one or more resources in a computing environment based on varying workload levels. The automated techniques may comprise predicting a future workload level based on data associated with the computing environment. Then, an estimation is performed to determine whether a current resource deployment is insufficient, sufficient, or overly sufficient to satisfy the future workload level. Then, one or more actions are caused to be taken when the current resource deployment is estimated to be insufficient or overly sufficient to satisfy the future workload level. Actions may comprise resource provisioning, resource tuning and/or admission control.
摘要:
A method is provided for cache flushing in a computer system having a processor, a cache, a synchronization primitive detector, and a cache flush engine. The method includes providing a synchronization primitive from the processor into the computer system; detecting the synchronization primitive in the synchronization primitive detector; providing a trigger signal from the synchronization primitive detector in response to detection of the synchronization primitive; providing cache information from the recall unit into the computer system in response to the trigger signal; and flushing the cache in response to the cache information in the computer system.
摘要:
Accurate recovery of a primary server at a recovery server is accomplished with reduced network overhead by maintaining at the recovery server checkpoints taken over time of the primary server and an intention log of logical transactions pending and complete by the primary server. The recovery server is brought to the operational state of the primary server as of failure at the primary server by recovering the most recent checkpoint and then executing the transactions indicated by the intention log as received but not complete as of the time that the checkpoint was taken.
摘要:
A system for coordinating error tracking, level setting and reporting among the various layers/components of a distributed storage system. Each component of the distributed system includes a trigger generation and response (TGR) utility, which generates an error tracking trigger (ETT), comprising: (1) an action that the initiator wants the stack's error tracking mechanisms to take; (2) a message that the initiator wants the stack to immediately post in its logs; and (3) a route/direction that the trigger is to be transmitted through the stack. The ETT is transmitted one layer at a time through the stack, and each intervening layer of the stack is equipped with a utility to examine the ETT and take the appropriate action(s), designated by the trigger. An error log is maintained by each layer of the stack and used to record information about the error and enable user determination of the source, timing and cause of errors.
摘要:
A cache coherent distributed shared memory multi-processor computer system is provided which supports transactional memory semantics. A cache flushing engine and temporary buffer allow selective forced write-backs of dirty cache lines to the home memory. A flush can be performed from the updated cache to the temporary buffer and then to the home memory after confirmation of receipt or from the updated cache to the home memory directly with the temporary buffer holding the old data until confirmation that the home memory contains the update.
摘要:
A method, apparatus, and computer program product for storage pools with write atomicity. An abstraction manager enforces write atomicity and disallows options which are inconsistent with write atomicity. The abstraction manager constructs through a physical device interface a logical continuous view of a storage pool in a manner consistent with write atomicity. Applications collect information specific to write atomicity from the abstraction manager through an application interface.
摘要:
A method, apparatus, and computer instructions for controlling data flow. A control message is formed for the data flow in response to an event while the data flow is occurring. The control message includes a data type, an action, and a duration. The control message is sent to a receiver data processing system, wherein the receiver data processing system modifies the data flow to the data processing system using the control message.
摘要:
A cache coherent distributed shared memory multi-processor computer system is provided with a memory controller which includes a recall unit. The recall unit allows selective forced write-backs of dirty cache lines to the home memory. After a request is posted in the recall unit, a recall (“flush”) command is issued which forces the owner cache to write-back the dirty cache line to be flushed. The memory controller will inform the recall unit as each recall operation is completed. The recall unit operation will be interrupted when all flush requests are completed.
摘要:
In a computer system divided into multiple separate workload partitions, wherein at least one of the partitions contains multiple applications, and a loopback interface is made available to each partition, a method is provided for routing packets from one application to another application. In one embodiment of the invention, each packet to be routed from one application to another application in the same home partition comprises a first packet, and all other packets comprise second packets. The method comprises the steps of assigning the same pre-specified IP address to each of the first packets, and routing each of the first packets from its home partition to a loopback interface at the IP network processing layer. The method further includes marking the first packets to identify the home partition of each first packet, and directing the marked first packets to a mechanism configured to deliver each first packet only to its home partition, in response to the markings. Each second packet is assigned the IP address of its intended destination, and routed thereto over the loopback interface.
摘要:
A method and system for enabling dynamic matching of storage utilization characteristics of a host system application with the characteristics of the available storage pools of an attached distributed storage system, in order to provide an optimal match between the application and selected storage pool. An abstraction manager is provided, enhanced with a storage device configuration utility/module, which performs a series of tasks to (1) obtain/collect the correct configuration information from each connected storage device or storage pools and/or (2) calculate the configuration information when the information is not readily available. The storage device configuration module then normalizes, collates and matches the configuration information to the various applications running on the host system and/or outputs the information to a user/administrator of the host system via a software interface. By using the collected/calculated configuration information, applications are assigned to an optimal storage pool in an intelligent and efficient manner.