摘要:
Embodiments are provided for managing the system performance of resources performing tasks in response to task requests from tenants. In one aspect, a system that comprises at least one resource configured to perform at least one admitted task with an impact under the control of a computer system. The computer system provides services to more than one tenant. The computer system comprises a strategist configured to assess the impact of the admitted task to create a cost function vector containing multiple cost function specifications and a budget policy vector containing multiple budget policies and an actuator. The actuator receives the cost function vector and the budget policy vector from the strategist, receives a task request one of the more than one tenants, and calculates cost functions based upon the cost function vector to predict the impact of the task request on the resources for each of the task requests. The actuator throttles the task requests based upon the budget policies for the impact on the resources to create at least one of the admitted task performed by the resource and a delayed task request.
摘要:
Embodiments are provided for managing the system performance of resources performing tasks in response to task requests from tenants. In one aspect, a system that comprises at least one resource configured to perform at least one admitted task with an impact under the control of a computer system. The computer system provides services to more than one tenant. The computer system comprises a strategist configured to assess the impact of the admitted task to create a cost function vector containing multiple cost function specifications and a budget policy vector containing multiple budget policies and an actuator. The actuator receives the cost function vector and the budget policy vector from the strategist, receives a task request one of the more than one tenants, and calculates cost functions based upon the cost function vector to predict the impact of the task request on the resources for each of the task requests. The actuator throttles the task requests based upon the budget policies for the impact on the resources to create at least one of the admitted task performed by the resource and a delayed task request.
摘要:
Methods, systems, and physical computer-readable storage medium for processing a plurality of IO request jobs are provided. The method includes determining whether one or more request jobs are not meeting a QoS target, each job of the one or more request jobs having a corresponding priority, selecting a highest priority job from the one or more request jobs, if one or more request jobs are not meeting the QoS target, determining whether the highest priority job has a corresponding effective rate limit imposed thereon, if so, relaxing the corresponding effective rate limit, and if not, selecting one or more lower priority jobs from the one or more request jobs and tightening a corresponding effective limit on the one or more lower priority jobs from the one or more request jobs in accordance with a delay factor limit.
摘要:
A method is disclosed that places data-intensive subprocesses in close physical and logical proximity to the facility responsible for storing the data, so that high efficiencies at reduced cost are achieved. In one specific example, new computer programs, termed adjuncts, are added and placed in a logical partition on a storage facility so that they can be invoked using appropriate commands issued on the I/O channel. Further, programs or changes are added to existing programs on the host machine, wherein such programs or changes discover the function extensions and invoke them to perform data processing.
摘要:
A method is disclosed that places data-intensive subprocesses in close physical and logical proximity to the facility responsible for storing the data, so that high efficiencies at reduced cost are achieved. In one specific example, new computer programs, termed adjuncts, are added and placed in a logical partition on a storage facility so that they can be invoked using appropriate commands issued on the I/O channel. Further, programs or changes are added to existing programs on the host machine, wherein such programs or changes discover the function extensions and invoke them to perform data processing.
摘要:
A method for video playback uses only resources universally supported by a browser (“inline playback”) operating in virtually all handheld media devices. In one case, the method first prepares a video sequence for display by a browser by (a) dividing the video sequence into a silent video stream and an audio stream; (b) extracting from the silent video stream a number of still images, the number of still images corresponding to at least one of a desired output frame rate and a desired output resolution; and (c) combining the still images into a composite image. In one embodiment, the composite image having a number of rows, with each row being formed by the still images created from a fixed duration of the silent video stream. Another method plays the still images of the composite image as a video sequence by (a) loading the composite image to be displayed through a viewport defined the size of one of the still images; (b) selecting one of the still images of the composite image; (c) setting the viewport to display the selected still image; and (d) setting a timer for a specified time period based on a frame rate, such that, upon expiration of the specified time period: (i) selecting a next one of the still images to be displayed in the viewport, unless all still images of the composite image have been selected; and (ii) return to step (c) if not all still images have been selected.
摘要:
A process for caching data in a cache memory includes upon detecting that a first page is in a first or second list, the first page is moved to a most recently used (MRU) position in the second list. Upon detecting that the first page is in a first history list, a first target size is updated to a second target size for the first and second lists, the first page is moved from the first history list to the MRU position in the second list, and the first page is fetched to the cache memory. Upon detecting that the first page is in a second history list, the second target size is updated to a third target size for the first and second lists, and the first page is moved from the second history list to the MRU position in the second list.
摘要:
A process for caching data in a cache memory includes upon detecting that a first page is in a first or second list, the first page is moved to a most recently used (MRU) position in the second list. Upon detecting that the first page is in a first history list, a first target size is updated to a second target size for the first and second lists, the first page is moved from the first history list to the MRU position in the second list, and the first page is fetched to the cache memory. Upon detecting that the first page is in a second history list, the second target size is updated to a third target size for the first and second lists, and the first page is moved from the second history list to the MRU position in the second list.
摘要:
Embodiments of the invention relate to block layout and block allocation in a file system to support write transactions. Regions in a cluster file system are defined to support a block allocation, include both write affinity and wide striping region. An allocation map is maintained to define and support the block allocation regions. For each write transaction, a copy of data is written to both regions, and in the event there is a shortage of blocks in one of the regions, at least one block in the other region is dynamically converted to support the write transaction.
摘要:
Data storage reliability is maintained in a write-back distributed data storage system including multiple nodes, each node comprising a processor and an array of failure independent data storage devices. Information is stored as a set of stripes, each stripe including a collection of multiple data strips and associated parity strips, the stripes distributed across multiple corresponding primary data nodes and multiple corresponding parity nodes. A primary data node maintains the data strip holding a first copy of data, and each parity node maintains a parity strip holding a parity for the multiple data strips. A read-modify-write parity update protocol is performed for maintaining parity coherency, the primary data node driving parity coherency with its corresponding parity nodes, independently of other data nodes, in order to keep its relevant parity strips coherent.