摘要:
An apparatus, system and method to optimize load distribution across logical and physical resources in a storage system. An apparatus in accordance with the invention may include an availability module and an allocation module. The availability module may dynamically assign values to resources in a hierarchical tree structure. Each value may correspond to an availability parameter such as allocated volumes, current resource utilization, and historic resource utilization. The allocation module may serially process the values and allocate a load to a least busy resource in the hierarchical tree structure based on the assigned values.
摘要:
A mechanism for selectively disabling and enabling read caching based on past performance of the cache and current read/write requests. The system improves overall performance by using an autonomic algorithm to disable read caching for regions of backend disk storage (i.e., the backstore) that have had historically low cache hit ratios. The result is that more cache becomes available for workloads with larger hit ratios, and less time and machine cycles are spent searching the cache for data that is unlikely to be there.
摘要:
An apparatus, system and method to optimize load distribution across logical and physical resources in a storage system. An apparatus in accordance with the invention may include an availability module and an allocation module. The availability module may dynamically assign values to resources in a hierarchical tree structure. Each value may correspond to an availability parameter such as allocated volumes, current resource utilization, and historic resource utilization. The allocation module may serially process the values and allocate a load to a least busy resource in the hierarchical tree structure based on the assigned values.
摘要:
A mechanism for selectively disabling and enabling read caching based on past performance of the cache and current read/write requests. The system improves overall performance by using an autonomic algorithm to disable read caching for regions of backend disk storage (i.e., the backstore) that have had historically low cache hit ratios. The result is that more cache becomes available for workloads with larger hit ratios, and less time and machine cycles are spent searching the cache for data that is unlikely to be there.