摘要:
An improved technique involves generating a predictive model for data storage system configuration management. A customer support center generates such a predictive model from detailed customer configuration and transaction history. For example, a population of customers submits transaction logs to the customer support center; such transaction logs provide details as to how the customers responded to various events. The population of customers may also submit data including various statistics such as load intensity, workload characteristics, data access patterns, data change patterns, and data fingerprints to the customer support center. The customer support center then performs an analysis on the data and, from the analysis, computes values of model parameters that define a predictive model. This predictive model is configured to take in a particular state of any data storage system and produce a configuration that optimizes performance of that data storage system.
摘要:
An improved technique involves consulting a predictive model to optimize a configuration of a data storage system. A customer at a data storage system requests a configuration optimization from a customer support center by sending values of input parameters to the customer support center. Such input parameters represent a current state of the data storage system, e.g., initial configuration. The customer support center then receives configuration parameters in response to inputting the input parameters into a predictive model running on a host computer. The customer uses the configuration parameters to define the configuration state in which the data storage system runs.
摘要:
Example embodiments provide various techniques for modeling a storage environment at various times. A computer model can be composed to simulate a storage environment at various time periods. In an example, the computer model may simulate the storage environment in the past. As the storage environment is running, it continuously stores its existing state. This existing state can be accessed and a simulated workload may be applied to this previous state to identify an impact of the simulated workload on the existing state. A computer model may also simulate the storage environment in the future. Here, the future state of the storage environment may be projected from the existing state. Various extrapolation techniques may be used to project the future state of the storage environment.
摘要:
Example embodiments provide various techniques for modeling network storage environments. To model a particular storage environment, component models that are associated with the components of the storage environment are loaded. Each component model is programmed to mathematically simulate one or more components of the storage environment. A system model is then composed from the component models and this system model is configured to simulate the storage environment.
摘要:
Provided are a method, system, and article of manufacture that maintain a plurality of nodes coupled over a network, wherein a first node of the plurality of nodes initially owns a plurality of storage objects accessible via the network. A serialization lock is maintained in association with the plurality of storage objects, wherein the first node becomes unavailable. A second node of the plurality of nodes determines that the second node is to preemptively assume ownership of a selected storage object of the plurality of storage objects initially owned by the first node that has become unavailable. The second node acquires the serialization lock. The second node secures ownership of the selected storage object, subsequent to acquiring the serialization lock.
摘要:
A method, apparatus and computer implemented instructions for minimizing inconsistency between a set of data sources in a data processing system. A first signal is sent indicating that new content is present for the set of data sources. The new content is transmitted to the set of data sources in which the new content is unavailable for distribution by the set of data sources until a second signal is received by the set of data sources. The second signal is sent to the set of data sources if an acknowledgment is received from all of the sets of data sources.
摘要:
Provided are techniques for workload balancing. A message is received on a channel. A thread in a thread pool is selected to process the message. In response to determining that the message has been processed and a response has been sent on the channel by the thread, it is determined whether a total number of threads in the thread pool is greater than a low water mark plus one and whether the channel has more than a maximum number of threads blocked on a receive, wherein the low water mark represents a minimum number of threads in the thread pool. In response to determining that a number of threads in the thread pool is greater than the low water mark plus one and that the channel has more than the maximum number of threads blocked on a receive, the thread is terminated. In response to determining at least one of the number of threads in the thread pool is less than or equal to the low water mark plus one and the channel has less than or equal to the maximum number of threads blocked on a receive, the thread is retained.
摘要:
Example embodiments provide various techniques for modeling a storage environment at various times. A computer model can be composed to simulate a storage environment at various time periods. In an example, the computer model may simulate the storage environment in the past. As the storage environment is running, it continuously stores its existing state. This existing state can be accessed and a simulated workload may be applied to this previous state to identify an impact of the simulated workload on the existing state. A computer model may also simulate the storage environment in the future. Here, the future state of the storage environment may be projected from the existing state. Various extrapolation techniques may be used to project the future state of the storage environment.
摘要:
Provided are techniques for workload balancing. A message is received on a channel. A thread in a thread pool is selected to process the message. In response to determining that the message has been processed and a response has been sent on the channel by the thread, it is determined whether a total number of threads in the thread pool is greater than a low water mark plus one and whether the channel has more than a maximum number of threads blocked on a receive, wherein the low water mark represents a minimum number of threads in the thread pool. In response to determining that a number of threads in the thread pool is greater than the low water mark plus one and that the channel has more than the maximum number of threads blocked on a receive, the thread is terminated. In response to determining at least one of the number of threads in the thread pool is less than or equal to the low water mark plus one and the channel has less than or equal to the maximum number of threads blocked on a receive, the thread is retained.
摘要:
An intermediate content transcoder. An intermediate content transcoder can include a communicative receiver configured to receive master content while in transit from a content source to at least one content sink in a content distribution network, and a communicative transmitter configured to transmit transcoded content to one or more of the content sinks. A set of content transformations also can be provided. Furthermore, the intermediate content transcoder can include a set of conditions for selecting individual ones of the content transformations to be applied to master content received through the communicative receiver. Finally, the intermediate content transcoder can include a transcoder configured to apply the selected individual ones of the content transformations to the master content. The application of the content transformations can produce transcoded content. The communicative transmitter, in turn, can forward the produced transcoded content to one or more of the content sinks.