摘要:
Embodiments of the invention relate to creating and maintaining consistent data sets in a shared pool of configurable computer resources to support disaster recovery support. Data from an application is stored in local data storage and replicated to another data storage. A consistency point of the data is created in both of the data storage, with the consistency point representing an identical data set at a point-in-time. Based upon the created consistency points, a consistent set of application data may be requested to support a read operation for a migrated application.
摘要:
Managing a data transfer from one or more source storage devices to one or more target storage devices. The data transfer comprises concurrent transfer of a multiplicity of data units pursuant to respective data transfer commands. The concurrent transfer of the multiplicity of data units is currently in-progress. A computer determines a currently-overloaded storage component involved in the data transfer. The computer determines a plurality of the data transfer commands that involve the overloaded storage component. The computer determines an approximately-minimum number of the data transfer commands to cancel to stop overload of the overloaded storage component. In response, the computer cancels the minimum number of the data transfer commands.
摘要:
Embodiments of the present invention provide an integrated host and subsystem port selection methodology that uses performance measurements combined with information about active data paths. This technique also helps in resilient fabric planning by selecting ports from redundant fabrics. In a typical embodiment, host port to storage port pairs that create a path between a host and a storage device will be identified. From these pairs, a set of host port to storage port candidates for communicate data from the host to the storage device will be identified based on a set of resiliency constraints. Then, a specific host port to storage port pair will be selected from the set based on a lowest joint workload measurement. A path will then be created between the specific host port and storage port, and data will be communicated from the host to the storage device via the path.
摘要:
Embodiments of the invention relate to creating and maintaining consistent data sets in a shared pool of configurable computer resources to support disaster recovery support. Data from an application is stored in local data storage and replicated to another data storage. A consistency point of the data is created in both of the data storage, with the consistency point representing an identical data set at a point-in-time. Based upon the created consistency points, a consistent set of application data may be requested to support a read operation for a migrated application.
摘要:
Managing a data transfer from one or more source storage devices to one or more target storage devices. The data transfer comprises concurrent transfer of a multiplicity of data units pursuant to respective data transfer commands. The concurrent transfer of the multiplicity of data units is currently in-progress. A computer determines a currently-overloaded storage component involved in the data transfer. The computer determines a plurality of the data transfer commands that involve the overloaded storage component. The computer determines an approximately-minimum number of the data transfer commands to cancel to stop overload of the overloaded storage component. In response, the computer cancels the minimum number of the data transfer commands.
摘要:
An approach for automatic storage planning and provisioning within a clustered computing environment is provided. Planning input for a set of storage area network volume controllers (SVCs) will be received within the clustered computing environment, the planning input indicating a potential load on the SVCs and its associated components. Analytical models (e.g., from vendors) can be also used that allow for a load to be accurately estimated on the storage components. Configuration data for a set of storage components (i.e., the set of SVCs, a set of managed disk (Mdisk) groups associated with the set of SVCs, and a set of backend storage systems) will also be collected. Based on this configuration data, the set of storage components will be filtered to identify candidate storage components capable of addressing the potential load. Then, performance data for the candidate storage components will be analyzed to identify an SVC and an Mdisk group to address the potential load.
摘要:
Scheduling a proposed additional data transfer from one or more source storage devices to one or more target storage devices. A computer receives a request for the proposed additional data transfer, and in response, determines a measure of the proposed additional data transfer. The computer determines a measure of recent actual data transfers. The recent actual data transfers involve one or more of the source storage devices and one or more of the target storage devices. In response to the request for the proposed additional data transfer, the computer estimates performance of one or more of the source storage devices and one or more of the target storage devices that would occur during the proposed additional data transfer based on the measure of recent actual data transfers combined with the measure of the proposed additional data transfer. The computer compares the estimated performance to a corresponding performance threshold, and if less, the computer postpones execution of the proposed additional data transfer.
摘要:
Embodiments of the present invention provide an approach for adapting an information extraction middleware for a clustered computing environment (e.g., a cloud environment) by creating and managing a set of statistical models generated from performance statistics of operating devices within the clustered computing environment. This approach takes into account the required accuracy in modeling, including computation cost of modeling, to pick the best modeling solution at a given point in time. When higher accuracy is desired (e.g., nearing workload saturation), the approach adapts to use an appropriate modeling algorithm. Adapting statistical models to the data characteristics ensures optimal accuracy with minimal computation time and resources for modeling. This approach provides intelligent selective refinement of models using accuracy-based and operating probability-based triggers to optimize the clustered computing environment, i.e., maximize accuracy and minimize computation time.
摘要:
The present invention proactively identifies hotspots in a cloud computing environment through cloud resource usage models that use workload parameters as inputs. In some embodiments the cloud resource usage models are based upon performance data from cloud resources and time series based workload trend models. Hotspots may occur and can be detected at any layer of the cloud computing environment, including the server, storage, and network level. In a typical embodiment, parameters for a workload are identified in the cloud computing environment and inputted into a cloud resource usage model. The model is run with the inputted workload parameters to identify potential hotspots, and resources are then provisioned for the workload so as to avoid these hotspots.
摘要:
In general, embodiments of present invention provide an approach for calibrating a cloud computing environment. Specifically, embodiments of the present invention provide an empirical approach for obtaining end-to-end performance characteristics for workloads in the cloud computing environment (hereinafter the “environment”). In a typical embodiment, different combinations of cloud server(s) and cloud storage unit(s) are determined. Then, a virtual machine is deployed to one or more of the servers within the cloud computing environment. The virtual machine is used to generate a desired workload on a set of servers within the environment. Thereafter, performance measurements for each of the different combinations under the desired workload will be taken. Among other things, the performance measurements indicate a connection quality between the set of servers and the set of storage units, and are used in calibrating the cloud computing environment to determine future workload placement. Along these lines, the performance measurements can be populated into a table or the like, and a dynamic map of a data center having the set of storage units can be generated.