摘要:
Embodiments of the present invention provide an approach for prioritizing a set of data packets associated with an application running in a networked computing environment (e.g., a cloud computing environment). In a typical embodiment, an SLA associated with a particular application and at least one other SLA associated with at least one other application are accessed (e.g., in a database or the like). A proposed modification to a priority of a set of data packets associated with the particular application (e.g., corresponding to a workload being processed via the particular application) is then determined based on a comparison of a current performance (e.g., transmission rate) of the set of data packets versus a specified performance of the set of data packets as set forth in the SLA. Then, an effect of the proposed modification on a capability of the particular application to meet a set of terms of the SLA, and of the at least one other application to meet a set of terms of the at least one other SLA will be evaluated. Based on this evaluation, it will be determined whether to implement the proposed modification. For example, the proposed modification can be implemented if it will not cause any of the SLAs to be unmet.
摘要:
Embodiments of the present invention provide a workload optimization approach that measures workload performance across combinations of hardware (platform, network configuration, storage configuration, etc.) and operating systems, and which provides a workload placement on the platforms where jobs perform most efficiently. This type of placement may be based on performance measurements (e.g., throughput, response, and other such service levels), but it can also be based on other factors such as power consumption or reliability. In a typical embodiment, ideal platforms are identified for handling workloads based on performance measurements and any applicable service level agreement (SLA) terms.
摘要:
Embodiments of the present invention provide an approach to prioritize jobs (e.g., within a cloud computing environment) so as to maximize positive financial impacts (or to minimize negative financial impacts) for cloud service providers, while not exceeding processing capacity or failing to meet terms of applicable Service Level Agreements (SLAs). Specifically, under the present invention a respective income (i.e., a cost to the customer), a processing need, and set of SLA terms (e.g., predetermined priorities, time constraints, etc.) will be determined for each of a plurality of jobs to be performed. The jobs will then be prioritized in a way that: maximizes cumulative/collective income; stays within the total processing capacity of the cloud computing environment; and meets the SLA terms.
摘要:
Embodiments of the present invention provide an approach to prioritize jobs (e.g., within a cloud computing environment) so as to maximize positive financial impacts (or to minimize negative financial impacts) for cloud service providers, while not exceeding processing capacity or failing to meet terms of applicable Service Level Agreements (SLAs). Specifically, under the present invention a respective income (i.e., a cost to the customer), a processing need, and set of SLA terms (e.g., predetermined priorities, time constraints, etc.) will be determined for each of a plurality of jobs to be performed. The jobs will then be prioritized in a way that: maximizes cumulative/collective income; stays within the total processing capacity of the cloud computing environment; and meets the SLA terms.
摘要:
Embodiments of the present invention provide for optimizing workload placement in a networked computing environment (e.g., a cloud computing environment). Specifically, a workload placement technique is applied to determine an optimal platform for handling an identified workload. The workload placement technique comprises one or more of the following: a shadow placement technique; a simultaneous placement technique; and/or a single platform placement technique. Once an optimal platform is identified, a workload timing method may be applied to determine when the workload should be placed thereon. The workload timing method can comprise one or more of the following: a time-based method whereby the workload is placed on the optimal platform at a predetermined time or time interval; and/or an event-based method whereby the workload is placed on the optimal platform based on an occurrence of one or more events external to the workload itself.
摘要:
Embodiments of the present invention provide a workload optimization approach that measures workload performance across combinations of hardware (platform, network configuration, storage configuration, etc.) and operating systems, and which provides a workload placement on the platforms where jobs perform most efficiently. This type of placement may be based on performance measurements (e.g., throughput, response, and other such service levels), but it can also be based on other factors such as power consumption or reliability. In a typical embodiment, ideal platforms are identified for handling workloads based on performance measurements and any applicable service level agreement (SLA) terms.
摘要:
Embodiments of the present invention provide an approach for dynamically modifying Quality of Service (QoS) levels for resources (e.g., applications, processes, services, etc.) running in a networked computing environment. Specifically, embodiments of the present invention dynamically adjust transport level networking QoS parameters based on associated service level agreements (SLA) term. In a typical embodiment, a set of service level requirements associated with a resource running in the networked computing environment will first be identified (e.g., in a computer data structure). Then, the set of service level requirements will be mapped to a set of QoS parameters associated with a transport layer of the networked computing environment. A current performance of the resource within the transport layer will then be determined. Once the current performance has been determined, it will be further determined whether the current performance meets the set of service level requirements. Based on this determination/comparison, the set of QoS parameters can be adjusted accordingly.
摘要:
Embodiments of the present invention provide an approach for prioritizing a set of data packets associated with an application running in a networked computing environment (e.g., a cloud computing environment). In a typical embodiment, an SLA associated with a particular application and at least one other SLA associated with at least one other application are accessed (e.g., in a database or the like). A proposed modification to a priority of a set of data packets associated with the particular application (e.g., corresponding to a workload being processed via the particular application) is then determined based on a comparison of a current performance (e.g., transmission rate) of the set of data packets versus a specified performance of the set of data packets as set forth in the SLA. Then, an effect of the proposed modification on a capability of the particular application to meet a set of terms of the SLA, and of the at least one other application to meet a set of terms of the at least one other SLA will be evaluated. Based on this evaluation, it will be determined whether to implement the proposed modification. For example, the proposed modification can be implemented if it will not cause any of the SLAs to be unmet.
摘要:
Embodiments of the present invention provide an approach for dynamically modifying Quality of Service (QoS) levels for resources (e.g., applications, processes, services, etc.) running in a networked computing environment. Specifically, embodiments of the present invention dynamically adjust transport level networking QoS parameters based on associated service level agreements (SLA) term. In a typical embodiment, a set of service level requirements associated with a resource running in the networked computing environment will first be identified (e.g., in a computer data structure). Then, the set of service level requirements will be mapped to a set of QoS parameters associated with a transport layer of the networked computing environment. A current performance of the resource within the transport layer will then be determined. Once the current performance has been determined, it will be further determined whether the current performance meets the set of service level requirements. Based on this determination/comparison, the set of QoS parameters can be adjusted accordingly.
摘要:
Embodiments of the present invention provide an approach for optimizing workload placement in a networked computing environment (e.g., a cloud computing environment). Specifically, under embodiments of the present invention, a workload placement technique is applied to determine an optimal platform for handling an identified workload. The workload placement technique can comprise one or more of the following: a shadow placement technique whereby the workload is placed on multiple similar platforms substantially contemporaneously; a simultaneous placement technique whereby the workload is placed on multiple different platforms substantially contemporaneously; and/or a single platform placement technique whereby the workload is placed on a single platform at a given time. Once an optimal platform is identified, a workload timing method may be applied to determine when the workload should be placed thereon. The workload timing method can comprise one or more of the following: a time-based method whereby the workload is placed on the optimal platform at a predetermined time or time interval; and/or an event-based method whereby the workload is placed on the optimal platform based on an occurrence of one or more events external to the workload itself (e.g., a certain CPU or memory consumption, etc.). Once the workload is placed on the optimal platform, optimization data can be gathered for future assessments.