Adaptive CPU NUMA Scheduling
    12.
    发明申请
    Adaptive CPU NUMA Scheduling 审中-公开
    自适应CPU NUMA调度

    公开(公告)号:US20160085571A1

    公开(公告)日:2016-03-24

    申请号:US14492051

    申请日:2014-09-21

    Applicant: VMware, Inc.

    Abstract: Examples perform selection of non-uniform memory access (NUMA) nodes for mapping of virtual central processing unit (vCPU) operations to physical processors. A CPU scheduler evaluates the latency between various candidate processors and the memory associated with the vCPU, and the size of the working set of the associated memory, and the vCPU scheduler selects an optimal processor for execution of a vCPU based on the expected memory access latency and the characteristics of the vCPU and the processors. Some examples contemplate monitoring system characteristics and rescheduling the vCPUs when other placements may provide improved performance and/or efficiency.

    Abstract translation: 示例执行非均匀存储器访问(NUMA)节点的选择,用于将虚拟中央处理单元(vCPU)操作映射到物理处理器。 CPU调度器评估各种候选处理器与与vCPU相关联的存储器之间的延迟以及相关联存储器的工作集合的大小,并且vCPU调度器基于预期的存储器访问延迟选择用于执行vCPU的最优处理器 以及vCPU和处理器的特性。 一些示例考虑监视系统特性并重新安排vCPU,当其他布局可能提供改进的性能和/或效率时。

    Fine-grained application-aware latency optimization for virtual machines at runtime

    公开(公告)号:US11429424B2

    公开(公告)日:2022-08-30

    申请号:US16936289

    申请日:2020-07-22

    Applicant: VMware, Inc.

    Abstract: A method of selectively assigning virtual CPUs (vCPUs) of a virtual machine (VM) to physical CPUs (pCPUs), where execution of the VM is supported by a hypervisor running on a hardware platform including the pCPUs, includes determining that a first vCPU of the vCPUs is scheduled to execute a latency-sensitive workload of the VM and a second vCPU of the vCPUs is scheduled to execute a non-latency-sensitive workload of the VM and assigning the first vCPU to a first pCPU of the pCPUs and the second vCPU to a second pCPU of the pCPUs. A kernel component of the hypervisor pins the assignment of the first vCPU to the first pCPU and does not pin the assignment of the second vCPU to the second pCPU. The method further comprises selectively tagging or not tagging by a user or an automated tool, a plurality of workloads of the VM as latency-sensitive.

    PAIRWISE COMPARISON FOR LOAD BALANCING

    公开(公告)号:US20210211492A1

    公开(公告)日:2021-07-08

    申请号:US17208284

    申请日:2021-03-22

    Applicant: VMware, Inc.

    Abstract: Load balancing across hosts in a computer system is triggered based on pairwise comparisons of resource utilization at different host. A method for load balancing across hosts includes the steps of determining a resource utilization difference between first and second hosts, wherein the first host has a higher resource utilization than the second host, comparing the resource utilization difference against a threshold difference, and upon determining that the resource utilization difference exceeds the threshold difference, selecting a workload executing in the first host for migration to the second host.

    Implicit co-scheduling of CPUs
    17.
    发明授权

    公开(公告)号:US10572282B2

    公开(公告)日:2020-02-25

    申请号:US15494170

    申请日:2017-04-21

    Applicant: VMware, Inc.

    Abstract: Techniques for implicit coscheduling of CPUs to improve corun performance of scheduled contexts are described. One technique minimizes skew by implementing corun migrations, and another technique minimizes skew by implementing a corun bonus mechanism. Skew between schedulable contexts may be calculated based on guest progress, where guest progress represents time spent executing guest operating system and guest application code. A non-linear skew catch-up algorithm is described that adjusts the progress of a context when the progress falls far behind its sibling contexts.

    Automatic system service resource management for virtualizing low-latency workloads that are input/output intensive

    公开(公告)号:US10452572B2

    公开(公告)日:2019-10-22

    申请号:US15381067

    申请日:2016-12-15

    Applicant: VMware, Inc.

    Abstract: An approach for providing low-latency and fast response times required by virtual machine (VM) telecommunication (telco) workloads is disclosed. In one embodiment, a kernel scheduler identifies hypervisor system contexts that are outside of traditional virtual central processing unit (CPU) contexts and servicing telco workloads. The scheduler grants each such context exclusive affinity to a respective physical CPU, up to a configurable maximum number of such grants. The scheduler also increases the CPU reservations in a system resource pool for the hypervisor by one CPU, indicating that there is one fewer physical CPU on which VMs may be placed or moved. The exclusive affinities and CPU reservations may be dynamically adjusted as, e.g., new latency-sensitive VMs are powered on with virtual CPUs that need to be granted exclusive affinity to physical CPUs or contexts no longer relevant to telco workloads are demoted to not have exclusive affinity to physical CPUs.

    Adaptive CPU NUMA scheduling
    19.
    发明授权

    公开(公告)号:US10255091B2

    公开(公告)日:2019-04-09

    申请号:US14492051

    申请日:2014-09-21

    Applicant: VMware, Inc.

    Abstract: Systems and methods for performing selection of non-uniform memory access (NUMA) nodes for mapping of virtual central processing unit (vCPU) operations to physical processors are provided. A CPU scheduler evaluates the latency between various candidate processors and the memory associated with the vCPU, and the size of the working set of the associated memory, and the vCPU scheduler selects an optimal processor for execution of a vCPU based on the expected memory access latency and the characteristics of the vCPU and the processors. The systems and methods further provide for monitoring system characteristics and rescheduling the vCPUs when other placements provide improved performance and efficiency.

Patent Agency Ranking