SIMULTANEOUS-MULTI-THREADING (SMT) AWARE PROCESSOR ALLOCATION FOR CLOUD REAL-TIME WORKLOADS

    公开(公告)号:US20230063893A1

    公开(公告)日:2023-03-02

    申请号:US17464155

    申请日:2021-09-01

    Applicant: Red Hat, Inc.

    Abstract: An example system includes a processor and a node agent executing on the processor. The node agent is configured to receive a message indicative of a workload, a processor policy of the workload, and a number of processor threads requested for the workload. The node agent is also configured to allow simultaneous allocation of a processor core to the workload and another workload based on the processor policy being a first policy. The node agent is also configured to prevent simultaneous allocation of the processor core to the workload and the other workload based on the processor policy being a second policy or a third policy. The node agent is also configured to allow simultaneous allocation of the processor core for two or more of the requested processor threads based on the processor policy being the second policy. The node agent is also configured to prevent simultaneous allocation of the processor core for more than one of the requested processor threads based on the processor policy being the third policy.

    Scheduling workloads on partitioned resources of a host system in a container-orchestration system

    公开(公告)号:US11868805B2

    公开(公告)日:2024-01-09

    申请号:US17228929

    申请日:2021-04-13

    Applicant: RED HAT, INC.

    CPC classification number: G06F9/4881 G06F9/54

    Abstract: Techniques of scheduling workload(s) on partitioned resources of host systems are described. The techniques can be used, for example, in a container-orchestration system. One technique includes retrieving information characterizing at least one schedulable partition and determining an availability and a suitability of one or more of the schedulable partition(s) for executing a workload in view of the information. Each of the schedulable partition(s) includes resources of one or more host systems. The technique also includes selecting one or more of the schedulable partition(s) to execute the workload in view of the availability and the suitability.

    Aggregating host machines into a single cloud node for workloads requiring excessive resources

    公开(公告)号:US11755375B2

    公开(公告)日:2023-09-12

    申请号:US17537037

    申请日:2021-11-29

    Applicant: Red Hat, Inc.

    CPC classification number: G06F9/505 G06F9/5077 H04L67/60 G06F9/54

    Abstract: A system and method for aggregating host machines into a single cloud node for workloads requiring excessive resources. The method includes providing a plurality of computing devices in association with a cloud service system. The method includes defining an aggregated node of the cloud service system corresponding to at least two computing devices of the plurality of computing devices. The method includes exposing an application programming interface (API) that is indicative of combined resources of the at least two computing devices of the plurality of computing devices. The method includes receiving a query to perform a workload requiring a set of resources that exceed the resources provided by each of the computing devices of the cloud service system. The method includes forwarding, to the aggregated node, the query to cause the at least two computing devices to perform the workload using the combined resources of the least two computing device.

    AGGREGATING HOST MACHINES INTO A SINGLE CLOUD NODE FOR WORKLOADS REQUIRING EXCESSIVE RESOURCES

    公开(公告)号:US20230168943A1

    公开(公告)日:2023-06-01

    申请号:US17537037

    申请日:2021-11-29

    Applicant: Red Hat, Inc.

    CPC classification number: G06F9/505 G06F9/5077 H04L67/32 G06F9/54

    Abstract: A system and method for aggregating host machines into a single cloud node for workloads requiring excessive resources. The method includes providing a plurality of computing devices in association with a cloud service system. The method includes defining an aggregated node of the cloud service system corresponding to at least two computing devices of the plurality of computing devices. The method includes exposing an application programming interface (API) that is indicative of combined resources of the at least two computing devices of the plurality of computing devices. The method includes receiving a query to perform a workload requiring a set of resources that exceed the resources provided by each of the computing devices of the cloud service system. The method includes forwarding, to the aggregated node, the query to cause the at least two computing devices to perform the workload using the combined resources of the least two computing device.

    SCHEDULING IN A CONTAINER ORCHESTRATION SYSTEM UTILIZING HARDWARE TOPOLOGY HINTS

    公开(公告)号:US20220405135A1

    公开(公告)日:2022-12-22

    申请号:US17352685

    申请日:2021-06-21

    Applicant: RED HAT, INC.

    Abstract: A request to execute a workload that utilizes an amount of resources to be executed is received from a client device. Corresponding resources that are available at multiple non-uniform memory access (NUMA) nodes are received from one or more host systems. A particular NUMA node of the multiple NUMA nodes is identified in view of the particular NUMA node having available resources that are greater than the amount of resources to execute the workload. A scheduling hint is assigned to the workload that indicates that the particular NUMA node is to be used to execute the workload.

    Scheduling in a container orchestration system utilizing hardware topology hints

    公开(公告)号:US11954534B2

    公开(公告)日:2024-04-09

    申请号:US17352685

    申请日:2021-06-21

    Applicant: RED HAT, INC.

    CPC classification number: G06F9/505 G06F9/5061 G06F9/45533

    Abstract: A request to execute a workload that utilizes an amount of resources to be executed is received from a client device. Corresponding resources that are available at multiple non-uniform memory access (NUMA) nodes are received from one or more host systems. A particular NUMA node of the multiple NUMA nodes is identified in view of the particular NUMA node having available resources that are greater than the amount of resources to execute the workload. A scheduling hint is assigned to the workload that indicates that the particular NUMA node is to be used to execute the workload.

    SCHEDULING WORKLOADS ON PARTITIONED RESOURCES OF A HOST SYSTEM IN A CONTAINER-ORCHESTRATION SYSTEM

    公开(公告)号:US20220326986A1

    公开(公告)日:2022-10-13

    申请号:US17228929

    申请日:2021-04-13

    Applicant: RED HAT, INC.

    Abstract: Techniques of scheduling workload(s) on partitioned resources of host systems are described. The techniques can be used, for example, in a container-orchestration system. One technique includes retrieving information characterizing at least one schedulable partition and determining an availability and a suitability of one or more of the schedulable partition(s) for executing a workload in view of the information. Each of the schedulable partition(s) includes resources of one or more host systems. The technique also includes selecting one or more of the schedulable partition(s) to execute the workload in view of the availability and the suitability.

    Simultaneous-multi-threading (SMT) aware processor allocation for cloud real-time workloads

    公开(公告)号:US12141611B2

    公开(公告)日:2024-11-12

    申请号:US17464155

    申请日:2021-09-01

    Applicant: Red Hat, Inc.

    Abstract: An example system includes a processor and a node agent executing on the processor. The node agent is configured to receive a message indicative of a workload, a processor policy of the workload, and a number of processor threads requested for the workload. The node agent is also configured to allow simultaneous allocation of a processor core to the workload and another workload based on the processor policy being a first policy. The node agent is also configured to prevent simultaneous allocation of the processor core to the workload and the other workload based on the processor policy being a second policy or a third policy. The node agent is also configured to allow simultaneous allocation of the processor core for two or more of the requested processor threads based on the processor policy being the second policy. The node agent is also configured to prevent simultaneous allocation of the processor core for more than one of the requested processor threads based on the processor policy being the third policy.

    INCREASE ASSIGNMENT EFFECTIVENESS OF KUBERNETES PODS BY REDUCING REPETITIVE POD MIS-SCHEDULING

    公开(公告)号:US20220350656A1

    公开(公告)日:2022-11-03

    申请号:US17243450

    申请日:2021-04-28

    Applicant: Red Hat, Inc.

    Abstract: Systems and methods for intelligently scheduling a pod in a cluster of worker nodes are described. A scheduling service may account for previous scheduling attempts by considering the time and node (scheduling data) on which a preceding attempt to schedule a node were made, and factoring this information into the scheduling decision. Upon making a determination of a node on which to attempt to schedule the pod, the scheduling data may be updated with the time and node ID of the determined node and the pod may be scheduled on the determined node. In response to determining that the pod has been evicted from the determined node, the above process may continue iteratively until the pod has been successfully scheduled.

Patent Agency Ranking