-
公开(公告)号:US20230062253A1
公开(公告)日:2023-03-02
申请号:US17883011
申请日:2022-08-08
申请人: Intel Corporation
摘要: Various systems and methods for implementing a multi-access edge computing (MEC) based system to realize 5G Network Edge and Core Service Dimensioning using Machine Learning and other Artificial Intelligence Techniques, for improved operations and usage of computing and networking resources, and are disclosed herein. In an example, processing circuitry of a compute node on a network is used to analyze execution of an application to obtain operational data. The compute node then may modularize functions of the application based on the operational data to construct modularized functions. A phase transition graph is constructed using a machine-learning based analysis, the phase transition graph representing state transitions from one modularized function to another modularized function, where the phase transition graph is used to dimension the application by distributing the modularized functions across the network.
-
公开(公告)号:US11451455B2
公开(公告)日:2022-09-20
申请号:US16540800
申请日:2019-08-14
申请人: Intel Corporation
IPC分类号: H04L41/5009 , G06F13/28 , G06F15/173 , H04L12/06 , H04L41/5022 , H04L43/0852 , H04L49/9047 , H04L47/62 , H04L12/46 , H04L41/5003 , H04L47/6275
摘要: Technologies for latency based service level agreement (SLA) management in remote direct memory access (RDMA) networks include multiple compute devices in communication via a network switch. A compute device determines a service level objective (SLO) indicative of a guaranteed maximum latency for a percentage of RDMA requests of an RDMA session. The compute device receives latency data indicative of latency of an RDMA request from a host device. The compute device determines a priority associated with the RDMA request as a function of the SLO and the latency data. The compute device schedules the RDMA request based on the priority. The network switch may allocate queue resources to the RDMA request based on the priority, reclaim the queue resources after the RDMA request is scheduled, and then return the queue resources to a free pool. Other embodiments are described and claimed.
-
公开(公告)号:US20220286399A1
公开(公告)日:2022-09-08
申请号:US17637416
申请日:2020-09-11
申请人: Intel Corporation
发明人: Niall McDonnell , Gage Eads , Mrittika Ganguli , Chetan Hiremath , John Mangan , Stephen Palermo , Bruce Richardson , Edwin Verplanke , Praveen Mosur , Bradley Chaddick , Abhishek Khade , Abhirupa Layek , Sarita Maini , Rahul Shah , Shrikant Shah , William Burroughs , David Sonnier
IPC分类号: H04L47/125 , H04L47/625 , H04L47/62 , H04L47/6275
摘要: Methods, apparatus, systems, and articles of manufacture are disclosed for hardware queue scheduling for multi-core computing environments. An example apparatus includes a first core and a second core of a processor, and circuitry in a die of the processor, at least one of the first core or the second core included in the die, the at least one of the first core or the second core separate from the circuitry, the circuitry to enqueue an identifier to a queue implemented with the circuitry, the identifier associated with a data packet, assign the identifier in the queue to a first core of the processor, and in response to an execution of an operation on the data packet with the first core, provide the identifier to the second core to cause the second core to distribute the data packet.
-
公开(公告)号:US11362904B2
公开(公告)日:2022-06-14
申请号:US16957626
申请日:2019-02-21
申请人: INTEL CORPORATION
发明人: Mrittika Ganguli , Dinesh Kumar , Robert Valiquette , Yadong Li , Mohan Kumar
摘要: Technologies for enhanced network discovery and configuration include a network with a fabric manager and multiple network devices. A network device requests platform information from a management controller and receives the platform information via a sideband interface. The network device broadcasts a discovery message indicative of the platform information on a link layer network. The fabric manager discovers the network topology with an enhanced link layer discovery protocol and creates a vPOD in the network. The vPOD includes an application network with multiple racks. The fabric manager creates a tagged network domain for the vPOD. The fabric manager sends an out-of-band configuration command to the network device with a tag associated with the vPOD. After receiving the out-of-band configuration command, the network device receives a packet, compares domain metadata of the packet to the tag received from the fabric manager, and routes the packet. Other embodiments are described and claimed.
-
公开(公告)号:US10735835B2
公开(公告)日:2020-08-04
申请号:US15395482
申请日:2016-12-30
申请人: Intel Corporation
发明人: Susanne M. Balle , Rahul Khanna , Nishi Ahuja , Mrittika Ganguli
IPC分类号: G06F1/20 , G06F3/06 , H04L12/24 , H04L29/08 , G06F13/42 , H04Q11/00 , H05K7/14 , G06F9/50 , H04L12/911 , B25J15/00 , B65G1/04 , H05K7/20 , H04L12/931 , H04L12/939 , H04W4/02 , H04L12/751 , H05K1/18 , G05D23/19 , G05D23/20 , H04L12/927 , H05K1/02 , H04L12/781 , H04Q1/04 , G06F12/0893 , H03M7/30 , H05K13/04 , G06F13/16 , G11C5/06 , G06F11/14 , G06F11/34 , G06F12/0862 , G06F15/80 , H04L12/919 , G06F12/10 , G06F13/40 , G06Q10/06 , G07C5/00 , H04L12/28 , H04L29/12 , H04L9/06 , H04L9/14 , H04L9/32 , H04L12/933 , H04L12/947
摘要: Technologies for allocating resources of a set of managed nodes to workloads to manage heat generation include an orchestrator server to receive resource allocation objective data including a target temperature for one or more of the managed nodes. The orchestrator server is also to determine an initial assignment of a set of workloads among the managed nodes, receive telemetry data from the managed nodes indicative of resource utilization by each of the managed nodes and one or more temperatures and fan speeds of the managed nodes as the workloads are performed, predict future heat generation of the workloads as a function of the telemetry data, determine, as a function of the predicted future heat generation, an adjustment to the assignment of the workloads to achieve the target temperature, and apply the adjustments to the assignments of the workloads among the managed nodes as the workloads are performed.
-
公开(公告)号:US10721137B2
公开(公告)日:2020-07-21
申请号:US15198300
申请日:2016-06-30
申请人: Intel Corporation
摘要: Systems described herein operate to improve network performance in a multi-tenant cloud computing environment. Systems can include communication circuitry and processing circuitry to generate a phase sequence matrix that indicates the identity and number of phases of a workload by measuring resources of the compute node during execution of the workload throughout a lifetime of the workload. The processing circuitry can generate a workload fingerprint that includes the phase sequence matrix and a phrase residency matrix. The phase residency matrix can indicate the fraction of execution time of the workload spent in each phase identified in the phase sequence matrix. A cloud controller can access the workload fingerprint for multiple workloads operating on multiple compute nodes in the cloud cluster to adjust workload allocations based at least on these workload fingerprints and on whether service level objectives (SLOs) are being met.
-
公开(公告)号:US10659386B2
公开(公告)日:2020-05-19
申请号:US15872928
申请日:2018-01-16
申请人: Intel Corporation
IPC分类号: G06F15/173 , H04L12/911 , G06F9/50 , H04L29/08 , G06F9/455 , H04L12/26
摘要: Technologies for contention-aware cloud compute scheduling include a number of compute nodes in a cloud computing cluster and a cloud controller. Each compute node collects performance data indicative of cache contention on the compute node, for example, cache misses per thousand instructions. Each compute node determines a contention score as a function of the performance data and stores the contention score in a cloud state database. In response to a request for a new virtual machine, the cloud controller receives contention scores for the compute nodes and selects a compute node based on the contention score. The cloud controller schedules the new virtual machine on the selected compute node. The contention score may include a contention metric and a contention score level indicative of the contention metric. The contention score level may be determined by comparing the contention metric to a number of thresholds. Other embodiments are described and claimed.
-
公开(公告)号:US20190319892A1
公开(公告)日:2019-10-17
申请号:US16451454
申请日:2019-06-25
申请人: Intel Corporation
IPC分类号: H04L12/873 , H04L12/911 , H04L12/875
摘要: Technologies for managing burst bandwidth requirements are disclosed. In the illustrative embodiment, a software-defined network (SDN) controller monitors storage devices in a data center. If a storage device fails, the SDN controller manages the bandwidth used to replicate the data that was stored on the failed storage device. The SDN controller may allocate an initial amount of bandwidth based on one or more parameters of the storage device, and the SDN controller may increase the bandwidth in a series of discrete steps. In another embodiment, the SDN controller may predict a bandwidth burst based on sequential writes at a storage sled from several compute devices, and allocate bandwidth accordingly in a tiered manner.
-
公开(公告)号:US11575609B2
公开(公告)日:2023-02-07
申请号:US16517358
申请日:2019-07-19
申请人: Intel Corporation
发明人: Arvind Srinivasan , Ramakrishna Huggahalli , Parthasarathy Sarangam , Sunil Ahluwalia , Mrittika Ganguli , Malek Musleh
IPC分类号: H04L47/2441 , H04L47/122 , H04L47/11
摘要: A switch or network interface can detect congestion caused by a flow of packets. The switch or network interface can generate a congestion hint packet and send the congestion hint packet directly to a source transmitter of the flow of packets that caused the congestion. The congestion hint packet can include information that the source transmitter can use to determine a remedial action to attempt to alleviate or stop congestion at the switch or network interface. For example, the transmitter can reduce a transmit rate of the flow of packets and/or select another route for the flow of packets. Some or all switches or network interfaces between the source transmitter and a destination endpoint can employ flow differentiation whereby a queue is selected to accommodate for a flow's sensitivity to latency.
-
公开(公告)号:US11575607B2
公开(公告)日:2023-02-07
申请号:US17018809
申请日:2020-09-11
申请人: Intel Corporation
发明人: Stephen Palermo , Bradley Chaddick , Gage Eads , Mrittika Ganguli , Abhishek Khade , Abhirupa Layek , Sarita Maini , Niall McDonnell , Rahul Shah , Shrikant Shah , William Burroughs , David Sonnier
IPC分类号: H04L47/125 , H04L47/62 , H04L47/625 , H04L47/6275
摘要: Methods, apparatus, systems, and articles of manufacture are disclosed for dynamic load balancing for multi-core computing environments. An example apparatus includes a first and a plurality of second cores of a processor, and circuitry in a die of the processor separate from the first and the second cores, the circuitry to enqueue identifiers in one or more queues in the circuitry associated with respective ones of data packets of a packet flow, allocate one or more of the second cores to dequeue first ones of the identifiers in response to a throughput parameter of the first core not satisfying a throughput threshold to cause the one or more of the second cores to execute one or more operations on first ones of the data packets, and provide the first ones to one or more data consumers to distribute the first data packets.
-
-
-
-
-
-
-
-
-