-
公开(公告)号:US12132790B2
公开(公告)日:2024-10-29
申请号:US17875672
申请日:2022-07-28
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Patrick Bohan , Kshitij Arun Doshi , Brinda Ganesh , Andrew J. Herdrich , Monica Kenguva , Karthik Kumar , Patrick G Kutch , Felipe Pastor Beneyto , Rashmin Patel , Suraj Prabhakaran , Ned M. Smith , Petar Torre , Alexander Vul
IPC: H04L67/148 , G06F9/48 , H04L41/5003 , H04L41/5019 , H04L43/0811 , H04L47/70 , H04L67/00 , H04L67/10 , H04W4/40 , H04W4/70
CPC classification number: H04L67/148 , G06F9/4856 , H04L41/5019 , H04L43/0811 , H04L47/82 , H04L67/10 , H04L67/34 , H04W4/40 , H04W4/70 , H04L41/5003
Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QoS pre-allocation; and automatic QoS migration across edge computing nodes. In a specific example, a technique for service migration includes: identifying a service operated with computing resources in an edge computing system, involving computing capabilities for a connected edge device with an identified service level; identifying a mobility condition for the service, based on a change in network connectivity with the connected edge device; and performing a migration of the service to another edge computing system based on the identified mobility condition, to enable the service to be continued at the second edge computing apparatus to provide computing capabilities for the connected edge device with the identified service level.
-
公开(公告)号:US11838138B2
公开(公告)日:2023-12-05
申请号:US17534132
申请日:2021-11-23
Applicant: Intel Corporation
Inventor: Dario Sabella , Ned M. Smith , Neal Oliver , Kshitij Arun Doshi , Suraj Prabhakaran , Miltiadis Filippou , Francesc Guim Bernat
IPC: H04L12/14 , H04L67/1087 , H04L67/1074 , H04M15/00 , H04L67/10 , H04L67/12 , H04W4/24
CPC classification number: H04L12/1407 , H04L67/10 , H04L67/1074 , H04L67/1087 , H04L67/12 , H04M15/66 , H04M15/8353 , H04M15/84 , H04M15/85 , H04M15/852 , H04M15/853 , H04M15/886 , H04W4/24
Abstract: An architecture to allow Multi-Access Edge Computing (MEC) billing and charge tracking, is disclosed. In an example, a tracking process, such as is performed by an edge computing apparatus, includes: receiving a computational processing request for a service operated with computing resources of the edge computing apparatus from a connected edge device within the first access network, wherein the computational processing request includes an identification of the connected edge device; identifying a processing device, within the first access network, for performing the computational processing request; and storing the identification of the connected edge device, a processing device identification, and data describing the computational processes completed by the processing device in association with the computational processing request.
-
公开(公告)号:US11824732B2
公开(公告)日:2023-11-21
申请号:US16235462
申请日:2018-12-28
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Suraj Prabhakaran , Kshitij A. Doshi , Brinda Ganesh , Timothy Verrall
IPC: H04L41/16 , G06N3/04 , G06N5/04 , H04L41/5019 , H04L41/5009 , H04L41/5051 , H04L41/0816
CPC classification number: H04L41/16 , G06N3/04 , G06N5/04 , H04L41/0816 , H04L41/5012 , H04L41/5019 , H04L41/5051
Abstract: Examples include techniques for artificial intelligence (AI) capabilities at a network switch. These examples include receiving a request to register a neural network for loading to an inference resource located at the network switch and loading the neural network based on information included in the request to support an AI service to be provided by users requesting the AI service.
-
公开(公告)号:US11625277B2
公开(公告)日:2023-04-11
申请号:US16878861
申请日:2020-05-20
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Kshitij Arun Doshi , Bassam N. Coury , Suraj Prabhakaran , Timothy Verrall
Abstract: Systems and methods may be used to determine where to run a service based on workload-based conditions or system-level conditions. An example method may include determining whether power available to a resource of a compute device satisfies a target power, for example to satisfy a target performance for a workload. When the power available is insufficient, an additional resource may be provided, for example on a remote device from the compute device. The additional resource may be used as a replacement for the resource of the compute device or to augment the resource of the compute device.
-
公开(公告)号:US11580428B2
公开(公告)日:2023-02-14
申请号:US17668844
申请日:2022-02-10
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Suraj Prabhakaran , Kshitij Arun Doshi , Da-Ming Chiang , Joe Cahill
IPC: G06N5/04
Abstract: Various systems and methods of initiating and performing contextualized AI inferencing, are described herein. In an example, operations performed with a gateway computing device to invoke an inferencing model include receiving and processing a request for an inferencing operation, selecting an implementation of the inferencing model on a remote service based on a model specification and contextual data from the edge device, and executing the selected implementation of the inferencing model, such that results from the inferencing model are provided back to the edge device. Also in an example, operations performed with an edge computing device to request an inferencing model include collecting contextual data, generating an inferencing request, transmitting the inference request to a gateway device, and receiving and processing the results of execution. Further techniques for implementing a registration of the inference model, and invoking particular variants of an inference model, are also described.
-
公开(公告)号:US20220269960A1
公开(公告)日:2022-08-25
申请号:US17668844
申请日:2022-02-10
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Suraj Prabhakaran , Kshitij Arun Doshi , Da-Ming Chiang , Joe Cahill
IPC: G06N5/04
Abstract: Various systems and methods of initiating and performing contextualized AI inferencing, are described herein. In an example, operations performed with a gateway computing device to invoke an inferencing model include receiving and processing a request for an inferencing operation, selecting an implementation of the inferencing model on a remote service based on a model specification and contextual data from the edge device, and executing the selected implementation of the inferencing model, such that results from the inferencing model are provided back to the edge device. Also in an example, operations performed with an edge computing device to request an inferencing model include collecting contextual data, generating an inferencing request, transmitting the inference request to a gateway device, and receiving and processing the results of execution. Further techniques for implementing a registration of the inference model, and invoking particular variants of an inference model, are also described.
-
公开(公告)号:US11379264B2
公开(公告)日:2022-07-05
申请号:US16384554
申请日:2019-04-15
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Felipe Pastor Beneyto , Kshitij A. Doshi , Timothy Verrall , Suraj Prabhakaran
IPC: G06F9/48 , G06F1/3296 , G06F1/20 , G06F9/50 , G06F1/3206 , G06F1/30 , G06F1/3234
Abstract: Some examples provide for uninterruptible power supply form (UPS) resources and non-UPS resources to be offered in a composite node for customers to use. For a workload run on the composite node, monitoring of non-UPS resource power availability, resource temperature, and/or cooling facilities can take place. In the event, a non-UPS resource experiences a power outage or reduction in available power, temperature that is at or above a threshold level, and/or cooling facility outage, monitoring of performance of a workload executing on the non-UPS resource can take place. If the performance is acceptable and the power available to the non-UPS resource exceeds a threshold level, the supplied power can be reduced. If the performance experiences excessive levels of errors or slows unacceptably, the workload can be migrated to another non-UPS or UPS compliant resource.
-
公开(公告)号:US20220166847A1
公开(公告)日:2022-05-26
申请号:US17542175
申请日:2021-12-03
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Karthik Kumar , Thomas Willhalm , Petar Torre , Ned Smith , Brinda Ganesh , Evan Custodio , Suraj Prabhakaran
IPC: H04L67/60 , H04L12/66 , H04L47/70 , H04L67/2885 , H04L67/5681
Abstract: Technologies for fulfilling service requests in an edge architecture include an edge gateway device to receive a request from an edge device or an intermediate tier device of an edge network to perform a function of a service by an entity hosting the service. The edge gateway device is to identify one or more input data to fulfill the request by the service and request the one or more input data from an edge resource identified to provide the input data. The edge gateway device is to provide the input data to the entity associated with the request.
-
公开(公告)号:US20220166846A1
公开(公告)日:2022-05-26
申请号:US17390658
申请日:2021-07-30
Applicant: Intel Corporation
Inventor: Ramanathan Sethuraman , Timothy Verrall , Ned M. Smith , Thomas Willhalm , Brinda Ganesh , Francesc Guim Bernat , Karthik Kumar , Evan Custodio , Suraj Prabhakaran , Ignacio Astilleros Diez , Nilesh K. Jain , Ravi Iyer , Andrew J. Herdrich , Alexander Vul , Patrick G. Kutch , Kevin Bohan , Trevor Cooper
IPC: H04L67/303 , H04L9/40 , H04L9/08 , H04L67/12
Abstract: Technologies for managing telemetry and sensor data on an edge networking platform are disclosed. According to one embodiment disclosed herein, a device monitors telemetry data associated with multiple services provided in the edge networking platform. The device identifies, for each of the services and as a function of the associated telemetry data, one or more service telemetry patterns. The device generates a profile including the identified service telemetry patterns.
-
10.
公开(公告)号:US11243817B2
公开(公告)日:2022-02-08
申请号:US16369036
申请日:2019-03-29
Applicant: Intel Corporation
Inventor: Evan Custodio , Francesc Guim Bernat , Suraj Prabhakaran , Trevor Cooper , Ned M. Smith , Kshitij Doshi , Petar Torre
Abstract: Technologies for migrating data between edge accelerators hosted on different edge locations include a device hosted on a present edge location. The device includes one or more processors to: receive a workload from a requesting device, determine one or more accelerator devices hosted on the present edge location to perform the workload, and transmit the workload to the one or more accelerator devices to process the workload. The one or more processor is further to determine whether to perform data migration from the one or more accelerator devices to one or more different edge accelerator devices hosted on a different edge location, and send, in response to a determination to perform the data migration, a request to the one or more accelerator devices on the present edge location for transformed workload data to be processed by the one or more different edge accelerator devices.
-
-
-
-
-
-
-
-
-