MEMORY MANAGEMENT SYSTEM, COMPUTING SYSTEM, AND METHODS THEREOF

    公开(公告)号:US20190095122A1

    公开(公告)日:2019-03-28

    申请号:US15717963

    申请日:2017-09-28

    Abstract: According to various aspects, a computing system may include one or more first memories of a first memory type and one or more second memories of a second memory type different from the first memory type and a memory controller. The memory controller may be configured to receive telemetry data associated with at least one of the one or more first memories and the one or more second memories, execute a data transfer between the one or more first memories and the one or more second memories in a first operation mode of the memory controller, suspend a data transfer between the one or more first memories and the one or more second memories in a second operation mode of the memory controller, and switch between the first operation mode and the second operation mode based on the telemetry data.

    SERVICE LEVEL AGREEMENT-BASED MULTI-HARDWARE ACCELERATED INFERENCE

    公开(公告)号:US20190044831A1

    公开(公告)日:2019-02-07

    申请号:US15857526

    申请日:2017-12-28

    Abstract: Various systems and methods for implementing a service-level agreement (SLA) apparatus receive a request from a requester via a network interface of the gateway, the request comprising an inference model identifier that identifies a handler of the request, and a response time indicator. The response time indicator relates to a time within which the request is to be handled indicates an undefined time within which the request is to be handled. The apparatus determines a network location of a handler that is a platform or an inference model to handle the request consistent with the response time indicator, and routes the request to the handler at the network location.

    Distributed and contextualized artificial intelligence inference service

    公开(公告)号:US12217192B2

    公开(公告)日:2025-02-04

    申请号:US18091874

    申请日:2022-12-30

    Abstract: Various systems and methods of initiating and performing contextualized AI inferencing, are described herein. In an example, operations performed with a gateway computing device to invoke an inferencing model include receiving and processing a request for an inferencing operation, selecting an implementation of the inferencing model on a remote service based on a model specification and contextual data from the edge device, and executing the selected implementation of the inferencing model, such that results from the inferencing model are provided back to the edge device. Also in an example, operations performed with an edge computing device to request an inferencing model include collecting contextual data, generating an inferencing request, transmitting the inference request to a gateway device, and receiving and processing the results of execution. Further techniques for implementing a registration of the inference model, and invoking particular variants of an inference model, are also described.

    TECHNOLOGIES FOR TRANSPARENT FUNCTION AS A SERVICE ARBITRATION FOR EDGE SYSTEMS

    公开(公告)号:US20230396669A1

    公开(公告)日:2023-12-07

    申请号:US18234791

    申请日:2023-08-16

    CPC classification number: H04L67/1004 H04L41/12 H04L41/5006 H04L41/5019

    Abstract: Technologies for function as a service (FaaS) arbitration include an edge gateway, multiple endpoint devices, and multiple service providers. The edge gateway receives a registration request from a service provider that is indicative of an FaaS function identifier and a transform function. The edge gateway verifies an attestation received from the service provider and registers the service provider. The edge gateway receives a function execution request from an endpoint device that is indicative of the FaaS function identifier. The edge gateway selects the service provider based on the FaaS function identifier, programs an accelerator with the transform function, executes the transform function with the accelerator to transform the function execution request to a provider request, and submits the provider request to the service provider. The service provider may be selected based on an expected service level included in the function execution request. Other embodiments are described and claimed.

    Technologies for accelerating edge device workloads

    公开(公告)号:US11706158B2

    公开(公告)日:2023-07-18

    申请号:US17510077

    申请日:2021-10-25

    Abstract: Technologies for accelerating edge device workloads at a device edge network include a network computing device which includes a processor platform that includes at least one processor which supports a plurality of non-accelerated function-as-a-service (FaaS) operations and an accelerated platform that includes at least one accelerator which supports a plurality of accelerated FaaS (AFaaS) operation. The network computing device is configured to receive a request to perform a FaaS operation, determine whether the received request indicates that an AFaaS operation is to be performed on the received request, and identify compute requirements for the AFaaS operation to be performed. The network computing device is further configured to select an accelerator platform to perform the identified AFaaS operation and forward the received request to the selected accelerator platform to perform the identified AFaaS operation. Other embodiments are described and claimed.

    TECHNOLOGIES FOR ACCELERATING EDGE DEVICE WORKLOADS

    公开(公告)号:US20220224657A1

    公开(公告)日:2022-07-14

    申请号:US17510077

    申请日:2021-10-25

    Abstract: Technologies for accelerating edge device workloads at a device edge network include a network computing device which includes a processor platform that includes at least one processor which supports a plurality of non-accelerated function-as-a-service (FaaS) operations and an accelerated platform that includes at least one accelerator which supports a plurality of accelerated FaaS (AFaaS) operation. The network computing device is configured to receive a request to perform a FaaS operation, determine whether the received request indicates that an AFaaS operation is to be performed on the received request, and identify compute requirements for the AFaaS operation to be performed. The network computing device is further configured to select an accelerator platform to perform the identified AFaaS operation and forward the received request to the selected accelerator platform to perform the identified AFaaS operation. Other embodiments are described and claimed.

Patent Agency Ranking