-
公开(公告)号:US11805027B2
公开(公告)日:2023-10-31
申请号:US17710853
申请日:2022-03-31
Applicant: Amazon Technologies, Inc.
Inventor: Maximiliano Maccanti , Gowda Dayananda Anjaneyapura Range , Rishabh Ray Chaudhury , Michael Pham , Shruti Sharma , Saumitra Vikram , James Alan Sanders , Mihir Sathe
Abstract: A serverless computing system is configured to provide access to a machine learning model by at least associating an endpoint, comprising code that accesses the machine learning model, with an extension that interfaces between a serverless compute architecture and the endpoint. A request to perform an inference is received by the system and processed by using the serverless compute architecture to execute a compute function. The compute function cases the extension to interface with the endpoint to cause the machine learning model to perform the inference.
-
公开(公告)号:US20230171164A1
公开(公告)日:2023-06-01
申请号:US17710853
申请日:2022-03-31
Applicant: Amazon Technologies, Inc.
Inventor: Maximiliano Maccanti , Gowda Dayananda Anjaneyapura Range , Rishabh Ray Chaudhury , Michael Pham , Shruti Sharma , Saumitra Vikram , James Alan Sanders , Mihir Sathe
Abstract: A serverless computing system is configured to provide access to a machine learning model by at least associating an endpoint, comprising code that accesses the machine learning model, with an extension that interfaces between a serverless compute architecture and the endpoint. A request to perform an inference is received by the system and processed by using the serverless compute architecture to execute a compute function. The compute function cases the extension to interface with the endpoint to cause the machine learning model to perform the inference.
-
公开(公告)号:US11550713B1
公开(公告)日:2023-01-10
申请号:US17105260
申请日:2020-11-25
Applicant: Amazon Technologies, Inc.
Inventor: Philip Daniel Piwonka , Mihir Sathe , Roger J. Tragin , Dmitry Kravtsov
IPC: G06F12/02
Abstract: Systems and methods are described for enabling garbage collection on data storage systems. Traditional garbage collection often attempts to track use of data items on an individual level, deleting each item when it is no longer used. In distributed systems, tracking use on an individual level is difficult, and may require centralized knowledge across the system with respect to individual data items. Provided herein is a “coarse-grained” garbage collection mechanism, which divides objects into logical groups referred to as “roots.” Each root has a life cycle. While active, new data can be stored in a root. While inactive, use of data within a root can cause that date to be copied to a different, active root. When the system detects that data hasn't been used in an inactive root for a threshold period, the root can be considered “dead” and data within the root may be deleted.
-
公开(公告)号:US11449777B1
公开(公告)日:2022-09-20
申请号:US16915916
申请日:2020-06-29
Applicant: Amazon Technologies, Inc.
Inventor: Mihir Sathe
Abstract: Systems and methods are described for providing for serverless inferences against a trained machine learning (ML) model. Rather than obtaining one or more dedicated devices to conduct inferences, users are enabled to create a task on a serverless system that, when invoked, passing input data to a trained ML model and provides a result. To satisfy varying user requirements for inference speed, the system includes a variety of hardware configurations. The system can efficiently allocate resources between different tasks by invoking the task on a particular hardware configuration that is selected based on a current availability of the selected hardware configuration to host an execution environment in which the task is implemented and an expected time to invoke the task on the individual hardware configuration. The system can therefore efficiently allocate resources among inferences using a variety of different ML models.
-
公开(公告)号:US11863613B1
公开(公告)日:2024-01-02
申请号:US17209008
申请日:2021-03-22
Applicant: Amazon Technologies, Inc.
Inventor: Mihir Sathe , Aravind Srinivasan , Pranav Rao Perampalli Nekkar
IPC: H04L67/1008
CPC classification number: H04L67/1008
Abstract: Systems and methods are described for allocating requests to implement new workloads within a dynamic set of servers. Existing load balancing techniques can result in “focus firing” on new servers added to the set, since a load balancer may view a new server as underloaded. With sufficient intensity, focus firing can result in overshooting target load for the new server, and the new server in fact becoming overloaded. The present disclosure modifies selection of servers as potential targets for a workload by at least partly biasing against selection of young servers. The bias imposed can be scaled to avoid overloading new servers.
-
6.
公开(公告)号:US11775640B1
公开(公告)日:2023-10-03
申请号:US16835166
申请日:2020-03-30
Applicant: Amazon Technologies, Inc.
Inventor: Mihir Sathe , Niall Mullen
CPC classification number: G06F21/566 , G06F21/54 , G06F21/552 , G06F9/5005
Abstract: Systems and methods are described for detecting and preventing execution of malware on an on-demand code execution system. An on-demand code execution system may execute user-submitted code on virtual machine instances, which may be provisioned with various computing resources (memory, storage, processors, network bandwidth, etc.). These resources may be utilized in varying amounts or at varying rates during execution of the user-submitted code. The user-submitted code may also be unavailable for inspection for security or other reasons. A malware detection system may thus identify user-submitted code that corresponds to malware by monitoring resource utilization during execution of the code and generating a resource utilization signature, which enables comparison between the signature of the user-submitted code and resource utilization signatures of codes previously identified as malware. The malware detection system may then take actions such as notifying the user who requested execution or preventing execution of the user-submitted code.
-
公开(公告)号:US20230169396A1
公开(公告)日:2023-06-01
申请号:US17710864
申请日:2022-03-31
Applicant: Amazon Technologies, Inc.
Inventor: Maximiliano Maccanti , Gowda Dayananda Anjaneyapura Range , Rishabh Ray Chaudhury , Michael Pham , Shruti Sharma , Saumitra Vikram , James Alan Sanders , Mihir Sathe
IPC: G06N20/00
CPC classification number: G06N20/00
Abstract: A system is configured to provide access to a machine learning model by using a hybrid configuration comprising a dedicate server on which an instance of a model server is installed, and a serverless compute architecture that interfaces with an instance of the model server using an extension. A first portion of requests directed to the model server are processed by the dedicated server, and a second portion of the requests is processed by the serverless compute architecture.
-
公开(公告)号:US20180032510A1
公开(公告)日:2018-02-01
申请号:US14671864
申请日:2015-03-27
Applicant: Amazon Technologies, Inc.
Inventor: Mihir Sathe , Paul Andrew Lafranchise , Joseph Barry Guglielmo
IPC: G06F17/28
CPC classification number: G06F17/289 , G06F9/454
Abstract: In some cases, a localization service may identify candidate strings in the source code of an application. Further, the localization service may determine whether the candidate strings are displayed literals in a first human-perceivable language. In addition, the localization service may replace the identified displayed literals with identification tokens to generate pivot source code. In some examples, an identification token may include a JavaScript function that returns a translation of a displayed literal in a second human-perceivable language or any other desired human-perceivable language. Further, the localization service may verify pivot source code by comparing a localized application corresponding to the pivot source code to the application with the original source code of the application.
-
-
-
-
-
-
-