-
公开(公告)号:US11249810B2
公开(公告)日:2022-02-15
申请号:US16362545
申请日:2019-03-22
Applicant: Amazon Technologies, Inc.
Inventor: Jacob Adam Gabrielson , Joshua M. Burgin , Brad Bonnett , Kai Fan Tang
Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
-
公开(公告)号:US11194688B1
公开(公告)日:2021-12-07
申请号:US16406354
申请日:2019-05-08
Applicant: Amazon Technologies, Inc.
Inventor: Malcolm Featonby , Jacob Adam Gabrielson , Kai Fan Tang , John Merrill Phillips , Leslie Johann Lamprecht , Letian Feng , Roberto Pentz De Faria
Abstract: Techniques for an optimization service of a service provider network to generate an architecture diagram that represents an architecture of a web-based application. The optimization service may use the architecture diagram to determine modifications or changes to make to the application. For example, the optimization service may compare the architecture diagram with optimized architecture diagrams that represent application best practices, and determine the modifications or change to make to the application to optimize the application and bring the application in-line with best practices. Further, the optimization service may use the architecture diagram to generate a visualization, and provide the user account with the visualization of the architecture diagram to show users their application architecture.
-
公开(公告)号:US10979436B2
公开(公告)日:2021-04-13
申请号:US16550096
申请日:2019-08-23
Applicant: Amazon Technologies, Inc.
Inventor: Christopher Thomas Lewis , Kai Fan Tang , Farzad Moghimi , Ahmed Usman Khalid , Stephan Weinwurm
Abstract: A policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the policy and the first current capacity. A second request is submitted, to the service, to adjust the capacity of the resource by the amount. A third request is submitted, to the service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the response to the third request.
-
公开(公告)号:US20190089714A1
公开(公告)日:2019-03-21
申请号:US16195645
申请日:2018-11-19
Applicant: Amazon Technologies, Inc.
Inventor: Christopher Thomas Lewis , Kai Fan Tang , Farzad Moghimi , Ahmed Usman Khalid , Stephan Weinwurm
CPC classification number: H04L63/104 , H04L63/083 , H04L63/105 , H04L63/205 , H04L67/1025 , H04Q9/02
Abstract: A scaling policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a software container service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the scaling policy and the first current capacity. A second request is submitted, to the software container service, to adjust the capacity of the resource by the amount. A third request is submitted, to the software container service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the amount.
-
公开(公告)号:US10135837B2
公开(公告)日:2018-11-20
申请号:US15194486
申请日:2016-06-27
Applicant: Amazon Technologies, Inc.
Inventor: Christopher Thomas Lewis , Kai Fan Tang , Farzad Moghimi , Ahmed Usman Khalid , Stephan Weinwurm
Abstract: A scaling policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a software container service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the scaling policy and the first current capacity. A second request is submitted, to the software container service, to adjust the capacity of the resource by the amount. A third request is submitted, to the software container service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the amount.
-
公开(公告)号:US20170339158A1
公开(公告)日:2017-11-23
申请号:US15194486
申请日:2016-06-27
Applicant: Amazon Technologies, Inc.
Inventor: Christopher Thomas Lewis , Kai Fan Tang , Farzad Moghimi , Ahmed Usman Khalid , Stephan Weinwurm
CPC classification number: H04L63/205 , H04L63/083 , H04L63/104 , H04L63/105 , H04L67/1025 , H04Q9/02
Abstract: A scaling policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a software container service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the scaling policy and the first current capacity. A second request is submitted, to the software container service, to adjust the capacity of the resource by the amount. A third request is submitted, to the software container service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the amount.
-
公开(公告)号:US20240112067A1
公开(公告)日:2024-04-04
申请号:US17936793
申请日:2022-09-29
Applicant: Amazon Technologies, Inc.
Inventor: Shreyas Vathul Subramanian , Amey K Dhavle , Guvenc Degirmenci , Kai Fan Tang , Daniel Romero
Abstract: A multitenant solver execution service provides managed infrastructure for defining and solving large-scale optimization problems. In embodiments, the service executes solver jobs on managed compute resources such as virtual machines or containers. The compute resources can be automatically scaled up or down based on client demand and are assigned to solver jobs in a serverless manner. Solver jobs can be initiated based on configured triggers. In embodiments, the service allows users to select from different types of solvers, mix different solvers in a solver job, and translate a model from one solver to another solver. In embodiments, the service provides developer interfaces to, for example, run solver experiments, recommend solver types or solver settings, and suggest model templates. The solver execution service relieves developers from having to manage infrastructure for running optimization solvers and allows developers to easily work with different types of solvers via a unified interface.
-
公开(公告)号:US11243794B2
公开(公告)日:2022-02-08
申请号:US16362542
申请日:2019-03-22
Applicant: Amazon Technologies, Inc.
Inventor: Jacob Adam Gabrielson , Joshua M. Burgin , Brad Bonnett , Kai Fan Tang
IPC: G06F9/455 , G06F9/50 , G06F9/48 , G06N3/04 , H04L12/24 , H04L29/08 , G06N3/08 , G06F3/0484 , G06F3/0486
Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
-
19.
公开(公告)号:US20200004590A1
公开(公告)日:2020-01-02
申请号:US16565051
申请日:2019-09-09
Applicant: Amazon Technologies, Inc.
Inventor: Kai Fan Tang , Ahmed Usman Khalid
IPC: G06F9/50
Abstract: A notification for an application stack is received, where the application stack includes a plurality of resource types. At least one policy associated with the notification is obtained, with the first policy being a policy for scaling a first resource of a first resource type and a second resource of a second resource type of the application stack. A first capacity for the first resource and a second capacity for the second resource is determined based at least in part on the at least one policy. The first resource and the second resource are caused to be scaled according to the first capacity and the second capacity respectively.
-
公开(公告)号:US10412022B1
公开(公告)日:2019-09-10
申请号:US15298152
申请日:2016-10-19
Applicant: Amazon Technologies, Inc.
Inventor: Kai Fan Tang , Christopher Thomas Lewis , Farzad Moghimi , Stephan Weinwurm , Nayanish Hinge
IPC: G06F15/173 , H04L12/911 , H04L12/26 , H04L12/24
Abstract: A scaling notification is received that includes a current capacity for a scalable target of a resource. A policy associated with the scaling notification is obtained. A new capacity is computed for the scalable target, based at least in part on the policy and the current capacity of the scalable target. A scaling request is made to an application programming interface endpoint for the scalable target, with the request including a new capacity for the scalable target, thereby causing the current capacity of the scalable target to change to the new capacity.
-
-
-
-
-
-
-
-
-