Patents by Inventor Kai Fan Tang

Kai Fan Tang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240111832
    Abstract: A multitenant solver execution service provides managed infrastructure for defining and solving large-scale optimization problems. In embodiments, the service executes solver jobs on managed compute resources such as virtual machines or containers. The compute resources can be automatically scaled up or down based on client demand and are assigned to solver jobs in a serverless manner. Solver jobs can be initiated based on configured triggers. In embodiments, the service allows users to select from different types of solvers, mix different solvers in a solver job, and translate a model from one solver to another solver. In embodiments, the service provides developer interfaces to, for example, run solver experiments, recommend solver types or solver settings, and suggest model templates. The solver execution service relieves developers from having to manage infrastructure for running optimization solvers and allows developers to easily work with different types of solvers via a unified interface.
    Type: Application
    Filed: September 29, 2022
    Publication date: April 4, 2024
    Applicant: Amazon Technologies, Inc.
    Inventors: Shreyas Vathul Subramanian, Amey K Dhavle, Guvenc Degirmenci, Kai Fan Tang, Daniel Romero
  • Publication number: 20240111831
    Abstract: A multitenant solver execution service provides managed infrastructure for defining and solving large-scale optimization problems. In embodiments, the service executes solver jobs on managed compute resources such as virtual machines or containers. The compute resources can be automatically scaled up or down based on client demand and are assigned to solver jobs in a serverless manner. Solver jobs can be initiated based on configured triggers. In embodiments, the service allows users to select from different types of solvers, mix different solvers in a solver job, and translate a model from one solver to another solver. In embodiments, the service provides developer interfaces to, for example, run solver experiments, recommend solver types or solver settings, and suggest model templates. The solver execution service relieves developers from having to manage infrastructure for running optimization solvers and allows developers to easily work with different types of solvers via a unified interface.
    Type: Application
    Filed: September 29, 2022
    Publication date: April 4, 2024
    Applicant: Amazon Technologies, Inc.
    Inventors: Shreyas Vathul Subramanian, Amey K Dhavle, Guvenc Degirmenci, Kai Fan Tang, Daniel Romero
  • Publication number: 20240112067
    Abstract: A multitenant solver execution service provides managed infrastructure for defining and solving large-scale optimization problems. In embodiments, the service executes solver jobs on managed compute resources such as virtual machines or containers. The compute resources can be automatically scaled up or down based on client demand and are assigned to solver jobs in a serverless manner. Solver jobs can be initiated based on configured triggers. In embodiments, the service allows users to select from different types of solvers, mix different solvers in a solver job, and translate a model from one solver to another solver. In embodiments, the service provides developer interfaces to, for example, run solver experiments, recommend solver types or solver settings, and suggest model templates. The solver execution service relieves developers from having to manage infrastructure for running optimization solvers and allows developers to easily work with different types of solvers via a unified interface.
    Type: Application
    Filed: September 29, 2022
    Publication date: April 4, 2024
    Applicant: Amazon Technologies, Inc.
    Inventors: Shreyas Vathul Subramanian, Amey K Dhavle, Guvenc Degirmenci, Kai Fan Tang, Daniel Romero
  • Patent number: 11550635
    Abstract: Techniques are described for filtering and normalizing training data used to build a predictive auto scaling model used by a service provider network to proactively scale users' computing resources. Further described are techniques for identifying collections of computing resources that exhibit suitably predictable usage patterns such that a predictive auto scaling model can be used to forecast future usage patterns with reasonable accuracy and to scale the resources based on such generated forecasts. The filtering of training data and the identification of suitably predictable collections of computing resources are based in part on autocorrelation analyses, and in particular on “delayed” autocorrelation analyses, of time series data, among other techniques described herein.
    Type: Grant
    Filed: March 28, 2019
    Date of Patent: January 10, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Manwah Wong, Kai Fan Tang, Christopher Thomas Lewis
  • Patent number: 11347549
    Abstract: A notification for an application stack is received, where the application stack includes a plurality of resource types. At least one policy associated with the notification is obtained, with the first policy being a policy for scaling a first resource of a first resource type and a second resource of a second resource type of the application stack. A first capacity for the first resource and a second capacity for the second resource is determined based at least in part on the at least one policy. The first resource and the second resource are caused to be scaled according to the first capacity and the second capacity respectively.
    Type: Grant
    Filed: September 9, 2019
    Date of Patent: May 31, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Kai Fan Tang, Ahmed Usman Khalid
  • Patent number: 11249810
    Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
    Type: Grant
    Filed: March 22, 2019
    Date of Patent: February 15, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Jacob Adam Gabrielson, Joshua M. Burgin, Brad Bonnett, Kai Fan Tang
  • Patent number: 11243794
    Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
    Type: Grant
    Filed: March 22, 2019
    Date of Patent: February 8, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Jacob Adam Gabrielson, Joshua M. Burgin, Brad Bonnett, Kai Fan Tang
  • Patent number: 11221887
    Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
    Type: Grant
    Filed: March 22, 2019
    Date of Patent: January 11, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Jacob Adam Gabrielson, Joshua M. Burgin, Brad Bonnett, Kai Fan Tang
  • Patent number: 11194688
    Abstract: Techniques for an optimization service of a service provider network to generate an architecture diagram that represents an architecture of a web-based application. The optimization service may use the architecture diagram to determine modifications or changes to make to the application. For example, the optimization service may compare the architecture diagram with optimized architecture diagrams that represent application best practices, and determine the modifications or change to make to the application to optimize the application and bring the application in-line with best practices. Further, the optimization service may use the architecture diagram to generate a visualization, and provide the user account with the visualization of the architecture diagram to show users their application architecture.
    Type: Grant
    Filed: May 8, 2019
    Date of Patent: December 7, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Malcolm Featonby, Jacob Adam Gabrielson, Kai Fan Tang, John Merrill Phillips, Leslie Johann Lamprecht, Letian Feng, Roberto Pentz De Faria
  • Patent number: 10979436
    Abstract: A policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the policy and the first current capacity. A second request is submitted, to the service, to adjust the capacity of the resource by the amount. A third request is submitted, to the service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the response to the third request.
    Type: Grant
    Filed: August 23, 2019
    Date of Patent: April 13, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Christopher Thomas Lewis, Kai Fan Tang, Farzad Moghimi, Ahmed Usman Khalid, Stephan Weinwurm
  • Publication number: 20200301723
    Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
    Type: Application
    Filed: March 22, 2019
    Publication date: September 24, 2020
    Inventors: Jacob Adam GABRIELSON, Joshua M. BURGIN, Brad BONNETT, Kai Fan TANG
  • Publication number: 20200301740
    Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
    Type: Application
    Filed: March 22, 2019
    Publication date: September 24, 2020
    Inventors: Jacob Adam GABRIELSON, Joshua M. BURGIN, Brad BONNETT, Kai Fan TANG
  • Publication number: 20200301741
    Abstract: Techniques are described for optimizing the allocation of computing resources provided by a service provider network—for example, compute resources such as virtual machine (VM) instances, containers, standalone servers, and possibly other types of computing resources—among computing workloads associated with a user or group of users of the service provider network. A service provider network provides various tools and interfaces to help businesses and other organizations optimize the utilization of computing resource pools obtained by the organizations from the service provider network, including the ability to efficiently schedule use of the resources among workloads having varying resource demands, usage patterns, relative priorities, execution deadlines, or combinations thereof. A service provider network further provides various graphical user interfaces (GUIs) to help users visualize and manage the historical and scheduled uses of computing resources by users' workloads according to user preferences.
    Type: Application
    Filed: March 22, 2019
    Publication date: September 24, 2020
    Inventors: Jacob Adam GABRIELSON, Joshua M. BURGIN, Brad BONNETT, Kai Fan TANG
  • Publication number: 20200004590
    Abstract: A notification for an application stack is received, where the application stack includes a plurality of resource types. At least one policy associated with the notification is obtained, with the first policy being a policy for scaling a first resource of a first resource type and a second resource of a second resource type of the application stack. A first capacity for the first resource and a second capacity for the second resource is determined based at least in part on the at least one policy. The first resource and the second resource are caused to be scaled according to the first capacity and the second capacity respectively.
    Type: Application
    Filed: September 9, 2019
    Publication date: January 2, 2020
    Inventors: Kai Fan Tang, Ahmed Usman Khalid
  • Publication number: 20190379674
    Abstract: A policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the policy and the first current capacity. A second request is submitted, to the service, to adjust the capacity of the resource by the amount. A third request is submitted, to the service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the response to the third request.
    Type: Application
    Filed: August 23, 2019
    Publication date: December 12, 2019
    Inventors: Christopher Thomas Lewis, Kai Fan Tang, Farzad Moghimi, Ahmed Usman Khalid, Stephan Weinwurm
  • Patent number: 10412022
    Abstract: A scaling notification is received that includes a current capacity for a scalable target of a resource. A policy associated with the scaling notification is obtained. A new capacity is computed for the scalable target, based at least in part on the policy and the current capacity of the scalable target. A scaling request is made to an application programming interface endpoint for the scalable target, with the request including a new capacity for the scalable target, thereby causing the current capacity of the scalable target to change to the new capacity.
    Type: Grant
    Filed: October 19, 2016
    Date of Patent: September 10, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Kai Fan Tang, Christopher Thomas Lewis, Farzad Moghimi, Stephan Weinwurm, Nayanish Hinge
  • Patent number: 10409642
    Abstract: A notification for an application stack is received, where the application stack includes a plurality of resource types. At least one policy associated with the notification is obtained, with the first policy being a policy for scaling a first resource of a first resource type and a second resource of a second resource type of the application stack. A first capacity for the first resource and a second capacity for the second resource is determined based at least in part on the at least one policy. The first resource and the second resource are caused to be scaled according to the first capacity and the second capacity respectively.
    Type: Grant
    Filed: November 22, 2016
    Date of Patent: September 10, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Kai Fan Tang, Ahmed Usman Khalid
  • Patent number: 10397240
    Abstract: A scaling policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a software container service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the scaling policy and the first current capacity. A second request is submitted, to the software container service, to adjust the capacity of the resource by the amount. A third request is submitted, to the software container service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the amount.
    Type: Grant
    Filed: November 19, 2018
    Date of Patent: August 27, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Christopher Thomas Lewis, Kai Fan Tang, Farzad Moghimi, Ahmed Usman Khalid, Stephan Weinwurm
  • Publication number: 20190089714
    Abstract: A scaling policy associated with a notification received by one or more computer systems is obtained. A first request is submitted, to a software container service, for a first current capacity of a resource. An amount by which to adjust a capacity of the resource is calculated, based at least in part on the scaling policy and the first current capacity. A second request is submitted, to the software container service, to adjust the capacity of the resource by the amount. A third request is submitted, to the software container service, for a second current capacity of the resource, and whether the second request has been fulfilled is determined based at least in part on a comparison between the second current capacity and the amount.
    Type: Application
    Filed: November 19, 2018
    Publication date: March 21, 2019
    Inventors: Christopher Thomas Lewis, Kai Fan Tang, Farzad Moghimi, Ahmed Usman Khalid, Stephan Weinwurm
  • Patent number: 10148592
    Abstract: Techniques are described for scaling a group of computing resources. A computing resource service receives a scaling policy for use in scaling the group of computing resources. The scaling policy specifies a target level for a resource utilization metric and magnitude-based changes to the group. The computing resource service receives information about a magnitude of a measurement for the resource utilization metric. The computing resource service determines, based at least in part on the scaling policy, one or more changes for the group and initiates the one or more changes in the group.
    Type: Grant
    Filed: June 29, 2015
    Date of Patent: December 4, 2018
    Assignee: Amazon Technologies, Inc.
    Inventors: Derek Solomon Pai, Alison Qing-Ning Truong, Eric Samuel Stone, Ahmed Usman Khalid, Kai Fan Tang