Patents by Inventor Gal Dalal

Gal Dalal has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11880261
    Abstract: A system, method, and apparatus of power management for computing systems are included herein that optimize individual frequencies of components of the computing systems using machine learning. The computing systems can be tightly integrated systems that consider an overall operating budget that is shared between the components of the computing system while adjusting the frequencies of the individual components. An example of an automated method of power management includes: (1) learning, using a power management (PM) agent, frequency settings for different components of a computing system during execution of a repetitive application, and (2) adjusting the frequency settings of the different components using the PM agent, wherein the adjusting is based on the repetitive application and one or more limitations corresponding to a shared operating budget for the computing system.
    Type: Grant
    Filed: March 31, 2022
    Date of Patent: January 23, 2024
    Assignee: NVIDIA Corporation
    Inventors: Evgeny Bolotin, Yaosheng Fu, Zi Yan, Gal Dalal, Shie Mannor, David Nellans
  • Publication number: 20240007403
    Abstract: In various embodiments, a congestion control modelling application automatically controls congestion in data transmission networks. The congestion control modelling application executes a trained neural network in conjunction with a simulated data transmission network to generate a training dataset. The trained neural network has been trained to control congestion in the simulated data transmission network. The congestion control modelling application generates a first trained decision tree model based on an initial loss for an initial model relative to the training dataset. The congestion control modelling application generates a final tree-based model based on the first trained decision tree model and at least a second trained decision tree model. The congestion control modelling application executes the final tree-based model in conjunction with a data transmission network to control congestion within the data transmission network.
    Type: Application
    Filed: April 11, 2023
    Publication date: January 4, 2024
    Inventors: Gal CHECHIK, Gal DALAL, Benjamin FUHRER, Doron HARITAN KAZAKOV, Shie MANNOR, Yuval SHPIGELMAN, Chen TESSLER
  • Publication number: 20230237342
    Abstract: A method is performed by an agent operating in an environment. The method comprises computing a first value associated with each state of a number of states in the environment, determining a lookahead horizon for each state of the number of states in the environment based on the computed first value for each state of the number of states, applying a first policy to compute a second value associated with each state of at least one state in the number of states in the environment for the at least one state in the number of states based on the determined lookahead horizons for the number of states, and determining a second policy based on the first policy and the second value for each state of the number of states in the environment.
    Type: Application
    Filed: January 24, 2023
    Publication date: July 27, 2023
    Inventors: Shie Mannor, Gal Chechik, Gal Dalal, Assaf Joseph Hallak, Aviv Rosenberg
  • Publication number: 20230137205
    Abstract: Introduced herein is a technique that uses ML to autonomously find a cache management policy that achieves an optimal execution of a given workload of an application. Leveraging ML such as reinforcement learning, the technique trains an agent in an ML environment over multiple episodes of a stabilization process. For each time step in these training episodes, the agent executes the application while making an incremental change to the current policy, i.e., cache-residency statuses of memory address space associated with the workload, until the application can be executed at a stable level. The stable level of execution, for example, can be indicated by performance variations, such as standard deviations, between a certain number of neighboring measurement periods remaining within a certain threshold. The agent, who has been trained in the training episodes, infers the final cache management policy during the final, inferring episode.
    Type: Application
    Filed: October 29, 2021
    Publication date: May 4, 2023
    Inventors: Yaosheng Fu, Shie Mannor, Evgeny Bolotin, David Nellans, Gal Dalal
  • Publication number: 20230079978
    Abstract: A system, method, and apparatus of power management for computing systems are included herein that optimize individual frequencies of components of the computing systems using machine learning. The computing systems can be tightly integrated systems that consider an overall operating budget that is shared between the components of the computing system while adjusting the frequencies of the individual components. An example of an automated method of power management includes: (1) learning, using a power management (PM) agent, frequency settings for different components of a computing system during execution of a repetitive application, and (2) adjusting the frequency settings of the different components using the PM agent, wherein the adjusting is based on the repetitive application and one or more limitations corresponding to a shared operating budget for the computing system.
    Type: Application
    Filed: March 31, 2022
    Publication date: March 16, 2023
    Inventors: Evgeny Bolotin, Yaosheng Fu, Zi Yan, Gal Dalal, Shie Mannor, David Nellans
  • Publication number: 20230041242
    Abstract: A reinforcement learning agent learns a congestion control policy using a deep neural network and a distributed training component. The training component enables the agent to interact with a vast set of environments in parallel. These environments simulate real world benchmarks and real hardware. During a learning process, the agent learns how maximize an objective function. A simulator may enable parallel interaction with various scenarios. As the trained agent encounters a diverse set of problems it is more likely to generalize well to new and unseen environments. In addition, an operating point can be selected during training which may enable configuration of the required behavior of the agent.
    Type: Application
    Filed: October 3, 2022
    Publication date: February 9, 2023
    Inventors: Shie Mannor, Chen Tessler, Yuval Shpigelman, Amit Mandelbaum, Gal Dalal, Doron Kazakov, Benjamin Fuhrer
  • Publication number: 20220398283
    Abstract: A method for performing a Tree-Search (TS) on an environment is provided. The method comprises generating a tree for a current state of the environment based on a TS policy, determining a corrected TS policy, and determining an action to apply to the environment based on the corrected TS policy. The tree comprises a plurality of nodes including a root node among the plurality of nodes corresponding to the current state of the environment. Each node other than the root node among the plurality of nodes corresponding to an estimated future state of the environment. The plurality of nodes in the tree are connected by a plurality of edges. Each edge among the plurality of edges is associated with an action causing a transition from a first state to a different sate of the environment.
    Type: Application
    Filed: May 25, 2022
    Publication date: December 15, 2022
    Inventors: Shie Mannor, Assaf Joseph Hallak, Gal Dalal, Steven Tarence Dalton, Iuri Frosio, Gal Chechik
  • Publication number: 20220231933
    Abstract: A reinforcement learning agent learns a congestion control policy using a deep neural network and a distributed training component. The training component enables the agent to interact with a vast set of environments in parallel. These environments simulate real world benchmarks and real hardware. During a learning process, the agent learns how maximize an objective function. A simulator may enable parallel interaction with various scenarios. As the trained agent encounters a diverse set of problems it is more likely to generalize well to new and unseen environments. In addition, an operating point can be selected during training which may enable configuration of the required behavior of the agent.
    Type: Application
    Filed: June 7, 2021
    Publication date: July 21, 2022
    Inventors: Shie Mannor, Chen Tessler, Yuval Shpigelman, Amit Mandelbaum, Gal Dalal, Doron Kazakov, Benjamin Fuhrer