Patents by Inventor Roland Hafner
Roland Hafner has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240062035Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.Type: ApplicationFiled: July 12, 2023Publication date: February 22, 2024Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
-
Patent number: 11893480Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for reinforcement learning with scheduled auxiliary tasks. In one aspect, a method includes maintaining data specifying parameter values for a primary policy neural network and one or more auxiliary neural networks; at each of a plurality of selection time steps during a training episode comprising a plurality of time steps: receiving an observation, selecting a current task for the selection time step using a task scheduling policy, processing an input comprising the observation using the policy neural network corresponding to the selected current task to select an action to be performed by the agent in response to the observation, and causing the agent to perform the selected action.Type: GrantFiled: February 28, 2019Date of Patent: February 6, 2024Assignee: DeepMind Technologies LimitedInventors: Martin Riedmiller, Roland Hafner
-
Patent number: 11741334Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.Type: GrantFiled: May 22, 2020Date of Patent: August 29, 2023Assignee: DeepMind Technologies LimitedInventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
-
Publication number: 20220237488Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for controlling an agent. One of the methods includes obtaining an observation characterizing a current state of the environment and data identifying a task currently being performed by the agent; processing the observation and the data identifying the task using a high-level controller to generate a high-level probability distribution that assigns a respective probability to each of a plurality of low-level controllers; processing the observation using each of the plurality of low-level controllers to generate, for each of the plurality of low-level controllers, a respective low-level probability distribution; generating a combined probability distribution; and selecting, using the combined probability distribution, an action from the space of possible actions to be performed by the agent in response to the observation.Type: ApplicationFiled: May 22, 2020Publication date: July 28, 2022Inventors: Markus Wulfmeier, Abbas Abdolmaleki, Roland Hafner, Jost Tobias Springenberg, Nicolas Manfred Otto Heess, Martin Riedmiller
-
Publication number: 20200285909Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.Type: ApplicationFiled: May 22, 2020Publication date: September 10, 2020Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
-
Patent number: 10664725Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.Type: GrantFiled: July 31, 2019Date of Patent: May 26, 2020Assignee: DeepMind Technologies LimitedInventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
-
Publication number: 20190354813Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.Type: ApplicationFiled: July 31, 2019Publication date: November 21, 2019Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
-
Patent number: 9528612Abstract: An explosion-proof sliding gate valve (1) for blocking a fluid flow (F) in a pipeline, which comprises: a sliding gate valve housing (2), a sealing plate (3) which is movably mounted in the slide gate valve housing (2) between an operating position (B) and a closed position (S), wherein in the closed position (S) the sealing plate (3) blocks the fluid flow (F), a mechanism for moving the sealing plate (3) from the operating position (B) into the closed position (S), and at least one deformation body (4) for damping movement of the sealing plate. The deformation body is arranged, according to the invention, at the side between sliding gate valve housing (2) and sealing plate (3).Type: GrantFiled: May 20, 2015Date of Patent: December 27, 2016Assignee: Rico-Sicherheitstechnik AGInventors: Daniel Zellweger, Roland Hafner, Giuseppe Walter Aloi
-
Publication number: 20150337967Abstract: An explosion-proof sliding gate valve (1) for blocking a fluid flow (F) in a pipeline, which comprises: a sliding gate valve housing (2), a sealing plate (3) which is movably mounted in the slide gate valve housing (2) between an operating position (B) and a closed position (S), wherein in the closed position (S) the sealing plate (3) blocks the fluid flow (F), a mechanism for moving the sealing plate (3) from the operating position (B) into the closed position (S), and at least one deformation body (4) for damping movement of the sealing plate. The deformation body is arranged, according to the invention, at the side between sliding gate valve housing (2) and sealing plate (3).Type: ApplicationFiled: May 20, 2015Publication date: November 26, 2015Inventors: Daniel ZELLWEGER, Roland HAFNER, Giuseppe Walter ALOI
-
Patent number: 8096284Abstract: In a method for determining a rail pressure setpoint value for a high-pressure rail of an internal combustion engine, the rail pressure setpoint value is modified to a maximum degree using a maximum gradient for changing the rail pressure setpoint value, and the maximum gradient is read from a characteristics map as a function of operating parameters of the internal combustion engine. The operating parameters include an engaged gear of a gear-change transmission.Type: GrantFiled: July 16, 2007Date of Patent: January 17, 2012Assignee: Robert Bosch GmbHInventors: Stefan Koidl, Guido Baumann, Antoine Combelle, Anthony Dieryckxvisschers, Jean-Daniel Mettetal, Pierre Mathis, Enrique Naupari, Martin Schwab, Roland Hafner
-
Publication number: 20090320798Abstract: In a method for determining a rail pressure setpoint value for a high-pressure rail of an internal combustion engine, the rail pressure setpoint value is modified to a maximum degree using a maximum gradient for changing the rail pressure setpoint value, and the maximum gradient is read from a characteristics map as a function of operating parameters of the internal combustion engine. The operating parameters include an engaged gear of a gear-change transmission.Type: ApplicationFiled: July 16, 2007Publication date: December 31, 2009Inventors: Stefan Koidl, Guido Baumann, Antoine Combelle, Anthony Dieryckxvisschers, Jean-Daniel Mettetal, Pierre Mathis, Enrique Naupari, Martin Schwab, Roland Hafner