Patents by Inventor Thomas Lampe

Thomas Lampe has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240062035
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.
    Type: Application
    Filed: July 12, 2023
    Publication date: February 22, 2024
    Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
  • Patent number: 11886997
    Abstract: An off-policy reinforcement learning actor-critic neural network system configured to select actions from a continuous action space to be performed by an agent interacting with an environment to perform a task. An observation defines environment state data and reward data. The system has an actor neural network which learns a policy function mapping the state data to action data. A critic neural network learns an action-value (Q) function. A replay buffer stores tuples of the state data, the action data, the reward data and new state data. The replay buffer also includes demonstration transition data comprising a set of the tuples from a demonstration of the task within the environment. The neural network system is configured to train the actor neural network and the critic neural network off-policy using stored tuples from the replay buffer comprising tuples both from operation of the system and from the demonstration transition data.
    Type: Grant
    Filed: October 7, 2022
    Date of Patent: January 30, 2024
    Assignee: DeepMind Technologies Limited
    Inventors: Olivier Pietquin, Martin Riedmiller, Wang Fumin, Bilal Piot, Mel Vecerik, Todd Andrew Hester, Thomas Rothoerl, Thomas Lampe, Nicolas Manfred Otto Heess, Jonathan Karl Scholz
  • Patent number: 11868882
    Abstract: An off-policy reinforcement learning actor-critic neural network system configured to select actions from a continuous action space to be performed by an agent interacting with an environment to perform a task. An observation defines environment state data and reward data. The system has an actor neural network which learns a policy function mapping the state data to action data. A critic neural network learns an action-value (Q) function. A replay buffer stores tuples of the state data, the action data, the reward data and new state data. The replay buffer also includes demonstration transition data comprising a set of the tuples from a demonstration of the task within the environment. The neural network system is configured to train the actor neural network and the critic neural network off-policy using stored tuples from the replay buffer comprising tuples both from operation of the system and from the demonstration transition data.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: January 9, 2024
    Assignee: DeepMind Technologies Limited
    Inventors: Olivier Claude Pietquin, Martin Riedmiller, Wang Fumin, Bilal Piot, Mel Vecerik, Todd Andrew Hester, Thomas Rothoerl, Thomas Lampe, Nicolas Manfred Otto Heess, Jonathan Karl Scholz
  • Patent number: 11835409
    Abstract: A pressure sensor apparatus and a method of operating the pressure sensor apparatus can include a hybrid pressure switch having a single switching point, the hybrid pressure switch comprising: a mechanical pressure switch; and an electronic pressure sensor, wherein the hybrid pressure switch monitors a media pressure with the mechanical pressure switch providing a switching function and the electronic pressure sensor providing a continuous pressure measurement with respect to the media pressure based on a continuous pressure output signal. The switching function can comprise a fail-safe switching function or a non-fail-safe switching function.
    Type: Grant
    Filed: December 26, 2020
    Date of Patent: December 5, 2023
    Assignee: Honeywell International Inc.
    Inventors: Christian Ruwe, Thomas Lampe, Tobias Otterpohl, Thomas Heyen, David Kucera
  • Patent number: 11741334
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.
    Type: Grant
    Filed: May 22, 2020
    Date of Patent: August 29, 2023
    Assignee: DeepMind Technologies Limited
    Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
  • Patent number: 11686601
    Abstract: A sensor system and method of operating the sensor system can include an indicator that is operable in resonance, the indicator being connected to a movable element of an actuator, and a sensor including sensor windings arranged in direct proximity to the indicator and external to a closed housing. The sensor can receive a position or an angle of the indicator, and can measure the position or the angle of the movable element inside the actuator through the closed housing based on inductive resonance facilitated by the indicator.
    Type: Grant
    Filed: September 18, 2020
    Date of Patent: June 27, 2023
    Assignee: Honeywell International Inc.
    Inventors: Christian Kamp, Thomas Lampe, Tobias Fechtel, Mathias Grewe
  • Publication number: 20230023189
    Abstract: An off-policy reinforcement learning actor-critic neural network system configured to select actions from a continuous action space to be performed by an agent interacting with an environment to perform a task. An observation defines environment state data and reward data. The system has an actor neural network which learns a policy function mapping the state data to action data. A critic neural network learns an action-value (Q) function. A replay buffer stores tuples of the state data, the action data, the reward data and new state data. The replay buffer also includes demonstration transition data comprising a set of the tuples from a demonstration of the task within the environment. The neural network system is configured to train the actor neural network and the critic neural network off-policy using stored tuples from the replay buffer comprising tuples both from operation of the system and from the demonstration transition data.
    Type: Application
    Filed: October 7, 2022
    Publication date: January 26, 2023
    Inventors: Olivier Pietquin, Martin Riedmiller, Wang Fumin, Bilal Piot, Mel Vecerik, Todd Andrew Hester, Thomas Rothoerl, Thomas Lampe, Nicolas Manfred Otto Heess, Jonathan Karl Scholz
  • Patent number: 11468321
    Abstract: An off-policy reinforcement learning actor-critic neural network system configured to select actions from a continuous action space to be performed by an agent interacting with an environment to perform a task. An observation defines environment state data and reward data. The system has an actor neural network which learns a policy function mapping the state data to action data. A critic neural network learns an action-value (Q) function. A replay buffer stores tuples of the state data, the action data, the reward data and new state data. The replay buffer also includes demonstration transition data comprising a set of the tuples from a demonstration of the task within the environment. The neural network system is configured to train the actor neural network and the critic neural network off-policy using stored tuples from the replay buffer comprising tuples both from operation of the system and from the demonstration transition data.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: October 11, 2022
    Assignee: DeepMind Technologies Limited
    Inventors: Olivier Claude Pietquin, Martin Riedmiller, Wang Fumin, Bilal Piot, Mel Vecerik, Todd Andrew Hester, Thomas Rothoerl, Thomas Lampe, Nicolas Manfred Otto Heess, Jonathan Karl Scholz
  • Patent number: 11377417
    Abstract: The present application relates to novel 3-phenylpropionic acid derivatives which carry a branched or cyclic alkyl substituent in the 3-position, to processes for their preparation, to their use for the treatment and/or prevention of diseases and to their use for preparing medicaments for the treatment and/or prevention of diseases, in particular for the treatment and/or prevention of cardiovascular diseases.
    Type: Grant
    Filed: June 18, 2020
    Date of Patent: July 5, 2022
    Assignee: Bayer Intellectual Property GmbH
    Inventors: Michael Hahn, Thomas Lampe, Johannes-Peter Stasch, Karl-Heinz Schlemmer, Frank Wunder, Volkhart Min-Jian Li, Eva Maria Becker-Pelster, Friederike Stoll, Andreas Knorr, Elisabeth Woltering
  • Publication number: 20220205859
    Abstract: A pressure sensor apparatus and a method of operating the pressure sensor apparatus can include a hybrid pressure switch having a single switching point, the hybrid pressure switch comprising: a mechanical pressure switch; and an electronic pressure sensor, wherein the hybrid pressure switch monitors a media pressure with the mechanical pressure switch providing a switching function and the electronic pressure sensor providing a continuous pressure measurement with respect to the media pressure based on a continuous pressure output signal. The switching function can comprise a fail-safe switching function or a non-fail-safe switching function.
    Type: Application
    Filed: December 26, 2020
    Publication date: June 30, 2022
    Inventors: Christian Ruwe, Thomas Lampe, Tobias Otterpohl, Thomas Heyen, David Kucera
  • Publication number: 20220090943
    Abstract: A sensor system and method of operating the sensor system can include an indicator that is operable in resonance, the indicator being connected to a movable element of an actuator, and a sensor including sensor windings arranged in direct proximity to the indicator and external to a closed housing. The sensor can receive a position or an angle of the indicator, and can measure the position or the angle of the movable element inside the actuator through the closed housing based on inductive resonance facilitated by the indicator.
    Type: Application
    Filed: September 18, 2020
    Publication date: March 24, 2022
    Inventors: Christian L. Kamp, Thomas Lampe, Tobias Fechtel, Mathias Grewe
  • Publication number: 20210103815
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a policy neural network for use in controlling a real-world agent in a real-world environment. One of the methods includes training the policy neural network by optimizing a first task-specific objective that measures a performance of the policy neural network in controlling a simulated version of the real-world agent; and then training the policy neural network by jointly optimizing (i) a self-supervised objective that measures at least a performance of internal representations generated by the policy neural network on a self-supervised task performed on real-world data and (ii) a second task-specific objective that measures the performance of the policy neural network in controlling the simulated version of the real-world agent.
    Type: Application
    Filed: October 7, 2020
    Publication date: April 8, 2021
    Inventors: Rae Chan Jeong, Yusuf Aytar, David Khosid, Yuxiang Zhou, Jacqueline Ok-chan Kay, Thomas Lampe, Konstantinos Bousmalis, Francesco Nori
  • Publication number: 20200385335
    Abstract: The present application relates to novel 3-phenylpropionic acid derivatives which carry a branched or cyclic alkyl substituent in the 3-position, to processes for their preparation, to their use for the treatment and/or prevention of diseases and to their use for preparing medicaments for the treatment and/or prevention of diseases, in particular for the treatment and/or prevention of cardiovascular diseases.
    Type: Application
    Filed: June 18, 2020
    Publication date: December 10, 2020
    Applicant: Bayer Intellectual Property GmbH
    Inventors: Michael Hahn, Thomas Lampe, Johannes-Peter Stasch, Karl-Heinz Schlemmer, Frank Wunder, Volkhart Min-Jian Li, Eva Maria Becker-Pelster, Friederike Stoll, Andreas Knorr, Elisabeth Woltering
  • Publication number: 20200285909
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.
    Type: Application
    Filed: May 22, 2020
    Publication date: September 10, 2020
    Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
  • Patent number: 10664725
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.
    Type: Grant
    Filed: July 31, 2019
    Date of Patent: May 26, 2020
    Assignee: DeepMind Technologies Limited
    Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
  • Publication number: 20200151562
    Abstract: An off-policy reinforcement learning actor-critic neural network system configured to select actions from a continuous action space to be performed by an agent interacting with an environment to perform a task. An observation defines environment state data and reward data. The system has an actor neural network which learns a policy function mapping the state data to action data. A critic neural network learns an action-value (Q) function. A replay buffer stores tuples of the state data, the action data, the reward data and new state data. The replay buffer also includes demonstration transition data comprising a set of the tuples from a demonstration of the task within the environment. The neural network system is configured to train the actor neural network and the critic neural network off-policy using stored tuples from the replay buffer comprising tuples both from operation of the system and from the demonstration transition data.
    Type: Application
    Filed: June 28, 2018
    Publication date: May 14, 2020
    Inventors: Olivier Pietquin, Martin Riedmiller, Wang Fumin, Bilal Piot, Mel Vecerik, Todd Andrew Hester, Thomas Rothörl, Thomas Lampe, Nicolas Manfred Otto Heess, Jonathan Karl Scholz
  • Publication number: 20190354813
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for data-efficient reinforcement learning. One of the systems is a system for training an actor neural network used to select actions to be performed by an agent that interacts with an environment by receiving observations characterizing states of the environment and, in response to each observation, performing an action selected from a continuous space of possible actions, wherein the actor neural network maps observations to next actions in accordance with values of parameters of the actor neural network, and wherein the system comprises: a plurality of workers, wherein each worker is configured to operate independently of each other worker, wherein each worker is associated with a respective agent replica that interacts with a respective replica of the environment during the training of the actor neural network.
    Type: Application
    Filed: July 31, 2019
    Publication date: November 21, 2019
    Inventors: Martin Riedmiller, Roland Hafner, Mel Vecerik, Timothy Paul Lillicrap, Thomas Lampe, Ivaylo Popov, Gabriel Barth-Maron, Nicolas Manfred Otto Heess
  • Publication number: 20190185415
    Abstract: The present application relates to novel 3-phenylpropionic acid derivatives which carry a branched or cyclic alkyl substituent in the 3-position, to processes for their preparation, to their use for the treatment and/or prevention of diseases and to their use for preparing medicaments for the treatment and/or prevention of diseases, in particular for the treatment and/or prevention of cardiovascular diseases.
    Type: Application
    Filed: December 21, 2018
    Publication date: June 20, 2019
    Applicant: BAYER INTELLECTUAL PROPERTY GMBH
    Inventors: Michael HAHN, Thomas LAMPE, Johannes-Peter STASCH, Karl-Heinz SCHLEMMER, Frank WUNDER, Volkhart Min-Jian LI, Eva-Maria BECKER-PELSTER, Friederike STOLL, Andreas KNORR, Elisabeth WOLTERING
  • Patent number: 10259776
    Abstract: The present application relates to novel 3-phenylpropionic acid derivatives which carry a branched or cyclic alkyl substituent in the 3-position, to processes for their preparation, to their use for the treatment and/or prevention of diseases and to their use for preparing medicaments for the treatment and/or prevention of diseases, in particular for the treatment and/or prevention of cardiovascular diseases.
    Type: Grant
    Filed: June 6, 2018
    Date of Patent: April 16, 2019
    Assignee: Bayer Intellectual Property GmbH
    Inventors: Michael Hahn, Thomas Lampe, Johannes-Peter Stasch, Karl-Heinz Schlemmer, Frank Wunder, Volkhart Min-Jian Li, Eva-Maria Becker-Pelster, Friederike Stoll, Andreas Knorr, Elisabeth Woltering
  • Patent number: RE49698
    Abstract: The invention relates to substituted dihydroquinazolines and to processes for their preparation and also to their use for preparing medicaments for the treatment and/or prophylaxis of diseases, in particular for use as antiviral agents, in particular against cytomegalo viruses.
    Type: Grant
    Filed: July 21, 2020
    Date of Patent: October 17, 2023
    Assignee: AIC246 AG & Co. KG
    Inventors: Tobias Wunberg, Judith Baumeister, Ulrich Betz, Mario Jeske, Thomas Lampe, Susanne Nikolic, Jurgen Reefschlager, Rudolf Schohe-Loop, Frank Sussmeier, Holger Zimmermann, Rolf Grosser, Kerstin Henninger, Guy Hewlett, Jorg Keldenich, Dieter Lang, Peter Nell