Patents by Inventor Joel William Veness

Joel William Veness has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11842264
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for a neural network system comprising one or more gated linear networks. A system includes: one or more gated linear networks, wherein each gated linear network corresponds to a respective data value in an output data sample and is configured to generate a network probability output that defines a probability distribution over possible values for the corresponding data value, wherein each gated linear network comprises a plurality of layers, wherein the plurality of layers comprises a plurality of gated linear layers, wherein each gated linear layer has one or more nodes, and wherein each node is configured to: receive a plurality of inputs, receive side information for the node; combine the plurality of inputs according to a set of weights defined by the side information, and generate and output a node probability output for the corresponding data value.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: December 12, 2023
    Assignee: DeepMind Technologies Limited
    Inventors: Agnieszka Grabska-Barwinska, Peter Toth, Christopher Mattern, Avishkar Bhoopchand, Tor Lattimore, Joel William Veness
  • Publication number: 20230079338
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer-readable storage media, for training a neural network to control a real-world agent interacting with a real-world environment to cause the real-world agent to perform a particular task. One of the methods includes training the neural network to determine first values of the parameters by optimizing a first task-specific objective that measures a performance of the policy neural network in controlling a simulated version of the real-world agent; obtaining real-world data generated from interactions of the real-world agent with the real-world environment; and training the neural network to determine trained values of the parameters from the first values of the parameters by jointly optimizing (i) a self-supervised objective that measures at least a performance of internal representations generated by the neural network on a self-supervised task performed on the real-world data and (ii) a second task-specific objective.
    Type: Application
    Filed: October 8, 2020
    Publication date: March 16, 2023
    Inventors: Eren Sezener, Joel William Veness, Marcus Hutter, Jianan Wang, David Budden
  • Patent number: 11429898
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for evaluating reinforcement learning policies. One of the methods includes receiving a plurality of training histories for a reinforcement learning agent; determining a total reward for each training observation in the training histories; partitioning the training observations into a plurality of partitions; determining, for each partition and from the partitioned training observations, a probability that the reinforcement learning agent will receive the total reward for the partition if the reinforcement learning agent performs the action for the partition in response to receiving the current observation; determining, from the probabilities and for each total reward, a respective estimated value of performing each action in response to receiving the current observation; and selecting an action from the pre-determined set of actions from the estimated values in accordance with an action selection policy.
    Type: Grant
    Filed: October 14, 2019
    Date of Patent: August 30, 2022
    Assignee: DeepMind Technologies Limited
    Inventors: Joel William Veness, Marc Gendron-Bellemare
  • Publication number: 20200349418
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for a neural network system comprising one or more gated linear networks. A system includes: one or more gated linear networks, wherein each gated linear network corresponds to a respective data value in an output data sample and is configured to generate a network probability output that defines a probability distribution over possible values for the corresponding data value, wherein each gated linear network comprises a plurality of layers, wherein the plurality of layers comprises a plurality of gated linear layers, wherein each gated linear layer has one or more nodes, and wherein each node is configured to: receive a plurality of inputs, receive side information for the node; combine the plurality of inputs according to a set of weights defined by the side information, and generate and output a node probability output for the corresponding data value.
    Type: Application
    Filed: November 30, 2018
    Publication date: November 5, 2020
    Inventors: Agnieszka Grabska-Barwinska, Peter Toth, Christopher Mattern, Avishkar Bhoopchand, Tor Lattimore, Joel William Veness
  • Patent number: 10445653
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for evaluating reinforcement learning policies. One of the methods includes receiving a plurality of training histories for a reinforcement learning agent; determining a total reward for each training observation in the training histories; partitioning the training observations into a plurality of partitions; determining, for each partition and from the partitioned training observations, a probability that the reinforcement learning agent will receive the total reward for the partition if the reinforcement learning agent performs the action for the partition in response to receiving the current observation; determining, from the probabilities and for each total reward, a respective estimated value of performing each action in response to receiving the current observation; and selecting an action from the pre-determined set of actions from the estimated values in accordance with an action selection policy.
    Type: Grant
    Filed: August 7, 2015
    Date of Patent: October 15, 2019
    Assignee: DeepMind Technologies Limited
    Inventors: Joel William Veness, Marc Gendron-Bellemare
  • Publication number: 20190236482
    Abstract: A method of training a machine learning model having multiple parameters, in which the machine learning model has been trained on a first machine learning task to determine first values of the parameters of the machine learning model.
    Type: Application
    Filed: July 18, 2017
    Publication date: August 1, 2019
    Inventors: Guillaume Desjardins, Razvan Pascanu, Raia Thais Hadsell, James Kirkpatrick, Joel William Veness, Neil Charles Rabinowitz