Patents by Inventor Thomas Unterthiner

Thomas Unterthiner has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11983903
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing images using self-attention based neural networks. One of the methods includes obtaining one or more images comprising a plurality of pixels; determining, for each image of the one or more images, a plurality of image patches of the image, wherein each image patch comprises a different subset of the pixels of the image; processing, for each image of the one or more images, the corresponding plurality of image patches to generate an input sequence comprising a respective input element at each of a plurality of input positions, wherein a plurality of the input elements correspond to respective different image patches; and processing the input sequences using a neural network to generate a network output that characterizes the one or more images, wherein the neural network comprises one or more self-attention neural network layers.
    Type: Grant
    Filed: November 1, 2023
    Date of Patent: May 14, 2024
    Assignee: Google LLC
    Inventors: Neil Matthew Tinmouth Houlsby, Sylvain Gelly, Jakob D. Uszkoreit, Xiaohua Zhai, Georg Heigold, Lucas Klaus Beyer, Alexander Kolesnikov, Matthias Johannes Lorenz Minderer, Dirk Weissenborn, Mostafa Dehghani, Alexey Dosovitskiy, Thomas Unterthiner
  • Publication number: 20240062426
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing images using self-attention based neural networks. One of the methods includes obtaining one or more images comprising a plurality of pixels; determining, for each image of the one or more images, a plurality of image patches of the image, wherein each image patch comprises a different subset of the pixels of the image; processing, for each image of the one or more images, the corresponding plurality of image patches to generate an input sequence comprising a respective input element at each of a plurality of input positions, wherein a plurality of the input elements correspond to respective different image patches; and processing the input sequences using a neural network to generate a network output that characterizes the one or more images, wherein the neural network comprises one or more self-attention neural network layers.
    Type: Application
    Filed: November 1, 2023
    Publication date: February 22, 2024
    Inventors: Neil Matthew Tinmouth Houlsby, Sylvain Gelly, Jakob D. Uszkoreit, Xiaohua Zhai, Georg Heigold, Lucas Klaus Beyer, Alexander Kolesnikov, Matthias Johannes Lorenz Minderer, Dirk Weissenborn, Mostafa Dehghani, Alexey Dosovitskiy, Thomas Unterthiner
  • Publication number: 20220375211
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing images using mixer neural networks. One of the methods includes obtaining one or more images comprising a plurality of pixels; determining, for each image of the one or more images, a plurality of image patches of the image, wherein each image patch comprises a different subset of the pixels of the image; processing, for each image of the one or more images, the corresponding plurality of image patches to generate an input sequence comprising a respective input element at each of a plurality of input positions, wherein a plurality of the input elements correspond to respective different image patches; and processing the input sequences using a neural network to generate a network output that characterizes the one or more images, wherein the neural network comprises one or more mixer neural network layers.
    Type: Application
    Filed: May 5, 2022
    Publication date: November 24, 2022
    Inventors: Ilya Tolstikhin, Neil Matthew Tinmouth Houlsby, Alexander Kolesnikov, Lucas Klaus Beyer, Alexey Dosovitskiy, Mario Lucic, Xiaohua Zhai, Thomas Unterthiner, Daniel M. Keysers, Jakob D. Uszkoreit, Yin Ching Jessica Yung, Andreas Steiner
  • Publication number: 20220172066
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network to process images. One of the methods includes obtaining a training image; processing the training image using a first subnetwork to generate, for each of a plurality of first image patches of the training image, a relevance score; generating, using the relevance scores, one or more second image patches of the training image by performing one or more differentiable operations on the relevance scores; processing the one or more second image patches using a second subnetwork to generate a prediction about the training image; determining an error of the training network output; and generating a parameter update for the first subnetwork, comprising backpropagating gradients determined according to the error of the training network output through i) the second subnetwork, ii) the one or more differentiable operations, and iii) the first subnetwork.
    Type: Application
    Filed: November 30, 2021
    Publication date: June 2, 2022
    Inventors: Thomas Unterthiner, Alexey Dosovitskiy, Aravindh Mahendran, Dirk Weissenborn, Jakob D. Uszkoreit, Jean-Baptiste Cordonnier
  • Publication number: 20220108478
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing images using self-attention based neural networks. One of the methods includes obtaining one or more images comprising a plurality of pixels; determining, for each image of the one or more images, a plurality of image patches of the image, wherein each image patch comprises a different subset of the pixels of the image; processing, for each image of the one or more images, the corresponding plurality of image patches to generate an input sequence comprising a respective input element at each of a plurality of input positions, wherein a plurality of the input elements correspond to respective different image patches; and processing the input sequences using a neural network to generate a network output that characterizes the one or more images, wherein the neural network comprises one or more self-attention neural network layers.
    Type: Application
    Filed: October 1, 2021
    Publication date: April 7, 2022
    Inventors: Neil Matthew Tinmouth Houlsby, Sylvain Gelly, Jakob D. Uszkoreit, Xiaohua Zhai, Georg Heigold, Lucas Klaus Beyer, Alexander Kolesnikov, Matthias Johannes Lorenz Minderer, Dirk Weissenborn, Mostafa Dehghani, Alexey Dosovitskiy, Thomas Unterthiner
  • Publication number: 20210383199
    Abstract: A method involves receiving a perceptual representation including a plurality of feature vectors, and initializing a plurality of slot vectors represented by a neural network memory unit. Each respective slot vector is configured to represent a corresponding entity in the perceptual representation. The method also involves determining an attention matrix based on a product of the plurality of feature vectors transformed by a key function and the plurality of slot vectors transformed by a query function. Each respective value of a plurality of values along each respective dimension of the attention matrix is normalized with respect to the plurality of values. The method additionally involves determining an update matrix based on the plurality of feature vectors transformed by a value function and the attention matrix, and updating the plurality of slot vectors based on the update matrix by way of the neural network memory unit.
    Type: Application
    Filed: July 13, 2020
    Publication date: December 9, 2021
    Inventors: Dirk Weissenborn, Jakob Uszkoreit, Thomas Unterthiner, Aravindh Mahendran, Francesco Locatello, Thomas Kipf, Georg Heigold, Alexey Dosovitskiy
  • Publication number: 20210256422
    Abstract: Provided are systems and methods for predicting machine learning model performance from the model parameter values, including for use in making improved decisions with regard to early stopping of training procedures. As one example, the present disclosure discusses the prediction of the accuracy (e.g., relative to a defined task and testing dataset such as a computer vision task) of trained neural networks (e.g., convolutional neural networks (CNNs)), using only the parameter values (e.g., the values of the network's weights) as inputs. As such, one example aspect of the present disclosure is directed to computing systems that include and use a machine-learned performance prediction model that has been trained to predict performance values of machine-learned models based on their parameter values (e.g., weight values and/or hyperparameter values).
    Type: Application
    Filed: February 17, 2021
    Publication date: August 19, 2021
    Inventors: Thomas Unterthiner, Daniel Martin Keysers, Sylvain Gelly, Olivier Jean Andre Bousquet, Ilya Tolstikhin