Patents by Inventor Klaus Greff

Klaus Greff has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240119697
    Abstract: Example embodiments of the present disclosure provide an example computer-implemented method for constructing a three-dimensional semantic segmentation of a scene from two-dimensional inputs. The example method includes obtaining, by a computing system comprising one or more processors, an image set comprising one or more views of a subject scene. The example method includes generating, by the computing system and based at least in part on the image set, a scene representation describing the subject scene in three dimensions. The example method includes generating, by the computing system and using a machine-learned semantic segmentation model framework, a multidimensional field of probability distributions over semantic categories, the multidimensional field defined over the three dimensions of the subject scene. The example method includes outputting, by the computing system, classification data for at least one location in the subject scene.
    Type: Application
    Filed: October 10, 2022
    Publication date: April 11, 2024
    Inventors: Daniel Christopher Duckworth, Suhani Deepak-Ranu Vora, Noha Radwan, Klaus Greff, Henning Meyer, Kyle Adam Genova, Seyed Mohammad Mehdi Sajjadi, Etienne François Régis Pot, Andrea Tagliasacchi
  • Publication number: 20240096001
    Abstract: Provided are machine learning models that generate geometry-free neural scene representations through efficient object-centric novel-view synthesis. In particular, one example aspect of the present disclosure provides a novel framework in which an encoder model (e.g., an encoder transformer network) processes one or more RGB images (with or without pose) to produce a fully latent scene representation that can be passed to a decoder model (e.g., a decoder transformer network). Given one or more target poses, the decoder model can synthesize images in a single forward pass. In some example implementations, because transformers are used rather than convolutional or MLP networks, the encoder can learn an attention model that extracts enough 3D information about a scene from a small set of images to render novel views with correct projections, parallax, occlusions, and even semantics, without explicit geometry.
    Type: Application
    Filed: November 15, 2022
    Publication date: March 21, 2024
    Inventors: Seyed Mohammad Mehdi Sajjadi, Henning Meyer, Etienne François Régis Pot, Urs Michael Bergmann, Klaus Greff, Noha Radwan, Suhani Deepak-Ranu Vora, Mario Lu¢i¢, Daniel Christopher Duckworth, Thomas Allen Funkhouser, Andrea Tagliasacchi
  • Publication number: 20220383628
    Abstract: A method includes obtaining first feature vectors and second feature vectors representing contents of a first and second image frame, respectively, of an input video. The method may also include generating, based on the first feature vectors, first slot vectors, where each slot vector represents attributes of a corresponding entity as represented in the first image frame, and generating, based on the first slot vectors, predicted slot vectors including a corresponding predicted slot vector that represents a transition of the attributes of the corresponding entity from the first to the second image frame. The method may additionally include generating, based on the predicted slot vectors and the second feature vectors, second slot vectors including a corresponding slot vector that represents the attributes of the corresponding entity as represented in the second image frame, and determining an output based on the predicted slot vectors or the second slot vectors.
    Type: Application
    Filed: April 21, 2022
    Publication date: December 1, 2022
    Inventors: Thomas Kipf, Gamaleldin Elsayed, Aravindh Mahendran, Austin Charles Stone, Sara Sabour Rouh Aghdam, Georg Heigold, Rico Jonschkowski, Alexey Dosovitskiy, Klaus Greff
  • Patent number: 11481585
    Abstract: Disclosed is a computer-implemented method for segmenting input data. In the method a plurality of tags is generated; the input data is masked with the plurality of tags; a plurality of output reconstructions is generated by inputting the plurality of masked input data to one of the following: a denoising neural network, a variational autoencoder; a plurality of values representing distances of each plurality of output reconstructions to the input data are determined; a plurality of updated versions of input data is generated by applying at least one of the determined values representing distances of each plurality of output reconstructions to the input data; and updated output reconstructions are generated by inputting the plurality of updated versions of input data to one of the networks. Also disclosed is a method for training the network and a processing unit.
    Type: Grant
    Filed: May 19, 2017
    Date of Patent: October 25, 2022
    Assignee: Canary Capital LLC
    Inventors: Harri Valpola, Klaus Greff
  • Patent number: 10984320
    Abstract: A computer-based method includes receiving an input signal at a neuron in a computer-based neural network that includes a plurality of neuron layers, applying a first non-linear transform to the input signal at the neuron to produce a plain signal, and calculating a weighted sum of a first component of the input signal and the plain signal at the neuron. In a typical implementation, the first non-linear transform is a function of the first component of the input signal and at least a second component of the input signal.
    Type: Grant
    Filed: May 1, 2017
    Date of Patent: April 20, 2021
    Assignee: Nnaisense SA
    Inventors: Rupesh Kumar Srivastava, Klaus Greff
  • Publication number: 20190220691
    Abstract: Disclosed is a computer-implemented method for segmenting input data. In the method a plurality of tags is generated; the input data is masked with the plurality of tags; a plurality of output reconstructions is generated by inputting the plurality of masked input data to one of the following: a denoising neural network, a variational autoencoder; a plurality of values representing distances of each plurality of output reconstructions to the input data are determined; a plurality of updated versions of input data is generated by applying at least one of the determined values representing distances of each plurality of output reconstructions to the input data; and updated output reconstructions are generated by inputting the plurality of updated versions of input data to one of the networks. Also disclosed is a method for training the network and a processing unit.
    Type: Application
    Filed: May 19, 2017
    Publication date: July 18, 2019
    Applicant: Curious AI Oy
    Inventors: Harri VALPOLA, Klaus GREFF
  • Publication number: 20170316308
    Abstract: A computer-based method includes receiving an input signal at a neuron in a computer-based neural network that includes a plurality of neuron layers, applying a first non-linear transform to the input signal at the neuron to produce a plain signal, and calculating a weighted sum of a first component of the input signal and the plain signal at the neuron. In a typical implementation, the first non-linear transform is a function of the first component of the input signal and at least a second component of the input signal.
    Type: Application
    Filed: May 1, 2017
    Publication date: November 2, 2017
    Inventors: Rupesh Kumar Srivastava, Klaus Greff