Patents by Inventor Suhani Deepak-Ranu Vora

Suhani Deepak-Ranu Vora has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240119697
    Abstract: Example embodiments of the present disclosure provide an example computer-implemented method for constructing a three-dimensional semantic segmentation of a scene from two-dimensional inputs. The example method includes obtaining, by a computing system comprising one or more processors, an image set comprising one or more views of a subject scene. The example method includes generating, by the computing system and based at least in part on the image set, a scene representation describing the subject scene in three dimensions. The example method includes generating, by the computing system and using a machine-learned semantic segmentation model framework, a multidimensional field of probability distributions over semantic categories, the multidimensional field defined over the three dimensions of the subject scene. The example method includes outputting, by the computing system, classification data for at least one location in the subject scene.
    Type: Application
    Filed: October 10, 2022
    Publication date: April 11, 2024
    Inventors: Daniel Christopher Duckworth, Suhani Deepak-Ranu Vora, Noha Radwan, Klaus Greff, Henning Meyer, Kyle Adam Genova, Seyed Mohammad Mehdi Sajjadi, Etienne François Régis Pot, Andrea Tagliasacchi
  • Publication number: 20240096001
    Abstract: Provided are machine learning models that generate geometry-free neural scene representations through efficient object-centric novel-view synthesis. In particular, one example aspect of the present disclosure provides a novel framework in which an encoder model (e.g., an encoder transformer network) processes one or more RGB images (with or without pose) to produce a fully latent scene representation that can be passed to a decoder model (e.g., a decoder transformer network). Given one or more target poses, the decoder model can synthesize images in a single forward pass. In some example implementations, because transformers are used rather than convolutional or MLP networks, the encoder can learn an attention model that extracts enough 3D information about a scene from a small set of images to render novel views with correct projections, parallax, occlusions, and even semantics, without explicit geometry.
    Type: Application
    Filed: November 15, 2022
    Publication date: March 21, 2024
    Inventors: Seyed Mohammad Mehdi Sajjadi, Henning Meyer, Etienne François Régis Pot, Urs Michael Bergmann, Klaus Greff, Noha Radwan, Suhani Deepak-Ranu Vora, Mario Lu¢i¢, Daniel Christopher Duckworth, Thomas Allen Funkhouser, Andrea Tagliasacchi