Patents by Inventor Mohammad Norouzi

Mohammad Norouzi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11978141
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating images. In one aspect, a method includes: receiving an input text prompt including a sequence of text tokens in a natural language; processing the input text prompt using a text encoder neural network to generate a set of contextual embeddings of the input text prompt; and processing the contextual embeddings through a sequence of generative neural networks to generate a final output image that depicts a scene that is described by the input text prompt.
    Type: Grant
    Filed: May 19, 2023
    Date of Patent: May 7, 2024
    Assignee: Google LLC
    Inventors: Chitwan Saharia, William Chan, Mohammad Norouzi, Saurabh Saxena, Yi Li, Jay Ha Whang, David James Fleet, Jonathan Ho
  • Publication number: 20240127058
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network using a priority queue. One of the methods includes maintaining data identifying a set of K output sequences that were previously generated; selecting at least one of the output sequences from the set of output sequences; for each selected output sequence, determining a respective score; determining, for each selected sequence, a respective first update to the current values of the controller parameters; generating a batch of new output sequences using the controller neural network; obtaining a respective reward for each of the new output sequences; determining, from the new output sequences and the output sequences in the maintained data, the K output sequences that have the highest rewards; and modifying the maintained data.
    Type: Application
    Filed: September 21, 2023
    Publication date: April 18, 2024
    Inventors: Mohammad Norouzi, Daniel Aaron Abolafia, Quoc V. Le
  • Patent number: 11960519
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for classifying data objects. One of the methods includes obtaining data that associates each term in a vocabulary of terms with a respective high-dimensional representation of the term; obtaining classification data for a data object, wherein the classification data includes a respective score for each of a plurality of categories, and wherein each of the categories is associated with a respective category label; computing an aggregate high-dimensional representation for the data object from high-dimensional representations for the category labels associated with the categories and the respective scores; identifying a first term in the vocabulary of terms having a high-dimensional representation that is closest to the aggregate high-dimensional representation; and selecting the first term as a category label for the data object.
    Type: Grant
    Filed: August 20, 2020
    Date of Patent: April 16, 2024
    Assignee: Google LLC
    Inventors: Gregory Sean Corrado, Tomas Mikolov, Samy Bengio, Yoram Singer, Jonathon Shlens, Andrea L Frome, Jeffrey Adgate Dean, Mohammad Norouzi
  • Publication number: 20240062062
    Abstract: A method for determining a placement for machine learning model operations across multiple hardware devices is described.
    Type: Application
    Filed: October 3, 2023
    Publication date: February 22, 2024
    Inventors: Samuel Bengio, Mohammad Norouzi, Benoit Steiner, Jeffrey Adgate Dean, Hieu Hy Pham, Azalia Mirhoseini, Quoc V. Le, Naveen Kumar, Yuefeng Zhou, Rasmus Munk Larsen
  • Patent number: 11847571
    Abstract: Systems, methods, and computer program products for performing semi-supervised contrastive learning of visual representations are provided. For example, the present disclosure provides systems and methods that leverage particular data augmentation schemes and a learnable nonlinear transformation between the representation and the contrastive loss to provide improved visual representations. Further, the present disclosure also provides improvements for semi-supervised contrastive learning.
    Type: Grant
    Filed: July 12, 2022
    Date of Patent: December 19, 2023
    Assignee: GOOGLE LLC
    Inventors: Ting Chen, Geoffrey Everest Hinton, Simon Kornblith, Mohammad Norouzi
  • Publication number: 20230385990
    Abstract: A method includes receiving, by a computing device, training data comprising a plurality of pairs of images, wherein each pair comprises an image and at least one corresponding target version of the image. The method also includes training a neural network based on the training data to predict an enhanced version of an input image, wherein the training of the neural network comprises applying a forward Gaussian diffusion process that adds Gaussian noise to the at least one corresponding target version of each of the plurality of pairs of images to enable iterative denoising of the input image, wherein the iterative denoising is based on a reverse Markov chain associated with the forward Gaussian diffusion process. The method additionally includes outputting the trained neural network.
    Type: Application
    Filed: July 27, 2023
    Publication date: November 30, 2023
    Inventors: Chitwan Saharia, Jonathan Ho, William Chan, Tim Salimans, David Fleet, Mohammad Norouzi
  • Publication number: 20230377226
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating images. In one aspect, a method includes: receiving an input text prompt including a sequence of text tokens in a natural language; processing the input text prompt using a text encoder neural network to generate a set of contextual embeddings of the input text prompt; and processing the contextual embeddings through a sequence of generative neural networks to generate a final output image that depicts a scene that is described by the input text prompt.
    Type: Application
    Filed: May 19, 2023
    Publication date: November 23, 2023
    Inventors: Chitwan Saharia, William Chan, Mohammad Norouzi, Saurabh Saxena, Yi Li, Jay Ha Whang, David James Fleet, Jonathan Ho
  • Patent number: 11803747
    Abstract: A method for determining a placement for machine learning model operations across multiple hardware devices is described.
    Type: Grant
    Filed: May 20, 2020
    Date of Patent: October 31, 2023
    Assignee: Google LLC
    Inventors: Samuel Bengio, Mohammad Norouzi, Benoit Steiner, Jeffrey Adgate Dean, Hieu Hy Pham, Azalia Mirhoseini, Quoc V. Le, Naveen Kumar, Yuefeng Zhou, Rasmus Munk Larsen
  • Publication number: 20230342616
    Abstract: Systems, methods, and computer program products for performing semi-supervised contrastive learning of visual representations are provided. For example, the present disclosure provides systems and methods that leverage particular data augmentation schemes and a learnable nonlinear transformation between the representation and the contrastive loss to provide improved visual representations. Further, the present disclosure also provides improvements for semi-supervised contrastive learning.
    Type: Application
    Filed: June 28, 2023
    Publication date: October 26, 2023
    Inventors: Ting Chen, Simon Komblith, Mohammad Norouzi, Geoffrey Everest Hinton, Kevin Jordan Swersky
  • Patent number: 11797839
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network using a priority queue. One of the methods includes maintaining data identifying a set of K output sequences that were previously generated; selecting at least one of the output sequences from the set of output sequences; for each selected output sequence, determining a respective score; determining, for each selected sequence, a respective first update to the current values of the controller parameters; generating a batch of new output sequences using the controller neural network; obtaining a respective reward for each of the new output sequences; determining, from the new output sequences and the output sequences in the maintained data, the K output sequences that have the highest rewards; and modifying the maintained data.
    Type: Grant
    Filed: October 29, 2018
    Date of Patent: October 24, 2023
    Assignee: Google LLC
    Inventors: Mohammad Norouzi, Daniel Aaron Abolafia, Quoc V. Le
  • Publication number: 20230325658
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating outputs conditioned on network inputs using neural networks. In one aspect, a method comprises obtaining the network input; initializing a current network output; and generating the final network output by updating the current network output at each of a plurality of iterations, wherein each iteration corresponds to a respective noise level, and wherein the updating comprises, at each iteration: processing a model input for the iteration comprising (i) the current network output and (ii) the network input using a noise estimation neural network that is configured to process the model input to generate a noise output, wherein the noise output comprises a respective noise estimate for each value in the current network output; and updating the current network output using the noise estimate and the noise level for the iteration.
    Type: Application
    Filed: September 2, 2021
    Publication date: October 12, 2023
    Inventors: Nanxin Chen, Byungha Chun, William Chan, Ron J. Weiss, Mohammad Norouzi, Yu Zhang, Yonghui Wu
  • Patent number: 11769228
    Abstract: A method includes receiving, by a computing device, training data comprising a plurality of pairs of images, wherein each pair comprises an image and at least one corresponding target version of the image. The method also includes training a neural network based on the training data to predict an enhanced version of an input image, wherein the training of the neural network comprises applying a forward Gaussian diffusion process that adds Gaussian noise to the at least one corresponding target version of each of the plurality of pairs of images to enable iterative denoising of the input image, wherein the iterative denoising is based on a reverse Markov chain associated with the forward Gaussian diffusion process. The method additionally includes outputting the trained neural network.
    Type: Grant
    Filed: August 2, 2021
    Date of Patent: September 26, 2023
    Assignee: Google LLC
    Inventors: Chitwan Saharia, Jonathan Ho, William Chan, Tim Salimans, David Fleet, Mohammad Norouzi
  • Patent number: 11756166
    Abstract: A method includes receiving, by a computing device, training data comprising a plurality of pairs of images, wherein each pair comprises an image and at least one corresponding target version of the image. The method also includes training a neural network based on the training data to predict an enhanced version of an input image, wherein the training of the neural network comprises applying a forward Gaussian diffusion process that adds Gaussian noise to the at least one corresponding target version of each of the plurality of pairs of images to enable iterative denoising of the input image, wherein the iterative denoising is based on a reverse Markov chain associated with the forward Gaussian diffusion process. The method additionally includes outputting the trained neural network.
    Type: Grant
    Filed: January 17, 2023
    Date of Patent: September 12, 2023
    Assignee: Google LLC
    Inventors: Chitwan Saharia, Jonathan Ho, William Chan, Tim Salimans, David Fleet, Mohammad Norouzi
  • Publication number: 20230260652
    Abstract: Systems and methods can perform self-supervised machine learning for improved medical image analysis. As one example, self-supervised learning on ImageNet, followed by additional self-supervised learning on unlabeled medical images from the target domain of interest, followed by fine-tuning on labeled medical images from the target domain significantly improves the accuracy of medical image classifiers such as, for example diagnostic models. Another example aspect of the present disclosure is directed to a novel Multi-Instance Contrastive Learning (MICLe) method that uses multiple different medical images that share one or more attributes (e.g., multiple images that depict the same underlying pathology and/or the same patient) to construct more informative positive pairs for self-supervised learning.
    Type: Application
    Filed: December 10, 2021
    Publication date: August 17, 2023
    Inventors: Shekoofeh Azizi, Wen Yau Aaron Loh, Zachary William Beaver, Ting Chen, Jonathan Paul Deaton, Jan Freyberg, Alan Prasana Karthikesalingam, Simon Kornblith, Basil Mustafa, Mohammad Norouzi, Vivek Natarajan, Fiona Keleher Ryan
  • Publication number: 20230252974
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating waveforms conditioned on phoneme sequences. In one aspect, a method comprises: obtaining a phoneme sequence; processing the phoneme sequence using an encoder neural network to generate a hidden representation of the phoneme sequence; generating, from the hidden representation, a conditioning input; initializing a current waveform output; and generating a final waveform output that defines an utterance of the phoneme sequence by a speaker by updating the current waveform output at each of a plurality of iterations, wherein each iteration corresponds to a respective noise level, and wherein the updating comprises, at each iteration: processing (i) the current waveform output and (ii) the conditioning input using a noise estimation neural network to generate a noise output; and updating the current waveform output using the noise output and the noise level for the iteration.
    Type: Application
    Filed: September 2, 2021
    Publication date: August 10, 2023
    Inventors: Byungha Chun, Mohammad Norouzi, Nanxin Chen, Ron J. Weiss, William Chan, Yu Zhang, Yonghui Wu
  • Patent number: 11699074
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a sequence generation neural network. One of the methods includes obtaining a batch of training examples; for each of the training examples: processing the training network input in the training example using the neural network to generate an output sequence; for each particular output position in the output sequence: identifying a prefix that includes the system outputs at positions before the particular output position in the output sequence, for each possible system output in the vocabulary, determining a highest quality score that can be assigned to any candidate output sequence that includes the prefix followed by the possible system output, and determining an update to the current values of the network parameters that increases a likelihood that the neural network generates a system output at the position that has a high quality score.
    Type: Grant
    Filed: January 17, 2020
    Date of Patent: July 11, 2023
    Assignee: Google LLC
    Inventors: Mohammad Norouzi, William Chan, Sara Sabour Rouh Aghdam
  • Publication number: 20230153959
    Abstract: A method includes receiving, by a computing device, training data comprising a plurality of pairs of images, wherein each pair comprises an image and at least one corresponding target version of the image. The method also includes training a neural network based on the training data to predict an enhanced version of an input image, wherein the training of the neural network comprises applying a forward Gaussian diffusion process that adds Gaussian noise to the at least one corresponding target version of each of the plurality of pairs of images to enable iterative denoising of the input image, wherein the iterative denoising is based on a reverse Markov chain associated with the forward Gaussian diffusion process. The method additionally includes outputting the trained neural network.
    Type: Application
    Filed: January 17, 2023
    Publication date: May 18, 2023
    Inventors: Chitwan Saharia, Jonathan Ho, William Chan, Tim Salimans, David Fleet, Mohammad Norouzi
  • Publication number: 20230103638
    Abstract: A method includes receiving training data comprising a plurality of pairs of images. Each pair comprises a noisy image and a denoised version of the noisy image. The method also includes training a multi-task diffusion model to perform a plurality of image-to-image translation tasks, wherein the training comprises iteratively generating a forward diffusion process by predicting, at each iteration in a sequence of iterations and based on a current noisy estimate of the denoised version of the noisy image, noise data for a next noisy estimate of the denoised version of the noisy image, updating, at each iteration, the current noisy estimate to the next noisy estimate by combining the current noisy estimate with the predicted noise data, and determining a reverse diffusion process by inverting the forward diffusion process to predict the denoised version of the noisy image. The method additionally includes providing the trained diffusion model.
    Type: Application
    Filed: October 5, 2022
    Publication date: April 6, 2023
    Inventors: Chitwan Saharia, Mohammad Norouzi, William Chan, Huiwen Chang, David James Fleet, Christopher Albert Lee, Jonathan Ho, Tim Salimans
  • Publication number: 20230075716
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for sequence modeling. One of the methods includes receiving an input sequence having a plurality of input positions; determining a plurality of blocks of consecutive input positions; processing the input sequence using a neural network to generate a latent alignment, comprising, at each of a plurality of input time steps: receiving a partial latent alignment from a previous input time step; selecting an input position in each block, wherein the token at the selected input position of the partial latent alignment in each block is a mask token; and processing the partial latent alignment and the input sequence using the neural network to generate a new latent alignment, wherein the new latent alignment comprises, at the selected input position in each block, an output token or a blank token; and generating, using the latent alignment, an output sequence.
    Type: Application
    Filed: February 8, 2021
    Publication date: March 9, 2023
    Inventors: William Chan, Chitwan Saharia, Geoffrey E. Hinton, Mohammad Norouzi, Navdeep Jaitly
  • Publication number: 20230067841
    Abstract: A method includes receiving, by a computing device, training data comprising a plurality of pairs of images, wherein each pair comprises an image and at least one corresponding target version of the image. The method also includes training a neural network based on the training data to predict an enhanced version of an input image, wherein the training of the neural network comprises applying a forward Gaussian diffusion process that adds Gaussian noise to the at least one corresponding target version of each of the plurality of pairs of images to enable iterative denoising of the input image, wherein the iterative denoising is based on a reverse Markov chain associated with the forward Gaussian diffusion process. The method additionally includes outputting the trained neural network.
    Type: Application
    Filed: August 2, 2021
    Publication date: March 2, 2023
    Inventors: Chitwan Saharia, Jonathan Ho, William Chan, Tim Salimans, David Fleet, Mohammad Norouzi