Patents by Inventor David Richard So

David Richard So has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240112027
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for performing neural architecture search for machine learning models. In one aspect, a method comprises receiving training data for a machine learning, generating a plurality of candidate neural networks for performing the machine learning task, wherein each candidate neural network comprises a plurality of instances of a layer block composed of a plurality of layers, for each candidate neural network, selecting a respective type for each of the plurality of layers from a set of layer types that comprises, training the candidate neural network and evaluating performance scores for the trained candidate neural networks as applied to the machine learning task, and determining a final neural network for performing the machine learning task based at least on the performance scores for the candidate neural networks.
    Type: Application
    Filed: September 28, 2023
    Publication date: April 4, 2024
    Inventors: Yanqi Zhou, Yanping Huang, Yifeng Lu, Andrew M. Dai, Siamak Shakeri, Zhifeng Chen, James Laudon, Quoc V. Le, Da Huang, Nan Du, David Richard So, Daiyi Peng, Yingwei Cui, Jeffrey Adgate Dean, Chang Lan
  • Publication number: 20230274151
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for searching for an architecture for a neural network that performs a multi-modal task that requires operating on inputs that each include data from multiple different modalities.
    Type: Application
    Filed: March 30, 2021
    Publication date: August 31, 2023
    Inventors: Zhen Xu, David Richard So, Andrew M. Dai
  • Publication number: 20220383195
    Abstract: A method for searching for an output machine learning (ML) algorithm to perform an ML task is described. The method includes: receiving a set of training examples and a set of validation examples, and generating a sequence of candidate ML algorithms to perform the task. For each candidate ML algorithm in the sequence, the method includes: setting up one or more training parameters for the candidate ML algorithm by executing a respective candidate setup function, training the candidate ML algorithm by processing the set of training examples using a respective candidate predict function and a respective candidate learn function, and evaluating a performance of the trained candidate ML algorithm by executing the respective candidate predict function on the set of validation examples to determine a performance metric. The method includes selecting a trained candidate ML algorithm with the best performance metric as the output ML algorithm for the task.
    Type: Application
    Filed: February 8, 2021
    Publication date: December 1, 2022
    Inventors: Chen Liang, David Richard So, Esteban Alberto Real, Quoc V. Le
  • Publication number: 20220383119
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for performing a machine learning task on a network input to generate a network output. One of the systems includes an attention neural network configured to perform the machine learning task. The attention neural network includes one or more attentions layers that each include a squared ReLU activation layer, a depth-wise convolution layer, or both.
    Type: Application
    Filed: May 27, 2022
    Publication date: December 1, 2022
    Inventors: David Richard So, Quoc V. Le, Jr., Hanxiao Liu, Wojciech Andrzej Manke, Zihang Dai, Noam M. Shazeer
  • Publication number: 20220367052
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for performing a machine learning task on a network input to generate a network output. In one aspect, one of the systems includes a neural network configured to perform the machine learning task, the neural network including one or more blocks that each include a feedforward spatial transformation unit.
    Type: Application
    Filed: May 16, 2022
    Publication date: November 17, 2022
    Inventors: Hanxiao Liu, David Richard So, Quoc V. Le, Zihang Dai
  • Publication number: 20210256390
    Abstract: A method for receiving training data for training a neural network to perform a machine learning task and for searching for, using the training data, an optimized neural network architecture for performing the machine learning task is described. Searching for the optimized neural network architecture includes: maintaining population data; maintaining threshold data; and repeatedly performing the following operations: selecting one or more candidate architectures from the population data; generating a new architecture from the one or more selected candidate architectures; for the new architecture: training a neural network having the new architecture until termination criteria for the training are satisfied; and determining a final measure of fitness of the neural network having the new architecture after the training; and adding data defining the new architecture and the final measure of fitness for the neural network having the new architecture to the population data.
    Type: Application
    Filed: May 3, 2021
    Publication date: August 19, 2021
    Inventors: David Martin Dohan, David Richard So, Chen Liang, Quoc V. Le
  • Patent number: 10997503
    Abstract: A method for receiving training data for training a neural network to perform a machine learning task and for searching for, using the training data, an optimized neural network architecture for performing the machine learning task is described. Searching for the optimized neural network architecture includes: maintaining population data; maintaining threshold data; and repeatedly performing the following operations: selecting one or more candidate architectures from the population data; generating a new architecture from the one or more selected candidate architectures; for the new architecture: training a neural network having the new architecture until termination criteria for the training are satisfied; and determining a final measure of fitness of the neural network having the new architecture after the training; and adding data defining the new architecture and the final measure of fitness for the neural network having the new architecture to the population data.
    Type: Grant
    Filed: June 20, 2019
    Date of Patent: May 4, 2021
    Assignee: Google LLC
    Inventors: David Martin Dohan, David Richard So, Chen Liang, Quoc V. Le
  • Publication number: 20200401899
    Abstract: A method for receiving training data for training a neural network to perform a machine learning task and for searching for, using the training data, an optimized neural network architecture for performing the machine learning task is described. Searching for the optimized neural network architecture includes: maintaining population data; maintaining threshold data; and repeatedly performing the following operations: selecting one or more candidate architectures from the population data; generating a new architecture from the one or more selected candidate architectures; for the new architecture: training a neural network having the new architecture until termination criteria for the training are satisfied; and determining a final measure of fitness of the neural network having the new architecture after the training; and adding data defining the new architecture and the final measure of fitness for the neural network having the new architecture to the population data.
    Type: Application
    Filed: June 20, 2019
    Publication date: December 24, 2020
    Inventors: David Martin Dohan, David Richard So, Chen Liang, Quoc V. Le