Patents by Inventor Ehsan Amid

Ehsan Amid has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240249193
    Abstract: Generally, the present disclosure is directed to enhanced federated learning (FL) that employs a set of clients with varying amounts of computational resources (e.g., system memory, storage, and processing bandwidth). To overcome limitations of conventional FL methods that employ a set of clients with varying amounts of computational resources, the embodiments run multi-directional knowledge distillation between the server models produced by each federated averaging (FedAvg) pool, using unlabeled server data as the distillation dataset. By co-distilling the two (or more) models frequently over the course of FedAvg rounds, information is shared between the pools without sharing model parameters. This leads to increased performance and faster convergence (in fewer federated rounds).
    Type: Application
    Filed: January 19, 2024
    Publication date: July 25, 2024
    Inventors: Jared Alexander Lichtarge, Rajiv Mathews, Rohan Anil, Ehsan Amid, Shankar Kumar
  • Publication number: 20240233707
    Abstract: A method includes receiving distillation data including a plurality of out-of-domain training utterances. For each particular out-of-domain training utterance of the distillation data, the method includes generating a corresponding augmented out-of-domain training utterance, and generating, using a teacher ASR model trained on training data corresponding to a target domain, a pseudo-label corresponding to the corresponding augmented out-of-domain training utterance. The method also includes distilling a student ASR model from the teacher ASR model by training the student ASR model using the corresponding augmented out-of-domain training utterances paired with the corresponding pseudo-labels generated by the teacher ASR model.
    Type: Application
    Filed: October 17, 2023
    Publication date: July 11, 2024
    Applicant: Google LLC
    Inventors: Tien-Ju Yang, You-Chi Cheng, Shankar Kumar, Jared Lichtarge, Ehsan Amid, Yuxin Ding, Rajiv Mathews, Mingqing Chen
  • Publication number: 20240221772
    Abstract: A method of phrase extraction for ASR models includes obtaining audio data characterizing an utterance and a corresponding ground-truth transcription of the utterance and modifying the audio data to obfuscate a particular phrase recited in the utterance. The method also includes processing, using a trained ASR model, the modified audio data to generate a predicted transcription of the utterance, and determining whether the predicted transcription includes the particular phrase by comparing the predicted transcription of the utterance to the ground-truth transcription of the utterance. When the predicted transcription includes the particular phrase, the method includes generating an output indicating that the trained ASR model leaked the particular phrase from a training data set used to train the ASR model.
    Type: Application
    Filed: March 19, 2024
    Publication date: July 4, 2024
    Applicant: Google LLC
    Inventors: Ehsan Amid, Om Dipakbhai Thakkar, Rajiv Mathews, Francoise Beaufays
  • Publication number: 20240194192
    Abstract: Information can be distilled from a global automatic speech recognition (ASR) model to a client ASR model. Many implementations include using an RNN-T model as the ASR model, where the global ASR model includes a global encoder, a joint network, a prediction network, and where the client ASR model includes a client encoder, the joint network, and the prediction network. Various implementations include using principal component analysis (PCA) while training the global ASR model to learn a mean vector and a set of principal components corresponding to the global ASR model. Additional or alternative implementations include training the client ASR model to generate one or more predicted coefficients of the global ASR model.
    Type: Application
    Filed: December 9, 2022
    Publication date: June 13, 2024
    Inventors: Ehsan Amid, Rajiv Mathews, Shankar Kumar, Jared Lichtarge, Mingqing Chen, Tien-Ju Yang, Yuxin Ding
  • Publication number: 20240135918
    Abstract: A method includes receiving distillation data including a plurality of out-of-domain training utterances. For each particular out-of-domain training utterance of the distillation data, the method includes generating a corresponding augmented out-of-domain training utterance, and generating, using a teacher ASR model trained on training data corresponding to a target domain, a pseudo-label corresponding to the corresponding augmented out-of-domain training utterance. The method also includes distilling a student ASR model from the teacher ASR model by training the student ASR model using the corresponding augmented out-of-domain training utterances paired with the corresponding pseudo-labels generated by the teacher ASR model.
    Type: Application
    Filed: October 16, 2023
    Publication date: April 25, 2024
    Applicant: Google LLC
    Inventors: Tien-Ju Yang, You-Chi Cheng, Shankar Kumar, Jared Lichtarge, Ehsan Amid, Yuxin Ding, Rajiv Mathews, Mingqing Chen
  • Patent number: 11955134
    Abstract: A method of phrase extraction for ASR models includes obtaining audio data characterizing an utterance and a corresponding ground-truth transcription of the utterance and modifying the audio data to obfuscate a particular phrase recited in the utterance. The method also includes processing, using a trained ASR model, the modified audio data to generate a predicted transcription of the utterance, and determining whether the predicted transcription includes the particular phrase by comparing the predicted transcription of the utterance to the ground-truth transcription of the utterance. When the predicted transcription includes the particular phrase, the method includes generating an output indicating that the trained ASR model leaked the particular phrase from a training data set used to train the ASR model.
    Type: Grant
    Filed: December 13, 2021
    Date of Patent: April 9, 2024
    Assignee: Google LLC
    Inventors: Ehsan Amid, Om Thakkar, Rajiv Mathews, Francoise Beaufays
  • Publication number: 20240095582
    Abstract: During a round of decentralized learning for updating of a global machine learning (ML) model, remote processor(s) of a remote system may transmit, to a population of computing devices, primary weights for a primary version of the global ML model, and cause each of the computing devices to generate a corresponding update for the primary version of the global ML model. Further, the remote processor(s) may cause the primary version of the global ML model to be updated based on the corresponding updates that are received during the round of decentralized learning. However, the remote processor(s) may receive other corresponding updates subsequent to the round of decentralized learning. Accordingly, various techniques described herein (e.g., FARe-DUST, FeAST on MSG, and/or other techniques) enable the other corresponding updates to be utilized in achieving a final version of the global ML model.
    Type: Application
    Filed: December 6, 2022
    Publication date: March 21, 2024
    Inventors: Andrew Hard, Sean Augenstein, Rohan Anil, Rajiv Mathews, Lara McConnaughey, Ehsan Amid, Antonious Girgis
  • Publication number: 20240070530
    Abstract: Implementations disclosed herein are directed to a hybrid federated learning (FL) technique that utilizes both federated averaging (FA) and federated distillation (FD) during a given round of FL of a given global machine learning (ML) model. Implementations may identify a population of client devices to participate in the given round of FL, determine a corresponding quantity of instances of client data available at each of the client devices that may be utilized during the given round of FL, and select different subsets of the client devices based on the corresponding quantity of instances of client data. Further, implementations may cause a first subset of the client devices to generate a corresponding FA update and a second subset of client devices to generate a corresponding FD update. Moreover, implementations may subsequently update the given global ML model based on the corresponding FA updates and the corresponding FD updates.
    Type: Application
    Filed: December 5, 2022
    Publication date: February 29, 2024
    Inventors: Ehsan Amid, Rajiv Mathews, Rohan Anil, Shankar Kumar, Jared Lichtarge
  • Publication number: 20230178094
    Abstract: A method of phrase extraction for ASR models includes obtaining audio data characterizing an utterance and a corresponding ground-truth transcription of the utterance and modifying the audio data to obfuscate a particular phrase recited in the utterance. The method also includes processing, using a trained ASR model, the modified audio data to generate a predicted transcription of the utterance, and determining whether the predicted transcription includes the particular phrase by comparing the predicted transcription of the utterance to the ground-truth transcription of the utterance. When the predicted transcription includes the particular phrase, the method includes generating an output indicating that the trained ASR model leaked the particular phrase from a training data set used to train the ASR model.
    Type: Application
    Filed: December 13, 2021
    Publication date: June 8, 2023
    Applicant: Google LLC
    Inventors: Ehsan Amid, Om Thakkar, Rajiv Mathews, Francoise Beaufays
  • Publication number: 20230103911
    Abstract: A method include obtaining a set of differentially private (DP) gradients each generated based on processing corresponding private data, and obtaining a set of public gradients each generated based on processing corresponding public data. The method also includes applying mirror descent to the set of public gradients to learn a geometry for the set of DP gradients, and reshaping the set of DP gradients based on the learned geometry. The method further includes training a machine learning model based on the reshaped set of DP gradients.
    Type: Application
    Filed: October 4, 2022
    Publication date: April 6, 2023
    Applicant: Google LLC
    Inventors: Om Dipakbhai Thakkar, Ehsan Amid, Arun Ganesh, Rajiv Mathews, Swaroop Ramaswamy, Shuang Song, Thomas Steinke, Vinith Suriyakumar, Abhradeep Guha Thakurta
  • Publication number: 20230044078
    Abstract: A method includes receiving training data for a machine learning model, the training data comprising a plurality of training examples and a corresponding plurality of labels. The method further includes dividing the training data into a plurality of training batches. For each training batch of the plurality of training batches, the method additionally includes learning a weight for each training example in the training batch that minimizes a sum of weighted losses for the training batch subject to a divergence constraint, where the divergence constraint limits a divergence of the learned weights for the training batch from a reference distribution, where the divergence is determined according to a chosen divergence measure. The method also includes training the machine learning model with each training batch of the plurality of training batches using the learned weight for each training example in the training batch. The method additionally includes providing the trained machine learning model.
    Type: Application
    Filed: July 29, 2022
    Publication date: February 9, 2023
    Inventors: Abhishek Kumar, Ehsan Amid
  • Publication number: 20220253713
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network using local layer-wise losses.
    Type: Application
    Filed: February 7, 2022
    Publication date: August 11, 2022
    Inventors: Ehsan Amid, Manfred Klaus Warmuth, Rohan Anil
  • Patent number: 10127694
    Abstract: The present disclosure relates to a triplet embedding system that improves dimensionality reduction through exponential triplet embedding. In particular, the triplet embedding system employs heavy-tailed properties of t-exponential distributions and robust non-convex loss functions to improve visualizations in the presence of noisy data. In addition, the triplet embedding system uses triplet similarity weighting and improved sampling to improve and accelerate triplet embedding in large datasets. Overall, the triplet embedding system produces improved dimensionality reduction visualizations, which accurately reveal the underlying structure of the real-world high-dimensional datasets in lower-dimensional space.
    Type: Grant
    Filed: November 18, 2016
    Date of Patent: November 13, 2018
    Assignee: ADOBE SYSTEMS INCORPORATED
    Inventors: Nikolaos Vlassis, Ehsan Amid
  • Publication number: 20180144518
    Abstract: The present disclosure relates to a triplet embedding system that improves dimensionality reduction through exponential triplet embedding. In particular, the triplet embedding system employs heavy-tailed properties of t-exponential distributions and robust non-convex loss functions to improve visualizations in the presence of noisy data. In addition, the triplet embedding system uses triplet similarity weighting and improved sampling to improve and accelerate triplet embedding in large datasets. Overall, the triplet embedding system produces improved dimensionality reduction visualizations, which accurately reveal the underlying structure of the real-world high-dimensional datasets in lower-dimensional space.
    Type: Application
    Filed: November 18, 2016
    Publication date: May 24, 2018
    Inventors: Nikolaos Vlassis, Ehsan Amid