Patents by Inventor Daniel Matthew Cer

Daniel Matthew Cer has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240020546
    Abstract: Systems and methods for prompt tuning can utilize previously-learned prompts for the initialization of tuning for prompts on different tasks that may differ from the task associated with the previously-learned prompt. The prompt being utilized for initialization can be a generic prompt and/or may be a prompt selected based on a determined similarity between two or more task embeddings.
    Type: Application
    Filed: July 13, 2022
    Publication date: January 18, 2024
    Inventors: Tu Thanh Vu, Daniel Matthew Cer, Noah Constant, Brian David Lester, Rami Al-Rfou
  • Patent number: 11769011
    Abstract: The present disclosure provides a novel sentence-level representation learning method Conditional Masked Language Modeling (CMLM) for training on large scale unlabeled corpora. CMLM outperforms the previous state-of-the-art English sentence embedding models, including those trained with (semi-)supervised signals. For multilingual representations learning, it is shown that co-training CMLM with bitext retrieval and cross-lingual natural language inference (NL) fine-tuning achieves state-of-the-art performance. It is also shown that multilingual representations have the same language bias and principal component removal (PCR) can eliminate the bias by separating language identity information from semantics.
    Type: Grant
    Filed: December 18, 2020
    Date of Patent: September 26, 2023
    Assignee: GOOGLE LLC
    Inventors: Yinfei Yang, Ziyi Yang, Daniel Matthew Cer
  • Publication number: 20220198144
    Abstract: The present disclosure provides a novel sentence-level representation learning method Conditional Masked Language Modeling (CMLM) for training on large scale unlabeled corpora. CMLM outperforms the previous state-of-the-art English sentence embedding models, including those trained with (semi-)supervised signals. For multilingual representations learning, it is shown that co-training CMLM with bitext retrieval and cross-lingual NLI fine-tuning achieves state-of-the-art performance. It is also shown that multilingual representations have the same language bias and principal component removal (PCR) can eliminate the bias by separating language identity information from semantics.
    Type: Application
    Filed: December 18, 2020
    Publication date: June 23, 2022
    Inventors: Yinfei Yang, Ziyi Yang, Daniel Matthew Cer